HR knowledge migration: when to rebuild versus retrofit before transformation
When standing up a new HR transformation initiative, one of the most costly early decisions is whether to migrate legacy knowledge assets or start fresh with structured capture. Most teams inherit a graveyard of transformation artifacts: half-transcribed stakeholder interviews, strategy decks from two leadership regimes ago, email threads that contain critical context buried in 47 replies. The instinct is to preserve everything. The smarter instinct is to be selective.
This piece provides a practical decision framework for HR IT and AI architects evaluating legacy knowledge migration, including a scoring matrix, cost realities, and a selective retrofit protocol for the artifacts worth saving.
Why the default answer is usually "start fresh"
Before walking through the framework, it helps to understand why fresh structured capture dominates retrofit in most HR transformation scenarios:
Legacy artifacts decay faster than you think. A stakeholder interview from 18 months ago reflects an org chart, a strategic context, and a set of leadership priorities that may no longer exist. The facts may still be accurate; the framing almost certainly is not.
Unstructured artifacts resist structuring. A 40-slide deck contains assertions without evidence trails. An email thread contains decisions without decision criteria. Converting these into a structured knowledge store requires reconstruction, not transcription, and reconstruction introduces interpretation bias.
The economics have shifted. An interview with a human researcher historically would have taken months to prepare, source, and deliver findings and at a tremendous cost. With AI-augmented capture methodology, that cost drops by an order of magnitude. At that price point, fresh structured capture becomes economically dominant over expensive retrospective structuring in almost every scenario.
Diagnostic question: "What percentage of our existing transformation artifacts could a new team member use to make a decision, without needing to call the person who created them?" A major question we ask clients before ISD engagement, and the answer is almost always under 10%.
Legacy knowledge assessment: a three-factor decision matrix

Not all legacy artifacts are equal. Score each artifact set on three dimensions, using a 1 to 5 scale:
Factor 1: Documentation quality (1 to 5)
This measures structural integrity, not content value. A fully transcribed, timestamped interview with clear speaker attribution scores a 5. A set of handwritten notes from a workshop scores a 1. The key question: can this artifact be processed programmatically, or does it require human interpretation to extract meaning?
Factor 2: Strategic value (1 to 5)
This measures relevance to current transformation priorities. A vendor evaluation that directly informs your upcoming technology selection scores high. A culture assessment from a pre-merger organization scores low. Be ruthless here; strategic value is perishable.
Factor 3: Recency (1 to 5)
Score 5 for artifacts less than 6 months old, 4 for 6 to 12 months, 3 for 12 to 18 months, 2 for 18 to 24 months, 1 for anything older. HR organizations evolve quickly. A people analytics maturity assessment from 2022 reflects a different team, different tools, and likely a different CHRO.
Interpreting the composite score:
12 to 15 (rare): Full retrofit. These artifacts are recent, well-documented, and strategically relevant. Invest in structuring them into your knowledge architecture.
8 to 11: Selective extraction. Pull specific facts, decisions, and data points. Discard the narrative framing.
7 or below: Treat as sunk cost. Archive for compliance if needed, but do not invest effort in migration.
Worked example
Consider a partially transcribed vendor evaluation from 14 months ago. Documentation quality: the transcription covers only 60% of the sessions, with no speaker attribution (score: 2). Strategic value: the vendor shortlist overlaps with your current evaluation, and several technical findings remain relevant (score: 4). Recency: 14 months places it in the 12 to 18 month band (score: 3). Composite: 9. This lands in selective extraction territory. You would pull the technical evaluation criteria and specific vendor findings, but discard the strategic rationale and recommendation narrative, which reflect a different leadership context.
The selective extraction protocol

For artifacts scoring 8 to 11, follow this four-step process:
Step 1: Identify extractable fact types. Not all information ages equally. Quantitative findings (cost data, headcount figures, system performance metrics) retain value longer than qualitative interpretations. Decision records ("we chose X because of Y") retain value longer than sentiment ("stakeholders felt positive about the direction").
Step 2: Extract and tag. Pull discrete facts from the source material and tag each with a confidence level (verified, plausible, requires revalidation) and an expiration indicator. For artifact sets that include recorded sessions or digital conversation exports, AI-assisted extraction can accelerate this step significantly. The human-in-the-loop checkpoint occurs at the tagging stage, where an analyst validates extraction accuracy and assigns confidence levels.
Step 3: Revalidate with current stakeholders. Any extracted fact with a "requires revalidation" tag gets a 60-second confirmation in your next structured interview. This is lightweight and compounds the value of new capture rather than creating a separate workstream.
Step 4: Integrate into the knowledge store. Validated extractions enter the same structured knowledge architecture as fresh capture. At Ikona Analytics, this means processing through a five-layer architecture: raw, transcribed, cleaned, fact-extracted, and anonymized. The result is a durable, queryable knowledge asset where legacy extractions and fresh interviews coexist in a single structured knowledge store.
Diagnostic question: "If we invested 40 hours in structuring legacy artifacts, would the resulting knowledge store be more or less valuable than 20 fresh structured interviews covering the same territory?" For most HR transformation initiatives, the math favors fresh capture decisively.
Preventing the next graveyard: knowledge asset governance
Solving the legacy migration problem without addressing governance just resets the clock. Three practices prevent your 2025 knowledge store from becoming the 2027 graveyard:
Ownership. Assign a named owner for the knowledge store, typically someone in the HR technology or people analytics function. This person does not create all content; they ensure it meets structural standards and remains queryable.
Deprecation triggers. Every knowledge asset gets a review date. Org-level findings expire at 12 months. Process documentation expires at 18 months. Technical architecture findings expire at 6 months. Expired assets move to an archive tier, not deletion, but they stop appearing in active queries.
Refresh integration. Build knowledge store maintenance into existing operating rhythms. Quarterly business reviews, annual planning cycles, and technology selection processes all generate new organizational intelligence. Capture it in the same structured format from the start.
Frequently asked questions
Should I pipe legacy Zoom recordings through an LLM for extraction?
You can, and for high-scoring artifacts it may be worthwhile. But recognize the limitation: automated extraction produces text, not structured knowledge. You still need the tagging, confidence scoring, and revalidation steps. LLM extraction reduces Step 2 labor; it does not eliminate Steps 3 and 4.
What format should the knowledge store take?
The specific tooling matters less than the structural discipline. Whether you use a graph database, a relational schema, or a well-tagged document store, the critical requirement is queryability: the ability to retrieve specific findings by topic, stakeholder, date, and confidence level without reading entire documents.
How do I handle politically sensitive legacy artifacts?
This is a real constraint the scoring matrix does not fully capture. Artifacts reflecting prior leadership decisions or contentious reorganizations require careful handling. Our recommendation: extract verifiable facts (timelines, costs, system configurations) and discard interpretive framing entirely. Fresh structured interviews with current stakeholders will surface the politically navigable version of the same organizational history.
The real cost is indecision
The most expensive outcome in legacy knowledge migration is not choosing wrong. It is not choosing at all. Teams that defer the rebuild-versus-retrofit decision end up doing both poorly: spending weeks structuring artifacts that will never inform a decision, while delaying the fresh capture that would.
The framework here is deliberately simple because the decision should be fast. Score your legacy artifacts honestly. Extract what scores high enough to warrant the effort. Let the rest go. And then invest your energy where the return is highest: structured, current, queryable knowledge captured with the methodological rigor to remain useful 12 months from now.
HR transformation has no shortage of complexity. Your knowledge architecture does not need to add to it.
If you are deciding what to carry forward and what to rebuild before a major HR transformation, that is exactly where an ISD engagement starts.
We'd welcome the conversation.
Written by
Richard Rosenow
Richard Rosenow is a founding partner at Ikona Analytics, bringing deep expertise in workforce intelligence, diagnostic methodology, and HR technology transformation from experience across Fortune 100 organizations.
Learn more about our team