I’ve seen configuration managed on paper, with change notices stacked on drawings like a to-do pile. I’ve seen teams lose the ability to rebuild the same product in another country because they couldn’t produce a complete BOM. The problem wasn’t “conservatism.” The problem was a broken evidence supply chain. Certification-ready digital threads fix that. Here’s what changed when we rebuilt that supply chain: Audit pack assembly: 2 months → 2 weeks Evidence control: Google Drive (no traceability, low access control) → controlled, end-to-end traceability + access control Audit pack assembly: 2 months → 2 weeks Audit pack assembly: 2 months → 2 weeks Evidence control: Google Drive (no traceability, low access control) → controlled, end-to-end traceability + access control Evidence control: Google Drive (no traceability, low access control) → controlled, end-to-end traceability + access control This wasn’t “we worked harder.” It was architectural. Paper vs. digital is the wrong fight In regulated deep-tech — nuclear, aerospace, defense, medical, advanced robotics — people argue about paper vs. 3D models, spreadsheets vs. metadata report, PDFs vs. PLM, “old-school authorities” vs. modern tools. That argument misses the point. Regulators don’t want paper. They want proof with properties that survive scrutiny: proof tied to a baseline (a frozen configuration) versioned and reproducible traceable end-to-end (requirements → design → verification → approvals) protected by access control and accountability supported by change impact analysis (so you can prove what changed and what didn’t) tied to a baseline (a frozen configuration) baseline versioned and reproducible versioned traceable end-to-end (requirements → design → verification → approvals) traceable end-to-end protected by access control and accountability access control supported by change impact analysis (so you can prove what changed and what didn’t) change impact analysis Paper often provides these properties by accident. Digital systems must provide them by design. Toolchain trust: the question behind “what tools do you use?” When auditors ask about your tools, they’re not shopping for brands. They’re drawing your trust boundary. trust boundary They want to know: Which systems are evidence-grade (system of record)? Which are supporting (convenient, but not trusted as proof)? How do you ensure outputs are reproducible (versions, baselines, controlled exports)? Who can change what — and where is the audit trail? Which systems are evidence-grade (system of record)? evidence-grade Which are supporting (convenient, but not trusted as proof)? supporting How do you ensure outputs are reproducible (versions, baselines, controlled exports)? reproducible Who can change what — and where is the audit trail? audit trail If you can’t answer those cleanly, you’ll be pushed back to PDFs. And you’ll deserve it. Two industries, same root cause: broken proof supply chains I’ve built and defended evidence chains across two very different contexts: Nuclear power plant projects: configuration management across multiple programmes, with customers in different countries and customer requirements coming from different sources (in one case, ~27,000 requirements). Hydrogen-electric aircraft powertrain: a certification route developed with certification bodies, where proven proof patterns were non-negotiable. Nuclear power plant projects: configuration management across multiple programmes, with customers in different countries and customer requirements coming from different sources (in one case, ~27,000 requirements). Nuclear power plant projects Hydrogen-electric aircraft powertrain: a certification route developed with certification bodies, where proven proof patterns were non-negotiable. Hydrogen-electric aircraft powertrain Different domains. Same failure mode: engineering work existed, but evidence was fragmented, inconsistent, and slow to assemble. Certification is not a document contest. It’s an evidence supply chain. evidence supply chain The moment it became obvious In an earlier environment, the “audit pack” was paper-based. Changes were happening in the build environment faster than documentation could keep up. On some drawings, there were piles of 4–5 unimplemented change notices. 4–5 unimplemented change notices A TechOps lead summed it up: “If I’m asked to do this again in the same style, I’m quitting.” Later, when the same product needed to be assembled outside the original site, the organization struggled — not because engineering was wrong, but because it couldn’t reliably produce a complete, trustworthy BOM across stakeholders. That’s the whole problem in one line: If you can’t reproduce proof, you don’t have control. If you can’t reproduce proof, you don’t have control. Digitalization is step 3, not step 1 Most “digital transformations” fail because they start with tools. The sequence that works is boring: Establish the process (what must be true, who approves what, what “done” means) Optimize the process (remove invented requirements and legacy rituals) Digitize (encode the optimized process into enforceable workflows) Establish the process (what must be true, who approves what, what “done” means) Establish the process Optimize the process (remove invented requirements and legacy rituals) Optimize the process Digitize (encode the optimized process into enforceable workflows) Digitize Skip #1 and #2 and you get expensive chaos — now with dashboards. The questions regulators keep asking (and what they really mean) Across industries, certification bodies come back to the same three questions: 1) “Which tools do you use, and why should we trust their outputs?” Translation: show the trust boundary, reproducibility, and audit trails. 2) “Do documented procedures match what’s implemented in the system? How is change managed?” Translation: does your process live in slides, or in enforced workflows with logs? 3) “Show traceability live in the system.” Translation: not a PDF export. A real trace from requirement to verified evidence, with access control. If you can answer these on demand, audits stop being a recurring emergency. The missing architecture: the Certification Interface Layer Engineering teams and regulators speak different languages. If you don’t build a translation layer, you translate manually during audit prep — at the worst possible time. I call this layer the Certification Interface Layer: Certification Interface Layer Map regulator questions → system objects + proof paths, not documents. Map regulator questions → system objects + proof paths, not documents. Proof pipeline Once this exists, “show me” becomes a query and a snapshot — not a two-month scramble. What changed for us (measurable outcomes) Over a two-year transition (Jan 2024 → Sep 2025), we rebuilt the evidence pipeline end-to-end. This work supported scaling our compliance posture and helped underpin a DOA journey (organizational capability recognised under Part 21 Subpart J frameworks). (easa.europa.eu) Jan 2024 → Sep 2025 easa.europa.eu Results: Audit pack assembly: 2 months → 2 weeks Evidence control: drive chaos → controlled access + full traceability Fragmentation reduction: 4 PDM systems → 1 PLM with a standardized approach Change control maturity: manual impact analysis → automated impact analysis Adoption scale: ~30 users → ~150 users Audit pack assembly: 2 months → 2 weeks Audit pack assembly: 2 months → 2 weeks Evidence control: drive chaos → controlled access + full traceability Evidence control: drive chaos → controlled access + full traceability Fragmentation reduction: 4 PDM systems → 1 PLM with a standardized approach Fragmentation reduction: 4 PDM systems → 1 PLM with a standardized approach Change control maturity: manual impact analysis → automated impact analysis Change control maturity: manual impact analysis → automated impact analysis Adoption scale: ~30 users → ~150 users Adoption scale: ~30 users → ~150 users No magic. Just proof as a system. The playbook: 7 steps to certification-ready digital threads Step 1) Define Units of Compliance Stop treating all requirements equally. Define what you certify: per requirement? per function? per subsystem? Output: compliance-unit catalog + definition of “proven.” Output: Step 2) Map Regulator Questions to Data Objects Build a one-page matrix: question → object → source of truth → owner → proof rule Output: Authority Q → Object map. Output: Step 3) Build the Evidence Graph (not a document tree) Documents are containers. Evidence is a graph: Requirement → Design → Implementation → Verification → Result → Approval Make key links enforced. Optional links disappear under pressure. Output: graph rules + completeness checks. Output: Step 4) Baselines and Snapshots Authorities certify a baseline, not your “current database.” You need repeatable baselining: freeze configuration capture linked evidence objects preserve versions, approvals, access control freeze configuration capture linked evidence objects preserve versions, approvals, access control Output: baseline freeze procedure + snapshot bundle definition. Output: Step 5) Change Impact Analysis Change management isn’t “we have change requests.” It’s: what becomes invalid and what must be re-verified when something changes? Define computable impact rules by change type: requirement change → which verification artifacts invalidate? BOM/part change → what downstream artifacts must update? software change → what verification is affected? requirement change → which verification artifacts invalidate? BOM/part change → what downstream artifacts must update? software change → what verification is affected? Output: impact rules + automated flags. Output: Step 6) Audit Pack Assembly Pipeline (treat it like a release) If audit packs are assembled manually, they will be wrong. Treat assembly like a release pipeline: select baseline → generate pack → review → sign → export every pack has an ID every pack is reproducible select baseline → generate pack → review → sign → export every pack has an ID every pack is reproducible Output: assembly pipeline + definition of done. Output: Step 7) Governance and Tool Boundaries Most “digital threads” fail here: no one defines decision rights. You need explicit answers to: which system is the source of truth for what? who can change baselines? what counts as approval/signature? what is evidence-grade vs supporting? which system is the source of truth for what? who can change baselines? what counts as approval/signature? what is evidence-grade vs supporting? Output: RACI + decision rights + signature model. Output: A 5-minute live traceability demo script If you can’t demo this live, you don’t have a digital thread. Open one compliance unit Show linked design artifact + rationale Show linked implementation item (BOM/config/software) Show linked verification activity and results Show approvals (who/when/under workflow) Create a change → show impact analysis outputs Generate a baseline snapshot and export the audit pack Open one compliance unit Show linked design artifact + rationale Show linked implementation item (BOM/config/software) Show linked verification activity and results Show approvals (who/when/under workflow) Create a change → show impact analysis outputs Generate a baseline snapshot and export the audit pack Auditors don’t need your entire tool landscape. They need reproducible proof. Traceability Why this applies beyond aerospace and nuclear (including humanoid robots and MedTech) Swap “airworthiness” for “safety case” and the structure stays the same: hazards ↔ mitigations ↔ requirements ↔ tests ↔ approvals baselines + access control change impact rules reproducible evidence exports toolchain trust boundaries hazards ↔ mitigations ↔ requirements ↔ tests ↔ approvals baselines + access control change impact rules reproducible evidence exports toolchain trust boundaries Different tech. Same proof problem. The decisive role: the translator One more uncomfortable truth: the hardest part isn’t integration — it’s requirements hygiene. We didn’t just ask teams how they work. We asked why. Most “requirements” turned out to be habits. why Using first principles, we reduced requirements to: standards and obligations that actually govern the function core requirements derived from those sources removal of invented constraints, with user alignment standards and obligations that actually govern the function core requirements derived from those sources removal of invented constraints, with user alignment Only after that did digitalization become scalable. What you can do in 30 days Create the Authority Q → Object map (one page, no tools required) Define compliance units and traceability completeness criteria Implement baseline snapshots for one subsystem Run the 5-minute demo weekly internally Then automate audit pack assembly Create the Authority Q → Object map (one page, no tools required) Define compliance units and traceability completeness criteria Implement baseline snapshots for one subsystem Run the 5-minute demo weekly internally Then automate audit pack assembly