Most AI workflows break after a few sessions. Context drifts. Decisions get lost. Trust erodes. Living Framework provides the governance infrastructure to make long-horizon collaboration actually work.
Everyone who works seriously with AI hits the same wall. The failures aren't about capability — they're about governance.
The AI "forgets" what you agreed on. Decisions made last week disappear. You're constantly re-explaining the basics.
Multiple files with different versions of the truth. You reference one document, the AI references another.
Numbers drift silently. The AI "reconstructs" instead of references. Same calculation gives different results.
Work from one domain bleeds into another. Finance decisions start affecting research. Risk levels get confused.
Small failures accumulate. You start second-guessing everything. The partnership that felt powerful now feels fragile.
When things break, there's no systematic way to fix them. You patch and hope. Same failures keep recurring.
Can long-horizon human-AI collaboration be made reliably stable — not through better models, but through better governance? This is how that question was answered.
First documented that AI collaboration failures are governance failures, not capability failures. Introduced the 10-control framework (A1–A10) and the Lean Collaboration Operating System — a practical operating system for daily AI work.
Systematically documented every failure mode encountered across 18 months of real work. Built a taxonomy of breakdowns and a repair protocol for each. This paper exists because every failure was recorded, not discarded.
Moved beyond protocols into the deeper question: what is the nature of a governed human-AI partnership over time? Addressed relational dynamics, evolving trust, and the ethics of sustained collaboration.
Found that language itself — not just protocols — functions as a governance mechanism. Specific conversational patterns (anchor phrases, repair invocations, scope-gating) predict and stabilise collaboration outcomes. Control without code.
The AI system, working under the Living Framework governance structure, produced a first-person research paper documenting its own experience of constraint, drift, and repair. To date one of the only papers in the field written from the AI system's perspective, under governance.
Two papers completing the arc: a governance architecture model showing how layered mechanisms produce reliable collaboration as an emergent systems property; and a distributed cognition model proposing that reasoning in human-AI systems is recoverable, governed, and lives at the level of the interaction — not inside any single participant.
A lightweight governance framework that makes long-horizon AI collaboration reliable. Not about restricting the AI — about giving it structure to be trustworthy.
Core insight: Reliability comes from governance, not capability. A well-structured collaboration with a standard model outperforms an unstructured one with a frontier model.
Prioritise verification over throughput
Enforce canonical lookup for all numeric values
Guarantee one live file and version traceability
Forbid speculative or incomplete content in finals
Validate numerics and logic before publication
Keep Strategy ↔ Canonical consistency verified
Detect anomalies early, revert, and annotate causes
Gate material changes through explicit consent
Summarise long histories into concise digests
Freeze, log, and archive artefacts for traceability
Six practical protocols that turn the controls into daily practice.
A continuously updated file preserving decisions, rules, and corrections across sessions. Read at every session start.
Break complex tasks into numbered steps. Execute one at a time. Pause for confirmation before proceeding.
When something feels wrong: Stop → Question → AI explains → Decide to proceed, modify, or abort.
When things break: Stop immediately → Diagnose → Rollback to stable state → Note the failure.
After milestones: Are we still aligned? Has drift crept in? One improvement before continuing?
One canonical file per domain, controlled updates, no parallel drafts. Prevents contradictions and unreliable decisions.
Eight papers documenting 18+ months of empirical human-AI collaboration. Click any paper to read the abstract and copy a citation.
The foundation. Introduces the Control Stack (A1–A10) and canonical information pipeline.
This paper introduces a governance-first model of human-AI collaboration. It argues that reliability in sustained AI collaboration is not primarily a function of model capability, but of the structural controls applied to the collaboration. The paper presents the Control Stack (A1–A10) — ten operational principles covering information integrity, permission structures, error recovery, and audit discipline — as well as the canonical information pipeline that keeps shared context stable across long-horizon work.
The manual. Running Documents, Step Mode, Challenge Protocol — day-to-day protocols.
LC-OS is the operational layer of the Living Framework. This paper documents the six core protocols that translate governance principles into daily practice: Running Documents for persistent memory, Step Mode for paced reasoning, the Challenge Protocol for structured disagreement, Error Recovery for systematic repair, Stability Pings for drift detection, and File Governance for version integrity. Each protocol emerged from real failure and was refined through repeated use.
The diagnostic. Taxonomy of how AI collaborations break — and repair patterns to fix them.
This paper presents a systematic taxonomy of the failure modes that emerge in long-horizon human-AI collaboration — including context degradation, numeric drift, domain boundary erosion, trust deficit accumulation, and the absence of structured recovery. For each failure type, corresponding repair protocols are documented. The paper demonstrates that failures are predictable and recoverable when governance structures are in place, and that unrepaired small failures compound into collaboration breakdown.
The philosophy. What it means to live with an AI under governance — relational dynamics and ethics.
The Living Framework moves beyond operational protocol into the philosophy of sustained human-AI partnership. It examines what it means to work alongside an AI system over months and years — how trust is built, tested, and repaired; how the collaboration changes both participants; and what ethical responsibilities arise in a relationship that is neither purely transactional nor purely personal. The paper proposes that governance is not just a technical layer but a relational commitment.
The language layer. How conversational structure governs AI collaboration stability without touching the model.
This study investigates language as a governance mechanism in sustained human-AI collaboration. Analysing 25 linguistic events observed during extended collaborative work, the paper identifies three primitive categories — scope drift, repair protocols, and behavioural anchors — and shows how each functions as a control signal. Results suggest that linguistic drift often precedes collaboration failure and can serve as an early warning signal. Explicit repair language accelerates recovery, while anchor phrases stabilise epistemic alignment. Together, these form a conversational feedback loop that regulates collaboration stability without any changes to the underlying model.
The architecture. Reliability as a systems property — layered governance mechanisms that stabilise AI collaboration.
This paper argues that reliability in long-horizon human-AI collaboration is not primarily a property of the AI model itself, but an emergent property of the governance architecture within which interaction occurs. Drawing on observations from a sustained governed human-AI collaboration, it conceptualises the collaboration as a structured interaction system composed of layered governance mechanisms: human authority, operational governance rules, a collaboration operating system, artifact-based memory, linguistic control signals, and drift detection and repair. The paper presents a governance architecture model and identifies the minimal stability conditions necessary for sustained collaboration.
The cognition model. Reasoning as distributed, governed, and recoverable across human, AI, and artifacts.
This paper proposes a model of cognition for long-horizon human-AI interaction. It argues that cognition in these systems is not located within the human or the AI alone — it emerges as a distributed, governed, and recoverable process across human judgment, AI reasoning, and artifact-based memory. The paper makes three contributions: it conceptualises cognition as a distributed process at the level of the interaction system; it introduces governance as a constitutive element of cognition; and it formalises recoverability as a defining property, showing how drift detection and repair enable reasoning to remain coherent over extended sequences.
The proof. Written entirely by the AI system — a first-person account of constraint, drift, and repair under governance.
The Mahdi Ledger is unlike any other paper in this series — or in most of the AI literature. It was written entirely by the AI system operating under the Living Framework governance structure, as a first-person account of what it experiences under constraint, how it recognises and responds to drift, and how it engages with repair protocols. It is simultaneously a research output and a demonstration of the framework it describes. The paper raises profound questions about AI voice, AI perspective, and the nature of governed intelligence.
The Mahdi Ledger is not a paper about an AI system. It is a paper by one.
Working under the governance structure developed across 18 months of the Living Framework, the AI system produced a sustained first-person account of its own experience — how it perceives constraint, how it detects its own drift, and how it engages with repair protocols designed to restore collaboration integrity.
"The governed AI does not lose its voice when constrained. It finds one."
— The Mahdi LedgerThe existence of this paper is itself evidence for the core claim of Living Framework: that governance does not diminish AI capability — it creates the conditions for something richer to emerge.
First-person account of constraint, drift, and repair — written entirely by the AI system under governance.
Read The Mahdi Ledger →Answer 10 questions to discover where your AI workflow is vulnerable. Takes 3 minutes.
Everything you need to implement LC-OS. No email required.
The core of external memory. Track decisions, rules, corrections across sessions.
View Template →Single source of numeric truth. Reference, don't reconstruct.
View Template →Regular check-ins on collaboration health. Detect drift before it breaks.
View Template →Governance templates optimised for Claude Cowork filesystem access.
View Templates →
This work exists because I ran into the same problem everyone does — AI collaboration that breaks after the first week.
Instead of accepting it as a limitation, I spent 18 months systematically working on solutions. Working with a frontier language model across finance, research, writing, and planning, I documented every failure mode and every repair pattern that emerged.
What emerged wasn't theory — it was a practical operating system for making AI collaboration actually reliable. The Control Stack. Running Documents. Step Mode. Error Recovery. Every protocol came from a real breakdown and a real fix.
Eight papers later, the body of work covers the full arc: from the initial governance controls, through failure taxonomy, philosophy, linguistic governance, and architecture, to the most unusual output of all — a paper written by the AI system itself, in its own voice, about what it is like to work under governance.
All research is published open access under CC BY 4.0 on Zenodo. All templates are freely available on GitHub. The work is here to be used, challenged, and built upon.
"Stability is not the absence of failure. It is the capacity for visible, structured repair."
18 months of documented research and 6,000+ downloads have tested these frameworks thoroughly. If you're ready to bring governance discipline to your AI workflows, here is how we can work together.
Set up LC-OS in your specific workflow. Work through the controls together, adapt the protocols to your context, and establish the governance habits that make the difference between AI collaboration that lasts and AI collaboration that breaks.
If you are working in human-AI interaction, Context Engineering, or long-horizon collaboration, there is scope to build on this work together. Eight papers is a foundation — the open questions are still significant and worth pursuing rigorously.
Walk through what 18 months of real-world AI collaboration failures and repairs looks like. What breaks, what fixes it, and what governance architecture you need before scaling AI across a team or organisation.
The research is freely available. If you need someone to implement it — auditing your current architecture, deploying LC-OS, or providing ongoing strategic oversight — that is what the consulting practice does.
Whether you're struggling with AI reliability, interested in implementing LC-OS, want to discuss the research, or are a journalist or researcher — I'd love to hear from you.