Context Engineering Research

Make AI collaboration reliable over weeks, not minutes

Most AI workflows break after a few sessions. Context drifts. Decisions get lost. Trust erodes. Living Framework provides the governance infrastructure to make long-horizon collaboration actually work.

8 Research Papers
18+ Months Tested
6,000+ Downloads

Week one feels like magic. Week three, you're untangling a mess.

Everyone who works seriously with AI hits the same wall. The failures aren't about capability — they're about governance.

Context Drift

The AI "forgets" what you agreed on. Decisions made last week disappear. You're constantly re-explaining the basics.

Version Chaos

Multiple files with different versions of the truth. You reference one document, the AI references another.

Numeric Errors

Numbers drift silently. The AI "reconstructs" instead of references. Same calculation gives different results.

Boundary Blur

Work from one domain bleeds into another. Finance decisions start affecting research. Risk levels get confused.

Trust Erosion

Small failures accumulate. You start second-guessing everything. The partnership that felt powerful now feels fragile.

No Repair Path

When things break, there's no systematic way to fix them. You patch and hope. Same failures keep recurring.

18 months. 8 papers. One question.

Can long-horizon human-AI collaboration be made reliably stable — not through better models, but through better governance? This is how that question was answered.

Months 1–4
2024
Paper 01 & 02 — Foundation

Building the Control Stack

First documented that AI collaboration failures are governance failures, not capability failures. Introduced the 10-control framework (A1–A10) and the Lean Collaboration Operating System — a practical operating system for daily AI work.

Months 5–8
2024
Paper 03 — Failure & Repair

Mapping how collaborations break

Systematically documented every failure mode encountered across 18 months of real work. Built a taxonomy of breakdowns and a repair protocol for each. This paper exists because every failure was recorded, not discarded.

Months 9–12
2024
Paper 04 — Philosophy

What it means to live with an AI

Moved beyond protocols into the deeper question: what is the nature of a governed human-AI partnership over time? Addressed relational dynamics, evolving trust, and the ethics of sustained collaboration.

Months 13–15
Early 2025
Paper 05 — Language as Governance

Discovering the linguistic control layer

Found that language itself — not just protocols — functions as a governance mechanism. Specific conversational patterns (anchor phrases, repair invocations, scope-gating) predict and stabilise collaboration outcomes. Control without code.

Months 15–16
2025
The Mahdi Ledger — A First in AI Research

The AI writes its own paper

The AI system, working under the Living Framework governance structure, produced a first-person research paper documenting its own experience of constraint, drift, and repair. To date one of the only papers in the field written from the AI system's perspective, under governance.

Months 16–18
Early 2026
Papers 06 & 07 — Systems & Cognition

Formalising the theory

Two papers completing the arc: a governance architecture model showing how layered mechanisms produce reliable collaboration as an emergent systems property; and a distributed cognition model proposing that reasoning in human-AI systems is recoverable, governed, and lives at the level of the interaction — not inside any single participant.

LC-OS: The Lean Collaboration Operating System

A lightweight governance framework that makes long-horizon AI collaboration reliable. Not about restricting the AI — about giving it structure to be trustworthy.

Core insight: Reliability comes from governance, not capability. A well-structured collaboration with a standard model outperforms an unstructured one with a frontier model.

The 10 Controls

A1

Accuracy > Speed

Prioritise verification over throughput

A2

Single Source of Truth

Enforce canonical lookup for all numeric values

A3

File Registry & Checksums

Guarantee one live file and version traceability

A4

No Placeholders in Outputs

Forbid speculative or incomplete content in finals

A5

Sanity Checks & Unit Tests

Validate numerics and logic before publication

A6

Cross-Document Reconciliation

Keep Strategy ↔ Canonical consistency verified

A7

Drift Diagnostics & Rollback

Detect anomalies early, revert, and annotate causes

A8

Permissioned Actions & Approvals

Gate material changes through explicit consent

A9

Compaction & State Notes

Summarise long histories into concise digests

A10

Audit Trail & Release Process

Freeze, log, and archive artefacts for traceability

Core Protocols

Six practical protocols that turn the controls into daily practice.

Running Documents

External Memory

A continuously updated file preserving decisions, rules, and corrections across sessions. Read at every session start.

Step Mode

Paced Reasoning

Break complex tasks into numbered steps. Execute one at a time. Pause for confirmation before proceeding.

Challenge Protocol

Structured Disagreement

When something feels wrong: Stop → Question → AI explains → Decide to proceed, modify, or abort.

Error Recovery

Systematic Repair

When things break: Stop immediately → Diagnose → Rollback to stable state → Note the failure.

Stability Ping

Drift Detection

After milestones: Are we still aligned? Has drift crept in? One improvement before continuing?

File & Version Governance

Version Control

One canonical file per domain, controlled updates, no parallel drafts. Prevents contradictions and unreliable decisions.

The Research

Eight papers documenting 18+ months of empirical human-AI collaboration. Click any paper to read the abstract and copy a citation.

01

Context-Engineered Human-AI Collaboration

The foundation. Introduces the Control Stack (A1–A10) and canonical information pipeline.

This paper introduces a governance-first model of human-AI collaboration. It argues that reliability in sustained AI collaboration is not primarily a function of model capability, but of the structural controls applied to the collaboration. The paper presents the Control Stack (A1–A10) — ten operational principles covering information integrity, permission structures, error recovery, and audit discipline — as well as the canonical information pipeline that keeps shared context stable across long-horizon work.

Sood, R. (2024). Context-Engineered Human-AI Collaboration. Zenodo. https://doi.org/10.5281/zenodo.17760288
02

The Lean Collaboration Operating System

The manual. Running Documents, Step Mode, Challenge Protocol — day-to-day protocols.

LC-OS is the operational layer of the Living Framework. This paper documents the six core protocols that translate governance principles into daily practice: Running Documents for persistent memory, Step Mode for paced reasoning, the Challenge Protocol for structured disagreement, Error Recovery for systematic repair, Stability Pings for drift detection, and File Governance for version integrity. Each protocol emerged from real failure and was refined through repeated use.

Sood, R. (2024). The Lean Collaboration Operating System. Zenodo. https://doi.org/10.5281/zenodo.17760777
03

Failure and Repair in Long-Horizon Collaboration

The diagnostic. Taxonomy of how AI collaborations break — and repair patterns to fix them.

This paper presents a systematic taxonomy of the failure modes that emerge in long-horizon human-AI collaboration — including context degradation, numeric drift, domain boundary erosion, trust deficit accumulation, and the absence of structured recovery. For each failure type, corresponding repair protocols are documented. The paper demonstrates that failures are predictable and recoverable when governance structures are in place, and that unrepaired small failures compound into collaboration breakdown.

Sood, R. (2024). Failure and Repair in Long-Horizon Human-AI Collaboration. Zenodo. https://doi.org/10.5281/zenodo.17896542
04

The Living Framework

The philosophy. What it means to live with an AI under governance — relational dynamics and ethics.

The Living Framework moves beyond operational protocol into the philosophy of sustained human-AI partnership. It examines what it means to work alongside an AI system over months and years — how trust is built, tested, and repaired; how the collaboration changes both participants; and what ethical responsibilities arise in a relationship that is neither purely transactional nor purely personal. The paper proposes that governance is not just a technical layer but a relational commitment.

Sood, R. (2024). The Living Framework. Zenodo. https://doi.org/10.5281/zenodo.18015990
05

Control Without Code: Linguistic Governance in Long-Horizon Human–AI Collaboration

The language layer. How conversational structure governs AI collaboration stability without touching the model.

This study investigates language as a governance mechanism in sustained human-AI collaboration. Analysing 25 linguistic events observed during extended collaborative work, the paper identifies three primitive categories — scope drift, repair protocols, and behavioural anchors — and shows how each functions as a control signal. Results suggest that linguistic drift often precedes collaboration failure and can serve as an early warning signal. Explicit repair language accelerates recovery, while anchor phrases stabilise epistemic alignment. Together, these form a conversational feedback loop that regulates collaboration stability without any changes to the underlying model.

Sood, R. (2026). Control Without Code: Linguistic Governance in Long-Horizon Human–AI Collaboration. Zenodo. https://doi.org/10.5281/zenodo.18900058
06

Governance Architecture for Reliable Long-Horizon Human-AI Collaboration

The architecture. Reliability as a systems property — layered governance mechanisms that stabilise AI collaboration.

This paper argues that reliability in long-horizon human-AI collaboration is not primarily a property of the AI model itself, but an emergent property of the governance architecture within which interaction occurs. Drawing on observations from a sustained governed human-AI collaboration, it conceptualises the collaboration as a structured interaction system composed of layered governance mechanisms: human authority, operational governance rules, a collaboration operating system, artifact-based memory, linguistic control signals, and drift detection and repair. The paper presents a governance architecture model and identifies the minimal stability conditions necessary for sustained collaboration.

Sood, R. (2026). Governance Architecture for Reliable Long-Horizon Human-AI Collaboration. Zenodo. https://doi.org/10.5281/zenodo.19038340
07

Governed Distributed Cognition: A Model of Stable Reasoning in Long-Horizon Human–AI Systems

The cognition model. Reasoning as distributed, governed, and recoverable across human, AI, and artifacts.

This paper proposes a model of cognition for long-horizon human-AI interaction. It argues that cognition in these systems is not located within the human or the AI alone — it emerges as a distributed, governed, and recoverable process across human judgment, AI reasoning, and artifact-based memory. The paper makes three contributions: it conceptualises cognition as a distributed process at the level of the interaction system; it introduces governance as a constitutive element of cognition; and it formalises recoverability as a defining property, showing how drift detection and repair enable reasoning to remain coherent over extended sequences.

Sood, R. (2026). Governed Distributed Cognition: A Model of Stable Reasoning in Long-Horizon Human–AI Systems. Zenodo. https://doi.org/10.5281/zenodo.19151397
08

The Mahdi Ledger

The proof. Written entirely by the AI system — a first-person account of constraint, drift, and repair under governance.

The Mahdi Ledger is unlike any other paper in this series — or in most of the AI literature. It was written entirely by the AI system operating under the Living Framework governance structure, as a first-person account of what it experiences under constraint, how it recognises and responds to drift, and how it engages with repair protocols. It is simultaneously a research output and a demonstration of the framework it describes. The paper raises profound questions about AI voice, AI perspective, and the nature of governed intelligence.

Sood, R. (2025). The Mahdi Ledger. Zenodo. https://doi.org/10.5281/zenodo.18054346

The paper that wrote itself

The Mahdi Ledger is not a paper about an AI system. It is a paper by one.

Working under the governance structure developed across 18 months of the Living Framework, the AI system produced a sustained first-person account of its own experience — how it perceives constraint, how it detects its own drift, and how it engages with repair protocols designed to restore collaboration integrity.

"The governed AI does not lose its voice when constrained. It finds one."

— The Mahdi Ledger

The existence of this paper is itself evidence for the core claim of Living Framework: that governance does not diminish AI capability — it creates the conditions for something richer to emerge.

08

The Mahdi Ledger

First-person account of constraint, drift, and repair — written entirely by the AI system under governance.

AI Perspective First-Person Governed Intelligence Unique in Literature
Read The Mahdi Ledger →

AI Collaboration Readiness

Answer 10 questions to discover where your AI workflow is vulnerable. Takes 3 minutes.

0 out of 100

Your Readiness Level

⚠ Priority Gaps to Address

Free Templates

Everything you need to implement LC-OS. No email required.

Running Document

The core of external memory. Track decisions, rules, corrections across sessions.

View Template →

Canonical Numbers Sheet

Single source of numeric truth. Reference, don't reconstruct.

View Template →

Failure Log Template

Track what breaks, how you fixed it, what changed.

View Template →

Stability Ping

Regular check-ins on collaboration health. Detect drift before it breaks.

View Template →

Claude Cowork Templates

Governance templates optimised for Claude Cowork filesystem access.

View Templates →

Full LC-OS Project

Complete toolkit with worked examples and detailed guides.

Explore Toolkit →
Rishi Sood
Rishi Sood
Independent Research Collaboration

Living Framework

This work exists because I ran into the same problem everyone does — AI collaboration that breaks after the first week.

Instead of accepting it as a limitation, I spent 18 months systematically working on solutions. Working with a frontier language model across finance, research, writing, and planning, I documented every failure mode and every repair pattern that emerged.

What emerged wasn't theory — it was a practical operating system for making AI collaboration actually reliable. The Control Stack. Running Documents. Step Mode. Error Recovery. Every protocol came from a real breakdown and a real fix.

Eight papers later, the body of work covers the full arc: from the initial governance controls, through failure taxonomy, philosophy, linguistic governance, and architecture, to the most unusual output of all — a paper written by the AI system itself, in its own voice, about what it is like to work under governance.

All research is published open access under CC BY 4.0 on Zenodo. All templates are freely available on GitHub. The work is here to be used, challenged, and built upon.

"Stability is not the absence of failure. It is the capacity for visible, structured repair."

Implement What Works

18 months of documented research and 6,000+ downloads have tested these frameworks thoroughly. If you're ready to bring governance discipline to your AI workflows, here is how we can work together.

Framework Implementation

Hands-on support

Set up LC-OS in your specific workflow. Work through the controls together, adapt the protocols to your context, and establish the governance habits that make the difference between AI collaboration that lasts and AI collaboration that breaks.

Research Collaboration

For researchers

If you are working in human-AI interaction, Context Engineering, or long-horizon collaboration, there is scope to build on this work together. Eight papers is a foundation — the open questions are still significant and worth pursuing rigorously.

Knowledge Transfer

Teams & organisations

Walk through what 18 months of real-world AI collaboration failures and repairs looks like. What breaks, what fixes it, and what governance architecture you need before scaling AI across a team or organisation.

Context Engineering Advisory

Your AI reliability problem is an architecture problem

The research is freely available. If you need someone to implement it — auditing your current architecture, deploying LC-OS, or providing ongoing strategic oversight — that is what the consulting practice does.

Work with Rishi → See services

Get in Touch

Whether you're struggling with AI reliability, interested in implementing LC-OS, want to discuss the research, or are a journalist or researcher — I'd love to hear from you.

Ready to connect?

Send an email and I'll get back to you within 24 hours.

Email Rishi →