Enterprise AI adoption has a blind spot. Organizations invest heavily in model selection, fine-tuning, and prompt engineering — but treat context as an afterthought. The knowledge that feeds AI systems is unversioned, ungoverned, and untraceable. When an AI agent gives a bad answer, no one can tell you which document it read, who wrote it, when it was last reviewed, or whether it was even approved for AI consumption.
This paper argues that context governance — not model capability — is the binding constraint on enterprise AI quality.
Abstract
Enterprise AI adoption has a blind spot. Organizations invest heavily in model selection, fine-tuning, and prompt engineering — but treat context as an afterthought. The knowledge that feeds AI systems is unversioned, ungoverned, and untraceable.
This paper argues that context governance — not model capability — is the binding constraint on enterprise AI quality. It introduces Context Nest, an open specification and reference implementation for structured, versioned, integrity-verified context for AI agents.
The core specification, engine, CLI, and MCP server are available now as open source. PromptOwl's commercial platform builds the governance, compliance, and collaboration layers on top.
1. The Problem: AI Quality Is a Context Problem
Every production AI system has the same architecture at its core:
Input → [Model + Context] → OutputModels are commoditizing. GPT-4, Claude, Gemini, Llama — they converge on capability quarter by quarter. The differentiation has shifted entirely to the right side of the plus sign: context.
Yet here's what “context management” looks like in most organizations today:
- Someone dumps PDFs into a vector database
- A RAG pipeline retrieves chunks based on semantic similarity
- The model generates an answer from whatever it found
- Nobody knows which chunks were used, who wrote them, or when they were last reviewed
This is the context governance gap — the space between “we have documents” and “our AI uses trusted, current, accountable knowledge.”
The Real Costs
Hallucination from stale context. A policy document was updated six months ago. The vector store still has the old version embedded. The AI confidently cites a policy that no longer exists. The legal team discovers this in production.
Accountability vacuum. An AI agent recommends a course of action to a customer. The recommendation is wrong. The post-mortem asks: What knowledge did the agent use? Who approved that knowledge? When was it last reviewed? Nobody can answer.
Shadow context. Individual teams maintain their own document collections for their own AI tools. No central visibility into what's being used where. No consistency. No governance. The same question returns different answers depending on which team's context the model draws from.
Compliance exposure. In regulated industries — finance, healthcare, legal — there's no audit trail for AI context. When the auditor asks “show me the provenance chain for this AI-generated output,” the answer is a shrug.
These aren't hypothetical. They're happening now, at scale, in organizations spending millions on AI.
Ready to stop guessing what your AI knows?
The open source specification, engine, CLI, and MCP server are available now.
Join the Alpha Waitlist2. Why RAG Isn't Enough
Retrieval-Augmented Generation solved the first problem: getting external knowledge into a model's context window. That was necessary. It wasn't sufficient.
RAG tells you nothing about:
| Question | RAG's Answer |
|---|---|
| Who wrote this content? | Unknown |
| When was it last reviewed? | Unknown |
| Is this the approved version? | No concept of versions |
| Who is responsible for keeping it current? | Nobody assigned |
| Should this model be allowed to see this document? | No permission model |
| What other documents does this one reference? | Embedding destroyed the structure |
| Can I trace this AI output back to its source? | Chunk IDs at best |
| Has someone tampered with this document since it was published? | No integrity verification |
RAG treats documents as opaque blobs to be chunked and embedded. It destroys the structure, relationships, and metadata that make knowledge trustworthy. The link between two documents? Gone. The author attribution? Gone. The “this was approved by the legal team on January 15th” status? Never captured.
Embeddings are lossy compression of knowledge. They're useful for retrieval. They're useless for governance.
The Vector Database Illusion
The market has convinced itself that a better vector database equals better AI. Faster retrieval. Better embeddings. Hybrid search. Re-ranking.
These are optimizations on a fundamentally ungoverned system. Retrieving the wrong document faster doesn't help. Retrieving an unapproved document with better semantic precision doesn't help. The problem isn't search quality. The problem is what you're searching over has no quality controls.
3. Context Nest: An Open Standard for Governed Context
What's missing is a governance layer between raw knowledge and AI consumption. We built one. Context Nest is an open specification and reference implementation that provides:
Authoring → Versioning → Integrity → Querying → Injection → TracingEach stage is a control point. The specification, engine, CLI, and MCP server are open source — available today on npm and GitHub. Here's what each layer does.
3.1 Structured, Connected Knowledge
Context is not a flat file. It's a connected network. Documents reference other documents. Concepts link to concepts. Live data sources connect to external systems.
Every document in a Context Nest is standard Markdown with YAML frontmatter. Open it in VS Code. Open it in Obsidian. Open it in any text editor. The content is always accessible. The metadata travels with it:
---
title: "API Design Guidelines"
type: document
tags:
- "#engineering"
- "#api"
status: published
version: 3
author: john.doe@example.com
checksum: "sha256:a1b2c3..."
---
# API Design Guidelines
All endpoints use REST conventions. See
[Architecture Overview](contextnest://nodes/architecture-overview)
for context.Documents link to each other using the contextnest:// URI scheme — stable, addressable references that survive moves, renames, and version changes. Backlinks are tracked automatically. Tags create a shared taxonomy. The raw Markdown is never mutated — editors render features via decoration, not transformation. Documents round-trip between any Markdown-compatible tool without loss.
The specification defines eight node types — document, snippet, glossary, persona, prompt, source, tool, and reference — each serving a distinct role in the knowledge graph. Source nodes are particularly powerful: they contain instructions for fetching live context from external services via MCP servers, REST APIs, or CLI tools. An agent resolving a source node doesn't read static content — it follows instructions to hydrate live data.
3.2 Hash-Chained Versioning
Every document version is cryptographically linked to its history. Not just “we saved a copy” — a SHA-256 hash chain that proves no version has been tampered with after the fact:
chain_hash[n] = SHA-256(chain_hash[n-1] : content_hash[n] : version[n] : author[n] : timestamp[n])The genesis entry uses a well-known sentinel (contextnest:genesis:v1). Any break in the chain — a silently edited version, a deleted entry, a backdated change — is detectable by recomputing hashes.
Version storage uses a keyframe + diff model: full snapshots at version 1 and every 10 versions, with unified diffs in between. Any historical version can be reconstructed. When an AI output is questioned, you trace it back: “The agent used Document X, version 4, published by Jane on February 1st.” Full provenance. No ambiguity.
3.3 Nest Checkpoints: Graph-Level Snapshots
Individual document versioning isn't enough. You need to know what the entire knowledge base looked like at a point in time.
A nest checkpoint is an atomic, immutable snapshot of the entire document graph — the Context Nest equivalent of a Git commit. Every time any document is published, a checkpoint is created. Each checkpoint records:
- Every published document and its version number at that instant
- Every document's
chain_hashat that version (cross-chain binding) - A
checkpoint_hashchaining this checkpoint to the previous one
This enables pinned resolution: contextnest://nodes/api-design@7 always resolves to the exact version of that document at checkpoint 7 — not “whatever's current,” but the precise content that existed at that moment. When an auditor asks “what did the AI know on March 1st?”, you reconstruct the checkpoint and show them exactly.
The checkpoint chain is independently verifiable. If context_history.yaml is corrupted, it can be deterministically rebuilt from per-document history files by replaying in chronological order.
3.4 Deterministic Query Language
RAG retrieval is probabilistic — “find documents that sound similar.” Context Nest selectors are deterministic — “give me exactly these documents, no surprises.”
The selector grammar is set-algebraic:
# All published engineering documents
ctx resolve '#engineering + status:published'
# API or architecture docs, excluding drafts
ctx resolve '(#api | #architecture) - status:draft'
# Everything in the onboarding pack
ctx resolve 'pack:onboarding.basics'
# Full-text search
ctx resolve 'contextnest://search/API design'Operators: + (AND), | (OR), - (NOT), () for grouping. Results are reproducible — the same query always returns the same documents. This is critical for compliance: you can prove what an agent was given, not guess at what a similarity search might have returned.
Context packs bundle selectors into named, curated collections with agent instructions:
id: onboarding.basics
label: "Onboarding Basics"
query: "#onboarding + type:document"
includes:
- "contextnest://nodes/architecture-overview"
agent_instructions: |
Present these documents in order.
Start with the architecture overview.3.5 Injection and Tracing
The final mile is getting governed context into the model's context window and maintaining the audit trail.
Agents request context by URI or selector query. The resolver returns only published content. Source nodes are returned in topological dependency order so agents can hydrate them correctly. Every access is logged:
| Trace Field | Value |
|---|---|
| Document | contextnest://nodes/api-design |
| Version | 4 |
| Checkpoint | 12 |
| Author | john.doe@example.com |
| Last edited | 2026-02-15T09:30:00Z |
When agents hydrate source nodes, the trace records what tools were called, which server was contacted, a hash of the result (not the content — proving what was seen without storing payloads), cache hit status, and duration. Complete provenance from question to answer.
3.6 MCP Integration: AI-Native Access
Context Nest ships with a Model Context Protocol server that exposes vault operations as tools for AI agents. Any MCP-compatible client — Claude Code, Claude Desktop, Cursor, or custom agents — can read, search, query, and mutate a vault through a standard interface:
{
"mcpServers": {
"contextnest": {
"command": "node",
"args": [
"node_modules/@promptowl/contextnest-mcp-server/dist/index.js"
],
"env": {
"CONTEXTNEST_VAULT_PATH": "/path/to/your/vault"
}
}
}
}The server exposes read tools (vault_info, resolve, read_document, search, verify_integrity) and mutation tools (create_document, update_document, delete_document, publish_document) — all auto-publishing and regenerating the context index. AI agents get structured, governed access to organizational knowledge without custom integrations.
4. Architecture Principles
4.1 The Format Is Not the Product
The most important design decision: the format is open. The governance layer is the product.
Context Nest documents are standard Markdown with YAML frontmatter. The specification is Apache-2.0 licensed. The reference engine is AGPL-3.0. Anyone can build tools that read and write Context Nest vaults. The ecosystem benefits from shared primitives — and adoption benefits PromptOwl.
Vendor lock-in is the number one objection from enterprise architects. A portable, open format eliminates that objection. If you stop using PromptOwl's commercial platform, your knowledge is still standard Markdown files in standard directories. No proprietary encoding. No data hostage.
4.2 Markdown-Native, Obsidian-Compatible
The specification supports two layout modes: a structured layout with dedicated nodes/, sources/, and packs/ directories, and a flat Obsidian-compatible layout where any .md file in the vault is a context node.
The raw markdown is never mutated. Wiki links, tags, mentions, and task checkboxes are rendered via decoration, not transformation. Documents round-trip between any editor without loss. Content is diffable in Git. No proprietary formatting traps content.
4.3 Extensible by Design
The specification is designed for extension via namespaced prefixes. Additional frontmatter fields (promptowl_feature: value), custom config files, custom selector syntax, custom URI patterns, additional transport types, and additional relationship edge types — all are supported without breaking compatibility.
This is how PromptOwl's commercial platform adds governance features on top of the open spec without forking it.
5. What's Available Now
The core platform is open source, shipped, and available on npm:
| Component | Package | License |
|---|---|---|
| Specification | context-nest-spec | Apache-2.0 |
| Engine | @promptowl/contextnest-engine | AGPL-3.0 |
| CLI | contextnest-cli | Apache-2.0 |
| MCP Server | @promptowl/contextnest-mcp-server | AGPL-3.0 |
# Get started in 30 seconds
npm install -g contextnest-cli
ctx init --name "My Knowledge Base"
ctx add nodes/first-doc --title "Getting Started"
ctx verifySource code: github.com/PromptOwl/context-nest
Why AGPL-3.0 for the Engine
The engine and MCP server are AGPL-3.0 — a strong copyleft license. This is intentional. If you use the engine in an open source project, the license is free. If you embed it in a proprietary product or offer it as a hosted service, you need a commercial license from PromptOwl. This protects the open source investment while ensuring the project is self-sustaining.
The CLI is Apache-2.0 — use it anywhere, no restrictions. The specification is Apache-2.0 — anyone can build their own implementation from the spec.
6. What PromptOwl Builds on Top
The open source layer is the foundation. PromptOwl's commercial platform adds the enterprise governance, collaboration, and compliance features that organizations need to deploy AI context at scale:
| Layer | Open Source | PromptOwl Commercial |
|---|---|---|
| Format & Storage | Context Nest spec, engine, CLI | — |
| AI Access | MCP server | — |
| Governance | — | Stewardship hierarchy, approval workflows, review queues |
| Permissions | — | Role-based access control, document/folder/tag-level policies |
| Collaboration | — | Real-time editing, team workspaces, commenting |
| Compliance | Hash chains (open) | Audit logging, analytics, SOC 2 reporting |
| Federation | URI scheme (open) | Registry API, cross-namespace resolution, connection pooling |
| Editor | — | Hootie Desktop (AI chat client), VS Code extension |
| Source Hydration | Basic resolution (open) | Caching, connection pooling, failure recovery, rate limiting |
The open source layer gives you versioned, integrity-verified, queryable context that any AI agent can access. The commercial layer gives you the people, process, and compliance controls that make it enterprise-ready.
7. The Case for Acting Now
Three converging trends make context governance urgent:
AI agents are going autonomous. As agents take actions — not just generating text — the quality of their context becomes safety-critical. An agent that emails a customer based on stale policy context isn't a quality issue. It's a liability issue.
Regulatory pressure is building. The EU AI Act requires transparency in AI decision-making. SOC 2 auditors are starting to ask about AI data provenance. Healthcare and financial services regulators want audit trails. “We use RAG” isn't going to satisfy these requirements. A hash-chained version history with checkpoint-level integrity verification will.
The model layer is commoditizing. When every vendor offers similar capability, the differentiator is what you feed the model. Organizations that govern their context well will outperform those that don't — systematically and measurably.
The organizations that build context governance now will have a structural advantage. Their AI will be more accurate, more auditable, and more trustworthy. The ones that wait will retrofit governance onto ungoverned systems — a harder, more expensive problem.
8. Alpha Program
We're looking for teams who share this conviction and want to shape the product.
Open Source — Available Now
The specification, engine, CLI, and MCP server are live. Install them, build a vault, point your AI agents at it. File issues, submit PRs, tell us what's missing.
- npm:
npm install -g contextnest-cli - GitHub: github.com/PromptOwl/context-nest
- Spec: github.com/PromptOwl/context-nest-spec
PromptOwl Governance Platform — Alpha
The commercial governance platform is in alpha with design partners. If you're deploying AI agents in production and your context layer keeps you up at night, we want to talk.
Ideal partners:
- AI-forward engineering or ops teams deploying agents in production
- Knowledge-heavy domains: legal, consulting, research, finance, healthcare
- 10-50 person teams — large enough for real governance needs, small enough to move fast
- Technical decision maker who can say yes without a six-month procurement cycle
What you get:
- Early access to the governance platform
- Direct influence on the roadmap
- Open source contributions with your use cases in mind
- Co-marketing and case study opportunity
About the Author

Misha Sulpovar is a Wise Owl at PromptOwl, a context engineering and governance platform, and the author of The AI Executive: Harnessing the Ungoverned Machine.
His work focuses on the infrastructure layer where context, governance, and AI execution intersect.
References
- Context Nest Specification — Apache-2.0
- Context Nest on npm — Engine, CLI, MCP Server
- Model Context Protocol (MCP) — Anthropic
- EU AI Act — Transparency Requirements
- NIST AI Risk Management Framework
Context Nest is an open specification by PromptOwl, Inc.
Engine & MCP Server: AGPL-3.0 | CLI: Apache-2.0 | Specification: Apache-2.0