Context Graphs: Capturing Decision Traces for Scalable AI Autonomy
TL;DR
- Context graphs capture "decision traces" -- the why behind actions, exceptions, and precedents -- enabling scalable AI autonomy beyond static systems of record.
- Enterprise AI agents require access to decision traces, currently residing in informal channels like Slack, to understand and replicate past human judgment and resolve conflicts.
- By persisting decision traces, agents can build queryable "context graphs" that reveal organizational ontology through actual usage, rather than predefined schemas.
- Context graphs allow companies to audit and debug AI autonomy, turning exceptions into searchable precedent and improving decision-making feedback loops over time.
- Designing systems for context graphs necessitates adapting organizational workflows to agentic needs, shifting individual contributors into roles managing and guiding AI agents.
- Allowing agents to discover organizational schemas on the fly, rather than predefining context graphs, reveals "policy in practice" and uncodified rules.
Deep Dive
Context graphs represent a pivotal advancement in enterprise AI, shifting the focus from merely organizing data to capturing the "why" behind decisions, which is critical for enabling scalable AI autonomy. While traditional systems of record store the "what" of business operations, context graphs preserve the decision traces--exceptions, overrides, and justifications--currently scattered across informal communication channels and human memory, thereby unlocking the next frontier for AI agents.
The core challenge for enterprise AI, particularly with the rise of autonomous agents, lies in ensuring these agents access and utilize the correct, canonical information at crucial decision points. Current systems of record, though vital for operational data, often fall short because they reflect only the final state of a transaction, not the complex reasoning or precedents that led to it. For instance, calculating Annual Recurring Revenue (ARR) can yield different figures depending on whether sales, finance, or accounting provides the data, each with its own set of exclusions and adjustments. When an agent is tasked with a decision, such as calculating ARR for a board presentation, it needs more than just raw data; it requires an understanding of which definition of ARR is authoritative and why. This "what versus why" gap is currently filled by tribal knowledge, Slack threads, deal desk conversations, and human judgment, none of which are easily queryable or scalable.
Context graphs address this by persisting the decision traces generated as agents execute tasks. When an agent interacts with various systems, evaluates policies, seeks approvals, or applies exceptions, these interactions create a rich record. By capturing and stitching together these decision events--the inputs gathered, policies evaluated, exceptions invoked, approvals granted, and the resulting state changes--a queryable "context graph" emerges. This graph acts as a living record of how decisions were made, providing the crucial "why" that empowers agents with deeper understanding and autonomy. For example, if a renewal discount policy caps at 10% but an agent approves a 20% discount, the context graph would record the specific justification, such as a prior VP approval for a similar deal or an approved service impact exemption, turning exceptions into searchable precedent. This allows organizations to audit and debug AI autonomy, prevent the repeated relearning of edge cases, and ultimately, compound intelligence over time as each automated decision adds to the graph.
The development of these context graphs also necessitates a shift in how we design and interact with AI systems. Rather than rigidly pre-defining structures, the most effective approach involves allowing agents to discover the organizational ontology dynamically through their real-world interactions. As agents traverse APIs, query documentation, and review past decisions, they uncover how entities genuinely relate and how rules are applied in practice, revealing the organizational schema organically. This emergent understanding, akin to a world model, allows for the identification of policies that are consistently broken and effectively become the de facto rules. Furthermore, the advent of agentic AI redefines human roles; individual contributors will transition into "managers of agents," responsible for providing oversight, directing context, and shepherding work between various AI agents, mirroring the coordination efforts of human team managers. This evolution means organizations will adapt to how AI works, optimizing workflows to provide agents with the necessary context, thereby leveraging human judgment for nuanced oversight and decision-making.
Ultimately, the ability to effectively engineer and leverage context graphs will be a key differentiator for enterprise AI in 2026. By moving beyond static data to capture the dynamic reasoning behind decisions, organizations can unlock scalable autonomy, improve the reliability of AI agents, and enable a more agile and responsive approach to business challenges. The human element remains critical, not in performing rote tasks, but in providing the judgment and context that fuels and refines the AI's decision-making process.
Action Items
- Create context graph schema: Define 5 key entities (e.g., deals, tickets, policies) and their relationships based on observed agent decision traces (ref: Foundation Capital essay).
- Audit 3-5 existing workflows: Identify and document decision traces (e.g., exceptions, overrides) currently residing in human communication channels (e.g., Slack, email).
- Design agent interaction protocol: Specify how agents will capture and persist decision traces during execution for 2-3 core business processes.
- Implement agent-driven policy refinement: Track instances where agents deviate from codified policy by 10% to identify implicit rules for 3-5 business areas.
- Evaluate 2-3 agent onboarding processes: Assess how effectively new agents are provided with historical decision traces to inform their actions.
Key Quotes
"The problem is that most of this lived downstream of the operational world the sales team still lived in salesforce the finance team still closed the books in netsuite the support team still worked tickets in zendesk the warehouse for lakehouse was the retrospective mirror not the transactional front door."
This quote from Jamin Ball highlights a fundamental disconnect in traditional enterprise data management. Ball argues that operational systems remain siloed within departments, and data warehouses or lakes, while retrospective, do not actively influence or guide daily operations. This separation means that even with advanced data infrastructure, the actual transactional front doors of businesses remain disconnected from a unified, operational view of data.
"Agents are inherently cross system meaning they don't live solely within one of those functions and they are action oriented they are not just trying to gather information they are trying to make use of that information to do things that combination he writes means agents are only as good as their understanding of which system owns which truth and what the contract is between those truths under the hood something still has to say this is the canonical customer record or this is the legally binding contract term or this number is the one we report to wall street that something might be a traditional system of record it might be a warehouse back semantic layer or it might be a new class of data control plane product but it is absolutely not going away."
Ball explains that agents, by their nature, operate across different systems and are designed to act on information, not just collect it. This capability makes their effectiveness dependent on understanding which system holds the definitive truth for specific data points and how these truths interrelate. Ball emphasizes that a central authority, whether a traditional system of record or a newer data control plane, is essential to establish and maintain these canonical truths for agents to function reliably.
"Ball's framing assumes the data agents need already lives somewhere and agents just need better access to it plus better governance semantic contracts and explicit rules about which definition wins for which purpose that's half the picture the other half is the missing layer that actually runs enterprises the decision traces the exceptions overrides precedents and cross system context that currently lives in slack threads deal desk conversations escalation calls and people's heads."
Gupta and Garg introduce the concept of "decision traces" as a critical missing layer in enterprise AI, complementing the idea of systems of record. They argue that while systems of record manage the "what" (e.g., a deal's discount), they fail to capture the "why" -- the specific context, exceptions, and precedents that led to a decision. This crucial "why" information, often residing in informal communication channels and human knowledge, is essential for agents to operate with true autonomy.
"The distinction the authors say is between rules and decision traces rules they say tell an agent what should happen in general whereas decision traces capture what happened in this specific case i think for our purposes on this episode an even simpler way to understand it is the what versus why gap and the simple idea here is that while systems of record are good at state i e this particular deal closed at a 20 discount they are bad at decision lineage why a 20 discount was allowed this time as the authors point out those decision traces i e the why lives in slack and dms and in meetings and in human heads limiting how much autonomy can then scale as the authors put it agents don't just need rules they need access to the decision traces that show how rules were applied in the past where exceptions were granted how conflicts were resolved who approved what and which precedents actually govern reality."
Gupta and Garg clarify the difference between rules and decision traces, framing it as the "what versus why" gap. They explain that rules provide general guidance, while decision traces document the specific rationale, exceptions, and approvals behind individual actions. This "why" information, often found in informal communications and human memory, is crucial for agents to understand past decision-making processes and scale autonomy effectively.
"The most counter intuitive development we shouldn't pre define these context graphs traditional knowledge graphs fail because they require pre defining structure upfront context graphs invert this completely modern agents act as informed walkers through your decision landscape as an agent solves a problem traversing through apis querying documentation reviewing past tickets it discovers the organizational ontology on the fly it learns which entities actually matter and how they genuinely relate through use not through a manual schema someone designed in a workshop each trajectory leaves a trace which systems were touched together which data points co occurred in decision chains how conflicts were resolved accumulate thousands of these walks and something remarkable emerges the organizational schema reveals itself from actual usage patterns rather than predetermined assumptions these become world models not just retrieval systems."
The Cogent Enterprise substack argues against pre-defining context graphs, contrasting them with traditional knowledge graphs that require upfront structure. They propose that modern agents can discover an organization's ontology organically by "walking" through decision landscapes, interacting with APIs, and reviewing data. This process, driven by actual usage patterns rather than manual schema design, allows the organizational schema to reveal itself, forming "world models" rather than mere retrieval systems.
"Designing our systems to get agents access to that data and ensuring that all of our agents can interoperate on that data is going to be incredibly important further companies will have to drive a substantial amount of change management to make this all work we imagined that ai systems would adapt to how we work but it turns out due to their extreme power and inherent limitations we will instead adapt to how they work this means we will have to optimize our organizations and workflows to best enable context for agents to be successful the core tenet of this change is that the user is responsible now for directing and guiding agents on how to do their work ensuring it gets the right context along the way in essence he writes the individual contributor of today becomes the manager of agents in the future their new responsibilities will be providing the oversight and escalation paths a meaningful amount of coordination throughout the work that the agents are doing and shepherding work between the various agents just like managers of teams in the pre ai era."
Aaron Levy emphasizes the critical role of "context engineering" and the need for organizational adaptation to AI. Levy posits that companies must design systems for agent data access and interoperability, which will necessitate significant change management. He suggests that humans will adapt to how AI systems work, optimizing workflows to provide agents with the necessary context, and that individual contributors will evolve into "managers of agents," responsible for oversight, coordination, and guiding AI tasks.
Resources
External Resources
Books
- "Long Live Systems of Record" by Jamin Ball - Mentioned as the starting point for a discussion on how agents intersect with human knowledge work and the concept of systems of record.
Articles & Papers
- "AI's Trillion Dollar Opportunity: Context Graphs" (Foundation Capital) - Discussed as an exploration of a missing layer of information in enterprises, specifically decision traces, exceptions, and cross-system context.
- "The Era of Context" by Aaron Levy - Explored the importance of context engineering and how human roles will shift to managing and guiding agents.
Websites & Online Resources
- aidbnewyear.com - Referenced as the location for information on a 10-week self-guided project to upgrade AI skills.
- www.kpmg.us/agents - Mentioned as the website to discover how KPMG can help accelerate AI adoption.
- bsuper.ai - Referenced as the website for Superintelligent's agent readiness audits.
Other Resources
- Context Graphs - Mentioned as a concept about what it takes to get agents to do more important work, involving access to decision traces, exceptions, and cross-system context.
- Systems of Record - Discussed as traditional systems that are good at state but bad at decision lineage, and how agents change the equation by being cross-system and action-oriented.
- Decision Traces - Referenced as the capture of what happened in a specific case, including how rules were applied, exceptions granted, conflicts resolved, and approvals made, forming the basis of context graphs.
- Context Engineering - Identified as a key prediction for enterprise AI in 2026, focusing on designing systems for agent access to data and interoperability.