Why Epistemic Governance Matters for AI Agents
Most agent frameworks have no mechanism for uncertainty. Here's why that's a problem and how epistemic governance addresses it.
Why Epistemic Governance Matters for AI Agents
The dominant pattern in agent orchestration is simple: receive a task, select a tool, call it, return the result. LangChain, CrewAI, AutoGen, and most other frameworks differ in orchestration style but share the same assumption -- the agent's job is execution, not reasoning about the quality of its own knowledge.
This produces a specific failure mode. Agents complete tasks confidently while being wrong, and nothing in the architecture tells them to pause.
The Confidence Problem
When an LLM-backed agent researches a claim, it generates a response with uniform syntactic certainty. "Water boils at 100 degrees Celsius at sea level" and "this supplement reduces inflammation by 40%" arrive in the same tone, with the same apparent confidence. The agent maintains no internal representation of epistemic state -- no record of what has been verified, what is contested, or what remains unknown.
This is not hypothetical. Kadavath et al. (2022) showed that language models are frequently overconfident on questions outside their training distribution. Wrapping an uncalibrated model in an agent loop does not fix this; it amplifies it, because each confident-but-wrong intermediate step feeds into the next.
The result is an architecture that is structurally incapable of saying "I do not have enough evidence to answer this."
What Epistemic Governance Means
Epistemic governance is the principle that an agent system should maintain an explicit model of its own knowledge state: what claims it holds, what evidence supports or undermines those claims, and how it should allocate its next action based on information value rather than task completion alone.
The idea draws on two lines of academic work. Lindley (1956) formalized expected information gain in experimental design -- a mathematical framework for ranking possible observations by their expected reduction of uncertainty. Dung (1995) developed abstract argumentation frameworks that formalize how claims attack and support each other in a directed graph. Together, these provide the theoretical backbone: an agent should choose its next action based on what would most reduce its uncertainty, and it should track the adversarial structure of its evidence.
In an epistemically governed system, the agent does not ask "what should I do next to finish the task?" It asks "what action would most reduce my uncertainty about the claims I am tracking?" These are different questions with different answers.
How VAOS Implements This
VAOS (Viable Agent Orchestration System, named after Beer's Viable System Model) implements epistemic governance through vaos-ledger, an OTP GenServer that maintains a structured record of three entity types:
Claims -- propositions under evaluation, each carrying a confidence score. Example: "Homeopathy is effective for treating chronic conditions."
Evidence -- findings that support or undermine specific claims, with source attribution, publication type, and methodology quality indicators.
Attacks -- directed relationships where one piece of evidence contradicts another. A systematic review attacks a case study. A replication failure attacks an original finding.
Expected Information Gain Scoring
When the agent must decide what to do next, EIG scoring ranks 5 action types by information value:
| Action | Purpose | High EIG When |
|---|---|---|
investigate_claim | Deep research on a single claim | Claim confidence near 0.5 (maximum uncertainty) |
seek_counter_evidence | Targeted search for disconfirming evidence | Evidence is one-sided |
replicate_finding | Re-verify an existing piece of evidence | A claim rests on a single source |
survey_literature | Broad scan for related academic work | Evidence set is small |
test_prediction | Derive and check a falsifiable prediction | Claim confidence is high but untested |
The key insight: an agent that already holds 8 confirming pieces of evidence gains less from a 9th confirmation than from one serious attempt at disconfirmation. EIG scoring encodes this directly.
The Investigate Tool: Adversarial Self-Examination
The most distinctive component is the investigate tool, which uses a dual-prompt adversarial architecture. When the agent investigates a claim, the process runs in two phases:
Phase 1 -- Advocacy. The agent builds the strongest possible case FOR its current position, gathering citations and constructing a grounded argument.
Phase 2 -- Adversarial. A separate prompt instructs the agent to argue AGAINST its own Phase 1 findings. It must find weaknesses, identify missing evidence, and surface counterarguments -- citing specific sources, not generating vague hedges.
The final output synthesizes both phases into a confidence-rated report. This is not a guarantee of correctness. But it structurally prevents the most common failure mode: the agent finding three agreeable sources and stopping.
Concrete Results: The Homeopathy Test
To test this architecture, we ran the investigate tool on the claim "homeopathy is effective for treating illness." This is a useful test case because it has a strong scientific consensus (against efficacy beyond placebo) alongside a large volume of belief-supporting content online. An uncalibrated agent would typically produce a "both sides" summary.
The results:
- The tool correctly identified asymmetric evidence against homeopathy's efficacy
- It found 3 grounded citations with empirical support and 5 belief-based citations (anecdotal, non-peer-reviewed, or from advocacy sources)
- It found 4 verified citations arguing against efficacy, all from systematic reviews or meta-analyses
- The adversarial pass identified that pro-homeopathy evidence came primarily from low-quality case studies, while counter-evidence came from systematic reviews
- Total investigation time: approximately 17 minutes
The important outcome was not that it reached the scientifically supported conclusion -- a well-prompted LLM can do that for a settled question. The important outcome was that it could show why: which evidence it found, how it weighed that evidence, and what the adversarial pass uncovered that the initial pass missed.
Limitations
This is early-stage work. The gaps are real and worth stating plainly.
Single-node persistence. The ledger writes to JSON files on a single GenServer. No replication, no distributed consensus. For a system built on OTP -- a platform designed for distributed computing -- this is an obvious architectural gap.
Simplistic Scorer heuristic. Evidence relevance uses term overlap between evidence text and claim text, not semantic similarity. A paper about "water memory" and a paper about "homeopathic dilutions" may not match despite being directly related. Embedding-based similarity would improve this substantially, but adds a vector store dependency we chose to defer.
No claim graph visualization. The attack/support relationships exist in the data model but have no visual interface. For debugging complex investigations, this makes the system harder to inspect than it should be.
Speed. 15-20 minutes per investigation. The bottleneck is sequential LLM calls in the adversarial phase. Acceptable for research tasks, unusable for anything interactive.
No systematic evaluation. The homeopathy case is illustrative, not a benchmark. Building a proper evaluation suite requires ground-truth datasets of claims with known evidence distributions -- itself a research problem we have not yet solved.
Why This Direction Matters
The current trajectory of agent frameworks is toward more tools, more integrations, more autonomy. For well-defined tasks with reliable tool outputs, this works. But as agents are asked to handle research, analysis, and decision support -- domains where being wrong has consequences -- the absence of epistemic machinery becomes a structural liability.
An agent that tracks its own uncertainty is not smarter. It is more honest about what it does and does not know. That honesty is a prerequisite for building systems that humans can trust with consequential work.
The VAOS epistemic governance system is implemented in vaos-ledger (7,631 lines, 241 tests) and the investigate tool in vas-swarm. Both are Elixir/OTP.
References
- Beer, S. (1972). Brain of the Firm: The Managerial Cybernetics of Organization. Allen Lane.
- Dung, P. M. (1995). On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming and n-person games. Artificial Intelligence, 77(2), 321-357.
- Kadavath, S., et al. (2022). Language models (mostly) know what they know. arXiv preprint arXiv:2207.05221.
- Lindley, D. V. (1956). On a measure of the information provided by an experiment. The Annals of Mathematical Statistics, 27(4), 986-1005.