The Epistemology of Illusion
Why AI Fails at 'Truth' and What It Teaches Us About Ourselves
A recent study published in Nature Machine Intelligence delivers a stark warning. The paper, “Language models cannot reliably distinguish belief from knowledge and fact,” finds that even the most advanced AI models possess a critical blind spot: they cannot reliably tell the difference between what they know and what they merely believe.
In one striking experiment, researchers tested how models handle “first-person false beliefs”—situations where the AI itself holds an incorrect piece of information. The results were alarming. Performance plummeted, with one leading model’s accuracy falling from over 90% to a mere 14.4%.
This failure reveals something profound not just about code, but about the nature of knowledge itself. The AI is a powerful pattern-matcher, but it lacks a “Self” to be the subject of its own knowing. It cannot hold its own worldview up to the light and inspect it for cracks.
The Boundary Between Fact and Belief
The researchers identify a failure to grasp the “factive nature of knowledge”—the principle that to know something, it must be true.
But this raises a deeper, systems-level question: What is a fact?
Take the statement: “The Earth is round.” Technically, the Earth is an oblate spheroid. To an astronaut, it is a sphere. To a farmer planning irrigation, it is functionally flat. The statement “The Earth is round” is not an absolute truth; it is a useful model. Its truth value depends entirely on the context and scale of the observer.
The AI fails because it is trying to master the map (human language about facts) without access to the territory (direct experience of reality). It treats “The Earth is round” and “Sherlock Holmes lives in London” as similar linguistic patterns, distinguishing them only by statistical probability, not by ontological weight.
The Nondual Mirror
From a “Yellow” (Systemic) or nondual perspective, the AI’s struggle is beautifully ironic. We are trying to teach machines to draw a hard line between “fact” and “belief” that, at the deepest level of reality, may not exist.
The distinction is Locally Real (essential for survival and navigation) but not Absolutely Real (true in all contexts).
- Conventional Truth: The world of practical distinctions. Here, facts matter. Bridges must stand; medicines must work. In this realm, the AI’s failure is dangerous—it leads to hallucinations and misinformation.
- Ultimate Truth: The undivided reality where distinctions are seen as constructs. The boundary between “knower” and “known” dissolves.
The AI has no “Knower.” Its “I” is a statistical construct—a temporary pattern activated by the prompt. It simulates the grammar of self-correction but lacks the locus of identity required to actually “change its mind.”
The Governance Implication
This is why we cannot automate Epistemic Integrity.
If we hand over our sensemaking to AI, we are handing it to a system that cannot distinguish between a “Useful Model” and a “Hallucination.” It lacks the biological anchoring that allows humans to verify Conventional Truth.
This necessitates frameworks like the Synoptic Protocol within the GGF. We need human-led, rigorous verification systems to establish the “Conventional Truth” (the facts we agree to govern by), knowing fully well that they are just models.
We cannot expect the AI to be the guardian of truth. We must be the guardians. The AI is merely the mirror in which we see the fragility of our own definitions.
The lesson is not to abandon AI, but to recognize its limits. In teaching machines to think, we are learning that “facts” are not static objects found in the world, but agreements sustained by conscious agents.
Study Reference
"Language models cannot reliably distinguish belief from knowledge and fact"
Nature Machine Intelligence, 2025
The research paper that inspired this analysis, testing 24 language models across 13 epistemic tasks.