When AI Says “This Quote Is Accurate,” You Shouldn’t Believe It
Verification, reliability, and the new epistemic burden in the age of generative systems
Educating AI publishes twice weekly. If this piece was useful, share it with a colleague who’s navigating these questions. And if you want to support the work, consider becoming a paid subscriber. It’s what keeps this going.
Check out my serial releases from my latest book, Thinking with AI: A Student’s Guide to Literacy in an AI-Rich World.
Intro and Ch. 1 Ch. 2-3 Ch. 4-5 Ch. 6-7
Next week, I hope to release some training roadmaps for educators and administrators interested in piloting discipline-specific AI Implementation in their contexts!
Nick
One of the most dangerous sentences an AI system can produce is also one of the most reassuring: “Yes, this quote matches the original document.”
It feels definitive. Settling the nerves. The kind of authoritative confirmation that lets you move on to the next task. In my experience, it is often wrong. Not dramatically wrong. Not in ways that announce themselves. But wrong in the subtle, structurally consequential ways that undermine the integrity of research, instruction, and professional judgment. This is not a fringe failure mode. It is a direct consequence of how generative AI systems work.
The Misconception
When you upload a document into a chat and ask the model to quote from it, there is a natural assumption that the system can simply “look it up.” That somewhere inside the machine, a search function locates exact text and compares it with mechanical precision.
This is not how large language models operate.
They do not retrieve text the way databases do. They reconstruct language probabilistically, token by token, based on patterns, likelihoods, and semantic approximation. Even when the original document is fully present in the prompt, the model does not perform exact character-by-character comparison. It generates what seems right. And what seems right is often close enough to feel authoritative without being literally accurate.
Reconstruction, Verification, and Why Both Fail
Consider a scenario many of us have already encountered. A teacher uploads a policy document and asks the model to quote the section addressing student data privacy. The model produces: “Schools must ensure student data privacy when implementing AI systems.” It sounds correct. It captures the meaning. But the actual document reads: “Schools are responsible for ensuring the privacy and security of student data when implementing AI systems.” The model did not retrieve that sentence. It reconstructed it. One is a verbatim quotation. The other is a paraphrase disguised as one. In education, in policy, in research, the difference between what a source actually says and what it approximately means is the difference between evidence and interpretation.
If the teacher then asks, “Does this quote match the original?” the model will frequently respond with confidence: yes, it does. That confident verification is produced by the same probabilistic process that generated the inaccurate quote in the first place. The system is not stepping outside itself to scan the document like software running a string match. It is generating a judgment about plausibility, evaluating whether the quote feels consistent with the source, not whether it is identical to it.
Researchers call this self-verification failure. Once the model generates a quote, it treats that output as a strong prior and tends to defend it when asked to verify. Because “careful” and “thorough” are semantically similar, the model will judge a substitution of one for the other as correct. Because a sentence with a dropped clause still carries the same general meaning, the model will affirm that nothing is missing. The verification step feels like a second opinion. It is not. It is the same opinion, restated with confidence.
What About RAG?
Some readers familiar with AI architecture may be wondering about Retrieval-Augmented Generation, or RAG, which works differently from standard chat. Instead of processing an entire document at once, a RAG system breaks it into smaller chunks, indexes them, and retrieves only the most relevant passages in response to a query. The model generates its answer based on those retrieved passages rather than the full document.
This offers real advantages for quotation reliability. The model works with a smaller, targeted passage, reducing the positional bias that plagues long-context processing. The retrieved chunk can be presented alongside the response, giving users something concrete to verify against. Well-designed RAG systems can even enforce extractive behavior, requiring the model to pull directly from retrieved text rather than generating freely. For many professional and educational use cases, a well-configured RAG pipeline will produce more grounded, more traceable outputs than a standard chat upload.
But RAG shifts the failure modes rather than eliminating them. If the chunking strategy splits a key sentence across two passages, or the embedding model fails to rank the right passage highly enough, the correct text never reaches the generator. Even with the correct passage retrieved, the generator still operates probabilistically and can still paraphrase or compress unless explicitly constrained. And most critically, recent research by Wallat et al. (2025) found that up to 57% of citations in RAG systems lack genuine faithfulness: the system attaches a citation that looks correct without having actually derived the claim from that passage. The grounding is decorative. RAG moves the needle. It does not solve the problem.
What This Means for Education
Teachers are increasingly turning to AI to support grading and content workflows. A teacher might upload a student essay alongside source materials and ask the AI to check whether the student accurately quoted a primary source. The AI may confidently affirm accuracy even when the student has subtly altered the original text. The error is not caught. It is certified. Students face a parallel risk. A student writing a research paper may rely on an AI system to locate and quote supporting evidence. The system may produce language that appears authoritative and convincingly specific, but that does not actually exist in the cited source. If the student trusts the verification step, the error becomes embedded in the academic record.
We need to start distinguishing more carefully between tasks that require semantic approximation and tasks that require literal fidelity. Generative AI is extraordinarily effective at semantic tasks: summarization, explanation, synthesis, brainstorming, feedback. These benefit from the model’s ability to reconstruct meaning flexibly. Quotation, citation, and verification are fundamentally different. They require extractive precision, the ability to identify and reproduce exact spans of text without alteration. By default, language models are not optimized for this. They are optimized to produce the most plausible continuation of a sequence, which is a different objective entirely. A teacher asking AI to generate feedback on student writing is operating within the model’s strengths. A teacher asking AI to verify whether a quote appears exactly in a source document is operating outside them.
This is why AI literacy cannot be reduced to prompt engineering or tool familiarity. It must include what I would call epistemic literacy: an understanding of what kinds of questions AI can answer reliably and what kinds require independent verification. Students must learn that AI systems are not authoritative sources. They are generative collaborators whose outputs must be treated as provisional until confirmed through external mechanisms. Teachers must similarly recalibrate their trust. The presence of a document within an AI chat does not guarantee faithful reproduction.
None of this diminishes the transformative potential of AI in education. I have written candidly in this newsletter about how much I benefit from working with these tools, and I am not about to pretend otherwise. But the time saved on generation must be partially reinvested in verification, and I mean genuine verification, not the model checking its own work.
Reliability Is a System Property
We are moving from a world where verification was primarily procedural to one where verification is interpretive. The tools we use to generate knowledge can also generate plausible errors. This places new responsibilities on all of us: to design workflows that preserve human verification at critical points, to teach students not just how to use AI but how to question it, and to develop institutional norms that distinguish between generative assistance and authoritative sourcing.
AI systems are extraordinarily fluent. They produce language that feels precise, confident, and complete. That fluency is their greatest strength and their greatest epistemic risk. When an AI system tells you that a quote is accurate, it is expressing a probabilistic judgment, not a deterministic fact. Understanding this distinction is foundational to every serious conversation we are going to have about AI in education going forward.
Reliability is not a property of the tool alone. It is a property of the human-tool system. And in that system, human judgment remains indispensable.
Nick Potkalitsky, Ph.D.
Further Reading
The research conversation behind this article is growing and worth following. Here are five entry points for readers who want to go deeper:
Lost in the Middle: How Language Models Use Long Contexts (Liu et al., 2024, Transactions of the Association for Computational Linguistics).
Correctness Is Not Faithfulness in RAG Attributions (Wallat et al., 2024/2025, ACM SIGIR ICTIR 2025).
When Can LLMs Actually Correct Their Own Mistakes? A Critical Survey of Self-Correction (Kamoi et al., 2024, Transactions of the Association for Computational Linguistics).
Not Wrong, But Untrue: LLM Overconfidence in Document-Based Queries (Hagar, 2025, Computation + Journalism Symposium).
LaRA: Benchmarking Retrieval-Augmented Generation and Long-Context LLMs (Li et al., 2025, ICML 2025).
Check out some of our favorite Substacks:
Mike Kentz’s AI EduPathways: Insights from one of our most insightful, creative, and eloquent AI educators in the business!!!
Terry Underwood’s Learning to Read, Reading to Learn: The most penetrating investigation of the intersections between compositional theory, literacy studies, and AI on the internet!!!
Suzi’s When Life Gives You AI: A cutting-edge exploration of the intersection among computer science, neuroscience, and philosophy
Alejandro Piad Morffis’s The Computerist Journal: Unmatched investigations into coding, machine learning, computational theory, and practical AI applications
Michael Woudenberg’s Polymathic Being: Polymathic wisdom brought to you every Sunday morning with your first cup of coffee
Rob Nelson’s AI Log: Incredibly deep and insightful essay about AI’s impact on higher ed, society, and culture.
Michael Spencer’s AI Supremacy: The most comprehensive and current analysis of AI news and trends, featuring numerous intriguing guest posts
Daniel Bashir’s The Gradient Podcast: The top interviews with leading AI experts, researchers, developers, and linguists.
Daniel Nest’s Why Try AI?: The most amazing updates on AI tools and techniques
Jason Gulya’s The AI Edventure: An important exploration of cutting-edge innovations in AI-responsive curriculum and pedagogy



