Abstract
Generative AI models continue to produce factual errors despite improvements in reliability, with the key challenge being distinguishing known from unknown information rather than simply expanding knowledge boundaries.
Despite significant strides in factual reliability, errors -- often termed hallucinations -- remain a major concern for generative AI, especially as LLMs are increasingly expected to be helpful in more complex or nuanced setups. Yet even in the simplest setting -- factoid question-answering with clear ground truth-frontier models without external tools continue to hallucinate. We argue that most factuality gains in this domain have come from expanding the model's knowledge boundary (encoding more facts) rather than improving awareness of that boundary (distinguishing known from unknown). We conjecture that the latter is inherently difficult: models may lack the discriminative power to perfectly separate truths from errors, creating an unavoidable tradeoff between eliminating hallucinations and preserving utility. This tradeoff dissolves under a different framing. If we understand hallucinations as confident errors -- incorrect information delivered without appropriate qualification -- a third path emerges beyond the answer-or-abstain dichotomy: expressing uncertainty. We propose faithful uncertainty: aligning linguistic uncertainty with intrinsic uncertainty. This is one facet of metacognition -- the ability to be aware of one's own uncertainty and to act on it. For direct interaction, acting on uncertainty means communicating it honestly; for agentic systems, it becomes the control layer governing when to search and what to trust. Metacognition is thus essential for LLMs to be both trustworthy and capable; we conclude by highlighting open problems for progress towards this objective.
Community
Why do frontier LLMs still hallucinate? ๐
Our ICML 2026 paper (position track) offers a simple diagnosis and a constructive path forward ๐
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Delineating Knowledge Boundaries for Honest Large Vision-Language Models (2026)
- Inducing Epistemological Humility in Large Language Models: A Targeted SFT Approach to Reducing Hallucination (2026)
- MARCH: Multi-Agent Reinforced Self-Check for LLM Hallucination (2026)
- Mitigating LLM Hallucinations through Domain-Grounded Tiered Retrieval (2026)
- PRISM: Probing Reasoning, Instruction, and Source Memory in LLM Hallucinations (2026)
- Facet-Level Tracing of Evidence Uncertainty and Hallucination in RAG (2026)
- Beyond"I Don't Know": Evaluating LLM Self-Awareness in Discriminating Data and Model Uncertainty (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2605.01428 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper