Curiosity · Concept
Hallucination
LLM hallucination is the phenomenon of a model producing fluent, confident, but false output — invented citations, wrong dates, nonexistent APIs, fabricated quotes. It's rooted in how models are trained: to predict plausible continuations, not to verify facts. Mitigations include grounding via RAG, tool use, decoding controls, self-consistency, and training techniques that penalize overconfident guesses.
Quick reference
- Proficiency
- Beginner
- Also known as
- LLM hallucination, confabulation, fabrication
- Prerequisites
- LLM basics
Frequently asked questions
What is an LLM hallucination?
It is when an LLM produces content that is wrong, fabricated, or unsupported — stated with the same confidence and fluency as correct content. Examples: inventing a court case, citing a nonexistent API, misquoting a real paper, or getting a simple date wrong.
Why do LLMs hallucinate?
They're trained to predict plausible next tokens, not to verify facts. When the model's parametric memory is thin (long-tail facts, post-cutoff events, ambiguous entities), it pattern-matches something that looks right. Sampling with temperature also introduces randomness that can tip into wrong outputs.
How can I reduce hallucinations in production?
Ground the model: use RAG to supply real sources, use tool calls for facts you can look up, require the model to cite retrieved passages, lower temperature, use structured outputs, and have the model explicitly say 'I don't know' when confidence is low. LLM-as-judge or groundedness checks post-hoc are also common.
Are reasoning models less prone to hallucinations?
Mixed. Reasoning models (o1, o3, R1, Claude thinking) can catch their own factual errors during deliberation, and tend to do better on math and logic. But they still hallucinate facts they don't know, and longer reasoning can sometimes let errors propagate. Grounding and tool use still matter.
Sources
- Huang et al. — A Survey on Hallucination in LLMs — accessed 2026-04-20
- Lin et al. — TruthfulQA — accessed 2026-04-20
- OpenAI — SimpleQA — accessed 2026-04-20
- Google — FACTS Grounding benchmark — accessed 2026-04-20