If you've spent any time using AI, you've probably noticed something unsettling: sometimes it just makes things up. It cites sources that don't exist, states facts that are wrong, and does all of this with total confidence. This phenomenon is called hallucination — and understanding it is essential for using AI effectively.
What Is a Hallucination?
In AI, a hallucination is when a language model generates output that is factually incorrect, fabricated, or not grounded in reality — presented as if it were true. The term is borrowed loosely from psychology (where it describes perceiving things that aren't there), and it's apt: the model is producing something that feels real but isn't.
Hallucinations range from subtle (slightly wrong dates or statistics) to dramatic (entirely made-up citations, fake quotes from real people, invented historical events).
Why Does AI Hallucinate?
The core reason is architectural. Language models don't work like databases — they don't look up stored facts. They generate text by predicting what word should come next based on patterns learned during training. This makes them fluent and flexible, but it also means there's no internal "fact-check" happening.
When a model encounters a question it doesn't have strong training signal for, it doesn't say "I don't know" — it generates the most plausible-sounding continuation. Sometimes that continuation is accurate. Sometimes it isn't.
Several factors increase hallucination risk:
- Obscure topics: Less training data means less reliable output
- Specific details: Numbers, dates, citations, and proper nouns are highest-risk
- Confident phrasing in the prompt: If you ask "Who said X?" the model feels compelled to name someone
- Model size: Smaller models hallucinate more frequently
The Confidence Problem
What makes hallucination especially dangerous is that the model's confidence level has essentially no correlation with accuracy. A model will state a fabricated citation in the same tone as a basic fact. There's no internal signal you can read to know when to trust it.
This is why "it sounded confident" is never a good reason to trust AI output on factual claims.
How to Catch Hallucinations
Verify specific claims. Any time the AI gives you a specific name, date, number, citation, or quote — verify it independently. This is non-negotiable for anything consequential.
Watch for citation red flags. Fake citations often have believable-sounding titles and authors but don't exist. Always look up cited papers, articles, or books directly.
Ask the model to express uncertainty. Prompt it with "Tell me if you're uncertain about any of this" or "Flag anything you're not sure about." This improves calibration somewhat, though it's not foolproof.
Use retrieval-augmented generation (RAG) for fact-critical tasks. RAG systems ground the model in specific documents before answering, significantly reducing hallucination for domain-specific questions.
Cross-check with a second model. If you get an important factual claim, ask a different AI system the same question. Disagreement is a signal to verify; agreement isn't a guarantee of accuracy.
What Hallucination Is Not
Hallucination is not dishonesty — the model has no intent. It's not stupidity — models that hallucinate can also do impressive reasoning. And it's not a problem that will simply disappear with the next model generation. Hallucination rates have improved significantly, but all current language models hallucinate to some degree.
The Right Mental Model
Think of an LLM like an extraordinarily well-read collaborator who has a genuinely poor memory for specific details. They can reason brilliantly about concepts, help you structure arguments, and generate creative ideas — but you wouldn't cite their memory for a specific statistic without checking it first. That's exactly the right relationship to have with AI output.