Skip to content
Notes on AIE033Act 3 — Steering & Trust

What Hallucinations Are

Define hallucinations accurately without mystique.

Full Explanation

A hallucination is what happens when an AI model produces plausible-sounding text that isn't actually true. The word suggests something unusual — a drift, a malfunction. But the model isn't doing anything unusual. It generates the most plausible continuation of its input, exactly as it was trained to. When that output happens to match reality, we call it correct. When it doesn't, we call it a hallucination. The model was doing the same thing both times.

There is no built-in step in the core generation process where the model checks its output against reality. Modern systems can reduce hallucinations by adding external checks — retrieving documents, verifying against sources, applying constraints — but these work around the model, not inside it. The underlying behavior doesn't change: the model produces fluent text with no mechanism for knowing whether that text is true. That gap between fluency and truth is the thing to understand.

---

Related AI Concepts

Resources

No dedicated resources for this episode yet.

Browse the resource library →
Alexey Makarov

Alexey Makarov

AI Enablement Strategist and Educator. Leading the AI Center of Excellence at SEFE. Creator of the Unreasonable AI YouTube channel. Based in Berlin.

About Alexey →