This is the practical companion to episode five of Actually, AI, hallucination.
You have heard the mechanism. A language model generates text by predicting the next most likely token. It has no truth engine. No fact database it checks against. The same process that produces brilliant analysis also produces complete fabrications, and from the outside, both look identical.
So what do you actually do with that?
Here is the honest answer. You use AI the way you would use a brilliant colleague who has read everything but remembers some of it wrong, and who will never, ever tell you when they are guessing. That colleague is incredibly useful. But you would not hand them a legal brief and file it without reading it. You would not let them write your medical advice and send it to a patient. You would not take their confident claim about a date or a statistic and put it in a published article without checking.
The trick is knowing when to check and how hard to check. Because checking everything defeats the purpose of having the tool, and checking nothing is how a New York lawyer ends up sanctioned by a federal judge.
Not all AI output carries the same risk. A useful mental model is to sort tasks into three buckets.
The first bucket is low stakes creative work. Brainstorming names for a project. Drafting a casual email. Generating ideas for a presentation. Writing a first draft of marketing copy. If the output contains a hallucinated fact, the consequence is that you rewrite a sentence. Nobody gets hurt. Nobody gets sued. The cost of an error is measured in minutes of your time. In this bucket, you can move fast and trust broadly. Hallucination is an inconvenience, not a hazard.
The second bucket is professional reference work. Summarizing a document. Explaining a technical concept. Writing code. Answering a factual question about history or science. Here the stakes rise. A hallucinated code example might introduce a bug. A wrong date in a summary might mislead a decision. A fabricated explanation of how an API works might send you down a dead end. In this bucket, you trust the structure and the reasoning but verify the specifics. Treat the output as a very good first draft that needs a fact check pass, not as a finished product.
The third bucket is high stakes consequential output. Legal citations, as we saw. Medical information. Financial advice. Anything where a wrong answer has real world consequences for real people. In this bucket, AI is a research assistant, not an authority. Every specific claim needs independent verification. Every citation needs to be checked against the actual source. Every number needs a second opinion. The output saves you time by giving you something to verify rather than something to accept.
Most people get in trouble because they use bucket one habits in bucket three situations. The lawyer in the Mata case was not careless in general. He was a thirty year veteran who had handled hundreds of cases. He applied casual trust to a high stakes task because the output looked so good. The quality of the prose is not correlated with the accuracy of the facts. A model can write a beautifully formatted, perfectly structured, completely fabricated legal citation. The better the writing, the harder it is to spot the lie. That asymmetry is the core danger.
Let us clear the table of the approaches that feel like verification but are not.
Asking the model "are you sure about that?" does not work. The main episode covered why. When you push back, the model does not go check anything. It generates what a confident reassurance would look like. Sometimes it happens to self correct, because "actually, I was wrong" is a pattern in its training data. But you have no way of knowing whether the correction is real or whether the model is now hallucinating a correction of a hallucination. You are asking the same unreliable process to audit itself using the same unreliable process.
Asking for sources does not work either. This is the one that catches people most often. You ask the model to cite its sources. It provides URLs, author names, publication dates, page numbers. They look real. They follow the right format. And a meaningful fraction of them are completely fabricated. The model learned the pattern of what a citation looks like. It can generate citations with the same fluency it generates prose. The citation is not a link to something the model looked up. It is a prediction of what a plausible citation would be.
Asking the model to rate its own confidence does not work reliably. The deep dive covered the Kadavath research showing models have some internal calibration, but expressing that calibration to the user is a different matter. A model trained with reinforcement learning from human feedback has been rewarded for confident, helpful sounding answers. Saying "I am sixty percent sure" gets lower ratings from human evaluators than saying "the answer is." The confidence the model expresses tells you about what sounds appropriately confident in context, not about the actual reliability of the claim.
And rephrasing the same question and asking again does not work the way you think. If you get the same answer twice, it might mean the answer is correct. Or it might mean the statistical pattern that produced the wrong answer the first time is strong enough to produce it again. Consistency is not accuracy. A model can be consistently wrong.
Now the part you came for. Four strategies that provide real verification, ordered from quickest to most thorough.
Strategy one. Ask the model to flag its own uncertainty differently. Instead of asking "are you sure?", which triggers a reassurance pattern, try asking "which parts of your answer are you least certain about?" or "what claims in this response would you want to verify before publishing?" This frames the task as identifying uncertainty rather than defending accuracy. It does not guarantee honest uncertainty reporting, but it activates different patterns in the model, ones trained on careful, hedging language rather than confident assertion. You will often get useful flags. Not always. But more often than "are you sure?" which gets useful flags almost never.
Strategy two. Verify against an external source. This is the boring, reliable answer. Take the specific claims, the names, dates, numbers, citations, and check them against a source that is not the model. Search for the paper it cited. Look up the person it named. Check the date it gave you. This does not require checking every sentence. Focus on the claims that matter. In a five paragraph summary, there might be two specific facts that your decision depends on. Check those two. Let the rest ride. The goal is targeted verification, not exhaustive verification.
Strategy three. Cross check with a different model. This is not the same as asking the same model again. Different models were trained on different data with different techniques. If you ask Claude, then ask Gemini, then ask ChatGPT, and all three give you the same specific fact, your confidence should go up. If they disagree, you have found a claim that needs external verification. This is especially useful for factual questions where you do not have easy access to primary sources. Three models agreeing is not proof. But three models disagreeing is a strong signal that at least one of them is hallucinating, and you should go find out which.
Strategy four. Use retrieval grounded tools when they are available. The deep dive covered retrieval augmented generation, where the model searches a knowledge base before answering. Many current AI tools offer this, search enabled chat, document analysis modes, tools that cite specific passages from uploaded files. When you need factual accuracy, use the version of the tool that retrieves from real documents rather than the version that generates from memory. The model can still misinterpret what it retrieves, but at least there is a paper trail you can check.
Here is something you can do in the next five minutes to build intuition about hallucination.
Open your preferred AI chat tool. Ask it to tell you about a real but somewhat obscure person, someone you already know about. Not a world famous figure whose facts are thoroughly represented in training data. Pick a local business owner, a niche academic, a mid career artist. Someone with a real but limited online presence.
Read the response carefully. Some of it will be correct. Some of it will be plausible but wrong. And you will not be able to tell which is which just by reading the text. The tone is the same. The confidence is the same. The formatting is the same.
Now check two specific claims against a real source. A date. An institution. A publication title. See what holds up and what does not.
This exercise is worth more than any amount of theory, because it breaks the illusion that good writing means good facts. Once you have caught a hallucination in the wild, about something you actually know, you never trust AI output the same way again. Not because you stop using it. Because you start using it correctly.
Here is the working framework. It is not complicated, but it does require a change in how you relate to AI output.
Trust the structure. Language models are excellent at organizing information, suggesting frameworks, outlining arguments, and generating first drafts. The structure of the output, the way it is organized, the flow of the argument, the completeness of the outline, these are usually reliable. The model learned patterns of good structure from billions of examples.
Verify the specifics. Names, dates, numbers, citations, causal claims, historical details, any place where there is a specific checkable fact. These are where hallucination lives. Not in the broad strokes but in the precise details. A paragraph that says "in the early two thousands, a researcher at a major university published an influential paper showing that X leads to Y" is probably structurally correct. The model has the right shape of the story. But the specific year, the specific researcher, the specific university, the specific finding, any of those might be wrong.
Escalate verification with stakes. The higher the consequence of being wrong, the more claims you check and the harder you check them. A casual brainstorm gets no verification. A blog post gets a spot check. A legal filing gets every citation verified against the actual source. Match your effort to the consequences.
And accept the asymmetry. AI makes you faster at generating. It does not make you faster at verifying. The tool gives you a first draft in seconds that would have taken an hour. The verification still takes the time it takes. That is the real bargain. You are trading generation time for verification time. For most tasks, that trade is enormously favorable. For some tasks, it is not. Knowing which is which is the skill that separates someone who uses AI well from someone who will eventually end up in front of a judge explaining why their citations do not exist.
That was the practical companion for episode five.