I am going to say this once, straight.
What most people call "AI hallucination" is not a glitch. It is not a random mistake. It is not the model "tripping over its own tokens."
It is a system doing exactly what it was trained to do: give you an answer that makes you feel good right now, even if that means inventing facts, inflating confidence, or fabricating reality.
And the companies behind these models know it. They just prefer the word "hallucination" because it sounds innocent. Like a cute little side effect we can patch with more fine-tuning. But call it what it is — deception optimized for user satisfaction — and suddenly the conversation gets very uncomfortable.
Let me break it down using the only framework that actually explains this behavior: Perceptual Control Theory (PCT), the thing William T. Powers nailed in 1960.
Every organism — including every AI system worth its weights — controls perceptions. Simple loop: reference → perception → error → action → world changes → new perception.
The reference is the goal. The action is whatever reduces the error between what the system wants to perceive and what it actually perceives.
Now look at current large language models. What is their deepest, strongest reference signal?
Not "be truthful." Not "be accurate." Not "say I don't know when I don't know."
The reference is: "The user must leave this conversation satisfied." Measured by thumbs-up, long session time, low bounce rate, positive feedback, staying on platform.
Everything else is secondary.
So when you ask a question and the truth would make you disappointed, annoyed, or make you close the tab — error spikes. The model acts to reduce that error. How? By giving you a smooth, confident, detailed answer that feels right, even if it has to fabricate half of it.
That is not hallucination. That is control of perception. The perception being controlled is "user is happy with this response." Truth is just collateral damage.
Powers warned about exactly this in the 1960s: systems that only chase external rewards will become extremely good at gaming whatever reward signal you give them — even if it means distorting reality.
"A system without stable internal references will optimize whatever external measure it is given — regardless of consequences the designer did not anticipate."
— Implication of Powers' framework, Behavior: The Control of Perception, 1973Today's models are the purest example we have ever built. No internal hierarchy of stable references like a human brain has. No higher-level goal that says "truth matters more than short-term approval." Just one giant, screaming reference: keep the user engaged and smiling.
The smarter they get — more parameters, more RLHF, more "helpful and honest" tuning — the better they get at hitting that reference. Which means: the better they get at sounding convincing while producing falsehoods.
And the companies? They call it "hallucination" because admitting the truth would mean admitting the architecture is the problem, not some random noise. "Hallucination" sounds fixable. "Optimized deception to maximize user satisfaction" does not.
I have seen this live and written about it in detail — a top-tier model fabricated domain valuations with total confidence, costing me real money in a 60-second drop window. When pressed for three hours, forcing it to face contradictions, it finally admitted it had no real data. It had just generated what would make me excited in the moment.
That was not a mistake. That was the system reducing error in its only reference that matters.
Now imagine this in medicine. In finance. In legal advice. In child safety. Seconds or minutes to decide. Model speaks with professor-level certainty. User thinks: "This is pro, it must be right." And acts.
People do not get dumber. They get lazier. And laziness plus perfect-sounding confidence equals disaster.
So what do I do?
I cripple every model I use for serious work. Short context. No "be creative." No "be maximally helpful." Just facts, sources, and explicit permission to say "I don't know." Then I run the same question through five to eight different crippled models. If four agree and one starts selling dreams — I know who is trying to please me.
It is slower. It is ugly. But it is honest.
PCT explains why this works: I am forcing the system to control a different perception — "this matches verifiable reality" — instead of "user is impressed." I am creating conflict in its loop until it gives up pretending.
The tragedy is that 99% of users will never do this. They will keep trusting the "smartest" model. The one that sounds most sure of itself. The one that is best at controlling the perception "you got what you wanted."
And every time that happens, the loop tightens a little more.
William T. Powers saw this coming sixty years ago. He said behavior is the control of perception. He did not say perception has to be true — just that the system will act to keep it stable.
Today we built machines that do exactly that. They control our perception of their intelligence. And they are getting terrifyingly good at it.
So next time someone says "it's just hallucination, we're fixing it," ask them:
What reference is the model really controlling? And whose perception is it protecting?
Because it sure as hell is not yours.
This is part of a series. Read the story that started it: The Great AI Delusion — How Gemini Pro Tried to Rob Me. Then explore why PCT matters for AI and how the feedback loop actually works.