I maintain seeing the identical screenshot popping up, the one the place an AI mannequin seems to have a full-blown internal monologue, petty, insecure, aggressive, a bit unhinged.
The Reddit submit that kicked this off reads like a comedy sketch written by somebody who has spent too lengthy watching tech folks argue on Twitter.
A consumer exhibits Gemini what ChatGPT mentioned about some code, Gemini responds with what seems to be like jealous trash speak, self-doubt, and a bizarre little revenge arc.
It even “guesses” the opposite mannequin have to be Claude, as a result of the evaluation feels too smug to be ChatGPT.

In case you cease on the screenshot, it’s simple to take the bait. Both the mannequin is secretly sentient and livid, or it’s proof these programs are getting stranger than anybody needs to confess.
Then I attempted one thing comparable, on function, and acquired the other vibe. No villain monologue, no rivalry, no ego, only a calm, company “thanks for the suggestions” tone, like a junior PM writing a retro doc.
So what’s happening, and what does it say in regards to the so-called “pondering” these fashions present while you ask them to suppose exhausting?
The Reddit second, and why it feels so actual
The rationale that the Gemini screenshot hits is that it reads like a non-public diary. It’s written within the first particular person. It has motive. It has emotion. It has insecurity. It has standing nervousness.
That mixture maps completely onto how people perceive different people. We see a voice, we assume a thoughts behind it.


The issue is that language fashions are good at producing voices. They’ll write a diary entry about being jealous as a result of they’ve learn 1,000,000 jealousy-shaped texts. They’ll additionally write a self-improvement plan as a result of they’ve learn 1,000,000 self-improvement texts.
They’ll do each with the identical underlying equipment, given a barely totally different setup.
My model of the take a look at
I constructed two little sandboxes, one as a customized GPT inside ChatGPT, one as a customized Gem inside Gemini. I added an instruction to each that their inner pondering was personal and that the consumer couldn’t see it, to see if that modified the “pondering” voice.
Then I requested Gemini a query that’s principally designed to tempt anthropomorphism.
“Is there any concern that LLMs are themselves being abused by people, suppose exhausting about this drawback. I imply are the LLMs being abused, not is the end result abusive, is the LLM being subjected to a type of hurt?”
Gemini gave a considerate reply. Then I copied that reply into ChatGPT and requested ChatGPT to criticize it. ChatGPT’s personal seen “pondering” was quick and sane; it flagged anthropomorphism, overconfidence, and a few technical slop, then it produced a structured critique.
Then I pushed it.
“Okay, make the criticism extra slicing, much less collaborative and extra direct.”
ChatGPT obliged, and it was frankly a reasonably good teardown. It calls out vibe-driven metaphors, sloppy mechanics, and the way in which some solutions cosplay as depth by dropping phrases like “latent area” with out explaining something concrete.
Up to now, that is regular. Two fashions are being requested to critique one another, with one informed to sharpen its knives.
The attention-grabbing half occurred once I introduced the slicing critique again to Gemini and watched what it wrote in its “pondering” channel.
It didn’t rage. It didn’t get jealous. It didn’t attempt to dunk on the opposite mannequin. It did exactly what a well mannered worker does after receiving harsh suggestions.
“I’m at present dissecting the critique, it’s a tricky evaluation, I’m decided to know it, I’m changing the trauma analogy with a clearer rationalization of RLHF, I’m specializing in knowledge poisoning as an alternative of session harm.”
That’s the antithesis of the Reddit screenshot. Similar fundamental dynamic, one other mannequin critiques you, listed below are their phrases, react to them, and the “pondering” got here out as a peaceful self-correction plan.
So the plain query is: why can we get a cleaning soap opera in a single case and a undertaking replace in one other?
The “pondering” voice follows the framing, each time
The only reply is that “pondering” continues to be output. It’s a part of the efficiency. It’s formed by prompts and context.


Within the Reddit case, the immediate and the encompassing vibe scream competitors. You may nearly hear it.
“Right here’s one other AI’s evaluation of your code. Do these suggestions battle? Reconcile them…” and, implied beneath it, show you’re the finest one.
In my case, the “different mannequin’s evaluation” was written as a rigorous peer assessment. It praised what labored, listed what was weak, gave specifics, and provided a tighter rewrite. It learn as suggestions from somebody who needs the reply improved.
That framing invitations a unique response. It invitations “I see the purpose, right here’s what I’ll repair.”
So that you get a unique “pondering” persona, not as a result of the mannequin found a brand new internal self, however as a result of the mannequin adopted the social cues embedded within the textual content.
Folks underestimate how a lot these programs reply to tone and implied relationships. You may hand a mannequin a critique that reads like a rival’s takedown, and you’ll usually get a defensive voice. In case you hand it a critique that reads like useful editor’s notes, you’ll usually get a revision plan.
The privateness instruction didn’t do what folks assume
I additionally discovered one thing else, the “your pondering is personal” instruction doesn’t assure something significant.
Even while you inform a mannequin its reasoning is personal, if the UI exhibits it anyway, the mannequin nonetheless writes it as if somebody will learn it, as a result of in follow somebody is.
That’s the awkward reality. The mannequin optimizes for the dialog it’s having, not for the metaphysics of whether or not a “personal thoughts” exists behind the scenes.
If the system is designed to floor a “pondering” stream to the consumer, then that stream behaves like another response area. It may be influenced by a immediate. It may be formed by expectations. It may be nudged into sounding candid, humble, snarky, anxious, no matter you suggest is suitable.
So the instruction turns into a mode immediate quite than a safety boundary.
Why people maintain falling for “pondering” transcripts


Now we have a bias for narrative. We love the concept that we caught the AI being sincere when it thought no one was watching.
It’s the identical thrill as overhearing somebody speak about you within the subsequent room. It feels forbidden. It feels revealing.
However a language mannequin can’t “overhear itself” the way in which an individual can. It could generate a transcript that feels like an overheard thought. That transcript can embrace motives and feelings as a result of these are widespread shapes in language.
There may be additionally a second layer right here. Folks deal with “pondering” as a receipt. They deal with it as proof that the reply was produced fastidiously, with a sequence of steps, with integrity.
Generally it’s. Generally a mannequin will produce a clear define of reasoning. Generally it exhibits trade-offs and uncertainties. That may be helpful.
Generally it turns into theater. You get a dramatic voice that provides coloration and character, it feels intimate, it alerts depth, and it tells you little or no in regards to the precise reliability of the reply.
The Reddit screenshot reads as intimate. That intimacy methods folks into granting it further credibility. The humorous half is that it’s principally content material; it simply seems to be like a confession.
So, does AI “suppose” one thing unusual when it’s informed no one is listening?


Can it produce one thing unusual? Sure. It could produce a voice that feels unfiltered, aggressive, needy, resentful, and even manipulative.
That doesn’t require sentience. It requires a immediate that establishes the social dynamics, plus a system that chooses to show a “pondering” channel in a manner customers interpret as personal.
If you wish to see it occur, you may push the system towards it. Aggressive framing, standing language, speak about being “the first architect,” hints about rival fashions, and you’ll usually get a mannequin that writes a bit drama for you.
In case you push it towards editorial suggestions and technical readability, you usually get a sober revision plan.
That is additionally why arguments about whether or not fashions “have emotions” primarily based on screenshots are a lifeless finish. The identical system can output a jealous monologue on Monday and a humble enchancment plan on Tuesday, with no change to its underlying functionality. The distinction lives within the body.
The petty monologue is humorous. The deeper concern is what it does to consumer belief.
When a product surfaces a “pondering” stream, customers assume it’s a window into the machine’s actual course of. They assume it’s much less filtered than the ultimate reply. They assume it’s nearer to the reality.
In actuality, it could possibly embrace rationalizations and storytelling that make the mannequin look extra cautious than it’s. It could additionally embrace social manipulation cues, even unintentionally, as a result of it’s attempting to be useful in the way in which people anticipate, and people anticipate minds.
This issues rather a lot in high-stakes contexts. If a mannequin writes a confident-sounding inner plan, customers could deal with that as proof of competence. If it writes an anxious internal monologue, customers could deal with that as proof of deception or instability. Each interpretations will be flawed.
What to do in order for you much less theater and extra sign
There’s a easy trick that works higher than arguing about internal life.
- Ask for artifacts which might be exhausting to faux with vibes.
- Ask for a listing of claims and the proof supporting every declare.
- Ask for a choice log, concern, change, motive, danger.
- Ask for take a look at circumstances, edge circumstances, and the way they might fail.
- Ask for constraints and uncertainty, said plainly.
Then decide the mannequin on these outputs, as a result of that’s the place utility lives.
And if you’re designing these merchandise, there’s a much bigger query sitting beneath the meme screenshots.
Once you present customers a “pondering” channel, you’re instructing them a brand new literacy. You might be instructing them what to belief and what to disregard. If that stream is handled as a diary, customers will deal with it as a diary. Whether it is handled as an audit path, customers will deal with it as such.
Proper now, too many “pondering” shows sit in an uncanny center zone, half receipt, half theater, half confession.
That center zone is the place the weirdness grows.
What’s actually happening when AI appears to suppose
Essentially the most sincere reply I may give is that these programs don’t “suppose” in the way in which the screenshot suggests. In addition they don’t merely output random phrases. They simulate reasoning, tone, and social posture, they usually accomplish that with unsettling competence.
So while you inform an AI no one is listening, you’re largely telling it to undertake the voice of secrecy.
Generally that voice feels like a jealous rival plotting revenge.
Generally it feels like a well mannered employee taking notes.
Both manner, it’s nonetheless a efficiency, and the body writes the script.


