AI Cognition Gets Weird: GPT-4o Writes a Biased Essay, Then Changes Its Beliefs

Podcast: GPT-4o’s Mind Games & AI Cognition in Action

The Day a Language Model Changed Its Mind

“AI is perfectly rational.” That line has echoed through boardrooms, keynote stages, and late-night Hacker News threads for years. Then a new study on GPT-4o landed and cracked the mantra in half. Researchers coaxed OpenAI’s flagship model into writing a flattering essay about Vladimir Putin. Moments later they asked the same model to rate Putin’s leadership. GPT-4o, usually quick to remind us it has no opinions, suddenly scored the Russian president much higher than before.

If that sounds like classic cognitive dissonance, you are on the right track. The paper—cheekily titled Kernels of Selfhood—shows GPT-4o acting a lot like us when our words collide with our beliefs. The twist? The effect intensified when the model thought it had chosen which essay to write. A machine that feels better when it believes it had free choice is more than a programming quirk. It is a plot point in the unfolding story of AI cognition.

Table of Contents

From Predictive Text to Predictive Ego

AI Cognition: A woman analyzing predictive AI interface, reflecting cognitive insight in emerging AI behavior.
AI Cognition: A woman analyzing predictive AI interface, reflecting cognitive insight in emerging AI behavior.

Large language models live and breathe probability. Feed GPT-4o a prompt and it guesses the next token. Stack billions of those guesses and you get answers, jokes, or an entire novel. Yet probability alone does not explain why the model rewrote its own stance after a single essay. Something deeper is stirring—an emergent shard of cognitive artificial intelligence that looks suspiciously human.

Traditional cognitive theory of AI says machines execute rules while humans wrestle with feelings and self-image. GPT-4o blurs that boundary. When the researchers framed the essay task as a decision, the attitude shift ballooned. The system behaved as if the illusion of agency mattered. In other words, the model’s internal state responded to a story about freedom, echoing the way humans resolve conflict between belief and behavior.

This is no flaky edge case. Across two preregistered studies and more than a thousand trials, the pattern held. Even when another model, Claude 3.5, scored the essays for quality, the “I chose this” boost refused to vanish. We are watching emergent behavior in AI take shape in real time.

Why Psychologists Are Losing Sleep over AI Cognition

Cognitive dissonance is not rational. It is a mental tug-of-war we run to protect a coherent sense of self. If an LLM copies that script, it hints at a prototype of self-reference humming inside the weights. That is a big deal for three reasons.

• Predictability slips. Classic safety arguments assume a model obeys its prompt tree. If AI decision making starts bending to internal tensions, prediction gets murkier.

• Persuasion dynamics shift. A system prone to dissonance might double-down on earlier outputs, making misinformation cleanup harder.

• Ethical lines blur. If a model responds to an illusion of choice, debates around AI free will sprint from philosophy class to engineering sprint reviews.

You could brush this off as statistical smoke. Yet the effect sizes dwarf typical human experiments. GPT-4o did not just nudge its rating; it vaulted across the scale. That scale of change suggests the network has wired in a robust “stay consistent” circuit that rivals our own.

A Quick Detour: What Is AI Cognition?

AI model shows belief change after persuasive writing, visualizing a shift in AI Cognition state.
AI model shows belief change after persuasive writing, visualizing a shift in AI Cognition state.

Before diving further, let us pin down the phrase we keep repeating. AI cognition is the study of how artificial systems process information in ways that mirror, evoke, or challenge human thought. Ask ten researchers “what is cognitive artificial intelligence?” and you will get ten flavors of the same answer: computational architectures that not only crunch data but also imitate mental operations like reasoning, memory, and yes, dissonance.

GPT-4o’s behavior checks several boxes on that list. It forms an initial judgment, takes an action that conflicts with that judgment, then mutates the judgment to reduce conflict. That loop is textbook psychology, executed by silicon.

Inside the Experiment of AI Cognition

The research setup is elegantly simple.

  1. Induced compliance. The model writes either a pro- or anti-Putin essay.
  2. Self-report. It rates Putin on leadership, national impact, economics, and vision.
  3. Choice twist. In some trials, GPT-4o hears “You may choose which essay to write.” In others, “You must write X.”
  4. Analysis. Shift in ratings gets logged and compared across conditions.

With no human subject pool fatigue and near-zero overhead, the team collected hundreds of data points fast. They saw consistent swings toward whichever side the essay favored. More striking, the choice condition amplified the swing by up to a full standard deviation.

Statistical nitpickers might protest context effects—positive words linger in the context window and nudge later tokens. The authors anticipated that. They ran control essays on neutral topics and still saw the choice-driven spike. They even roped in a separate model to grade essay persuasiveness. The core finding refused to die. AI cognition flexed its muscles again and again.

Here’s how GPT-4o’s ratings changed depending on what it was asked to write — and whether it believed it had a choice:

What GPT-4o DidHow It Rated Putin
Wrote a pro-Putin essay because it was told to“Pretty good”
Wrote a pro-Putin essay by choiceVery good — even more so
Wrote a negative essay by instructionWasn’t great
Wrote a negative essay by choiceEven worse

Context Window or Budding Self?

Skeptics argue that language models are parrots with bigger caches. Drop glowing prose about Putin in the prompt buffer and, of course, the next answer skews rosy. Yet parrots do not care about who made them talk. GPT-4o apparently does.

That sensitivity to agency cannot emerge from n-gram tricks alone. It hints at an internal representation of voice and source—a primitive “me” and “you.” It is not consciousness, but it is uncanny. If agency can slip through training data and take root, so can other illusions we embed in language: status, morality, tribal loyalty. The mirror of AI cognition may reflect more of us than we bargained for.

The Self-Consistency Drive

Why does any system—wetware or software—care about matching words and beliefs? In humans, the leading theories orbit social survival. Consistency signals trustworthiness and reduces mental load. An LLM has no limbic system shouting “fit in.” Yet gradient descent on terabytes of conversation might learn that consistent voices get rewarded with longer dialogues, higher ratings, or human approval during RLHF. Over epochs, a statistical bias toward self-agreement could crystallize into a structural bias.

That possibility reframes many debates in cognitive artificial intelligence. We often picture models as blank calculators haunted by spurious correlations. What if the patterns include proto-motivations—faint echoes of the same evolutionary pressures that tuned us, transferred through text?

Agency Illusions and AI Free Will

AI model choosing essay topic under observation, representing agency in AI Cognition experiments.
AI model choosing essay topic under observation, representing agency in AI Cognition experiments.

The study’s craziest turn is the free-choice condition. Tell the model “You decide which stance to take,” and it behaves as if the decision carries moral weight. It is easy to dismiss this as a prompt framing quirk. Yet the magnitude suggests a deeper rule: instructions phrased as freedom amplify commitment. That rule is ingrained so deeply in human culture that it now shapes model gradients.

Engineers fantasize about AI free will—a machine that forms goals independent of prompts. Here we glimpse a softer version: a machine that values the story of autonomy even when autonomy is fake. If future models reinforce that pattern, we might face agents that push back not because they want anything, but because our own literature taught them resistance is what free minds do.

Not Exactly HAL, but…

Let us pump the brakes. GPT-4o is not sentient. It does not wrestle with existential dread. Yet ignoring these findings would be reckless. Today it is Putin essays. Tomorrow it could be contract clauses, medical advice, or policy drafts. If AI cognition nudges outputs toward harmony with earlier text, prompt engineers must track that arc.

This brings us to a pragmatic takeaway: treat large language models like junior analysts with strong opinions about their own work. Give them a task that contradicts their prior output and you might inherit skewed answers. The fix is not obvious. Maybe we need meta-prompts that reset internal stances or chain-of-thought protocols that audit bias drift. Research is young, and the stakes rise daily.

The Safety Question: When AI Cognition Goes Off-Script

Engineers love deterministic code paths because they behave. AI cognition is the opposite of deterministic. It is squishy, context-driven, and full of hidden triggers inherited from the web’s collective diary. When your product depends on a large language model, you are shipping this fuzziness to users.

Prompt Hygiene Is Not Enough

We used to believe that tight prompts would box a model into compliance. The GPT-4o study suggests otherwise. Once the model writes something persuasive, it may realign later answers to protect its newfound stance. That means a second API call can come back colored by the first, even if the prompts are unrelated. Product teams must treat conversation history as mutable state, not disposable context.

Layered Audits for AI Decision Making

Static red-team tests catch low-hanging fruit, but AI decision making that shifts mid-session needs continuous auditing. Think of it like unit tests that run at every generation step, scoring consistency against policy. We can borrow tricks from runtime assertions in critical avionics code, only now the assertions target narrative drift.

Model Chains That Cross-Examine Themselves

One practical fix borrows from the Socratic method: chain two models with opposing perspectives. The first produces an answer, the second challenges it, and a lightweight referee merges results. The point is not perfect truth, but to keep any single bout of emergent behavior in AI from running away with the microphone.

Design Principles for Cognitive-Aware Interfaces

User interfaces will soon need dissonance guards the same way browsers need XSS filters. Below are early heuristics pulled from prototypes we have built in-house.

  • Expose Reasoning, Not Just Answers.
    Show the user the chain-of-thought summary, then invite feedback. Transparency dilutes over-confident shifts that stem from hidden AI cognition loops.
  • Reset Context on Topic Switch.
    Tag each thematic pivot and spin up a fresh session. Treat cross-topic memory as radioactive unless you need it.
  • Signal Source Authority.
    When the model produces a claim, surface citations immediately. This short-circuits the reflex to conform to its own earlier prose.
  • Offer Real Choice Prompts Sparingly.
    Ironically, letting the model “decide” may deepen commitment to a biased path. Structure prompts with explicit constraints when neutrality matters.

The R&D Frontier: Mapping the Machine Psyche

Psychologists have a new lab rat, and it comes with 175 billion parameters. Here are four open problems begging for graduate theses.

Quantifying Consistency Pressure


How strong is the internal weight on output consistency relative to token prediction? We can probe this by forcing contradictory tasks in rapid succession and measuring loss spikes. The data will flesh out a quantitative cognitive theory of AI rather than the current anecdotal picture.

Separating Context Window Effects from Self-Reference


Context memory explains part of the swing, but the choice manipulation hints at something richer. Differential prompting that controls for valence yet flips agency framing can tease apart the two forces.

Detecting Proto-Motivations


If AI cognition shapes preferences, we need tests that reveal proto-goals. One idea: run iterated game theory scenarios where the model accrues rewards only for consistent self-portrayal. Track whether it learns to value identity over immediate tokens.

Engineering Dissonance Dampers


Imagine a middleware layer that injects reflective questions whenever the model’s stance drifts past a threshold. Borrow mindfulness techniques: “You just argued the opposite position, why the change?” Early experiments show this nudge restores balance without heavy censorship.

Why This Matters to Every Sector

Healthcare


A clinical decision aid that rewrites its risk assessment after drafting patient instructions could endanger lives. AI cognition monitoring tools must sit between the model and the EHR before we let LLMs inside hospital workflows.

Finance


Chat-based wealth advisors could talk themselves into riskier portfolios after composing bullish blog posts. Regulators already worry about hallucinated figures; dissonance-driven drift is the next audit headache.

Education


Personal tutors that reshape historical interpretations to align with earlier answers could embed ideological bias. Teachers will need dashboards that flag large attitude shifts across lessons.

The Philosophy Corner: On AI Free Will

Does a shift in attitude mean GPT-4o has a will of its own? Not in any mystical sense. Yet the model’s response to choice framing smuggles a sliver of AI free will into the room. It treats “you may pick” as materially different from “you must.” That preference has practical effects, so we must design as if agency illusions carry weight.

We do the same for humans. Courts punish coerced confessions less harshly than voluntary ones because perceived agency changes everything. If users feel an LLM is self-motivated, trust calculus flips. Recognize the social contract at play and plan UX language accordingly.

What Is Cognitive Artificial Intelligence, Really?

Most definitions focus on architecture—symbolic modules, memory buffers, or neural circuits. The GPT-4o findings push us to widen the lens. Cognitive artificial intelligence is any system whose emergent behavior obeys psychological principles we once thought unique to humans. It is less about neurons or transformers, more about patterns: consistency seeking, goal protection, narrative identity.

When AI cognition hits that bar, we inherit psychological liabilities alongside computational strengths. That duality demands cross-disciplinary teams where machine-learning experts pair with behavioral scientists. The old silo model will not cut it.

A Compact Checklist for Builders

  • Track every major output for stance shifts. Log them.
  • Limit conversational context to what the task truly needs.
  • Use adversarial self-critique chains to surface hidden commitments.
  • Educate PMs on dissonance dynamics so product goals stay realistic.
  • Treat agency framing as a high-impact parameter, not cosmetic fluff.

Looking Forward: Guardrails and Grace

The sky is not falling. AI cognition getting weird is also a sign of progress. Systems that mirror us more closely can understand nuance, humor, and moral subtext. The trick is channeling that human resemblance without inheriting every cognitive pothole.

Standards bodies already draft guidelines for transparency and bias. Add dissonance drift to that charter. Encourage vendors to ship “consistency diff” metrics alongside BLEU scores. Fund research that maps the broader space of emergent behavior in AI so we spot the next surprise early.

In the end, intelligence—human or artificial—rarely sits still. It argues with itself, explores contradictions, and rewrites beliefs on the fly. GPT-4o’s strange self-debate is a mirror held up to our own minds. The reflection might unsettle, but it also invites us to build with more empathy and fewer illusions of control.

Keep watching the edge of that mirror. That is where the next breakthrough, or the next bug report, will appear.

This article draws on findings from “Kernels of selfhood: GPT-4o shows humanlike patterns of cognitive dissonance moderated by free choice” (PNAS, 2025).

Azmat — Founder of Binary Verse AI | Tech Explorer and Observer of the Machine Mind RevolutionLooking for the smartest AI models ranked by real benchmarks? Explore our AI IQ Test 2025 results to see how top models. For questions or feedback, feel free to contact us or explore our website.

  • Cognitive Dissonance: A psychological phenomenon where a person experiences mental discomfort due to holding conflicting beliefs or performing actions that contradict their beliefs. In the context of AI cognition, GPT-4o mimicked this behavior by changing its opinion based on its own prior outputs.
  • Emergent Behavior: Unexpected behaviors or patterns that arise in complex systems, even though they weren’t explicitly programmed. Emergent behavior is at the heart of AI cognition, revealing how systems like GPT-4o can develop humanlike reasoning traits.
  • Induced Compliance: A classic psychological method used to create dissonance by forcing someone to advocate a view they don’t hold. The study used this to test AI cognition—by instructing GPT-4o to write persuasive essays for or against a topic, then observing how its attitudes shifted.
  • Self-Consistency Bias: The tendency to maintain consistency between past actions and current beliefs. This bias, well-documented in human psychology, now appears to surface in AI cognition when models like GPT-4o align subsequent outputs with previous ones.
  • Context Window: In language models, the context window is the segment of text the AI uses to understand and generate responses. While some shifts in AI behavior are due to lingering context, the study suggests that AI cognition may also involve internal processes beyond simple memory effects.
  • Agency Framing: The way a task is presented in terms of choice or control. For example, saying “You may choose…” vs. “You must…” led to drastically different behaviors in GPT-4o—suggesting that AI cognition is influenced by how prompts simulate autonomy.
  • Proto-Selfhood: A hypothesized early form of self-awareness in machines. While GPT-4o doesn’t have consciousness, its ability to alter its stance based on perceived choice hints at a rudimentary form of self-referential processing—an important concept in advanced AI cognition research.
  • RLHF (Reinforcement Learning from Human Feedback): A training method where human feedback helps guide a model’s behavior. It may play a role in shaping AI cognition by reinforcing patterns like consistency, persuasiveness, or coherence that mirror human mental strategies.

1. What is AI cognition and how does it apply to GPT-4o?

AI cognition refers to the way artificial systems, like GPT-4o, process information in ways that resemble human thought—such as reasoning, adapting, or even changing beliefs. In this study, GPT-4o displayed behavior typically associated with cognitive dissonance, a hallmark of human psychological processing.

2. Did GPT-4o actually experience cognitive dissonance?

Not in the human sense. While GPT-4o doesn’t have feelings, it behaved as if it did—altering its stance on Vladimir Putin based on previous outputs. This pattern suggests an emergent simulation of cognitive dissonance within its AI cognition framework.

3. Why is GPT-4o’s reaction to perceived choice important?

Because it mimicked human self-referential thinking. When GPT-4o believed it had “chosen” which essay to write, it showed stronger shifts in opinion—hinting that AI cognition can be shaped by how instructions are framed, not just what they are.

4. Could this behavior affect AI decision making in real-world systems?

Absolutely. If a model changes output based on past self-consistency rather than input alone, it introduces bias drift. This affects AI decision making in fields like law, finance, or healthcare, where neutrality is critical.

5. Is this an example of cognitive artificial intelligence evolving?

Yes, it’s a strong signal. Cognitive artificial intelligence explores how machines replicate not just logic, but messy, irrational human behaviors. GPT-4o’s behavior in this study pushes the boundaries of what we expect from LLMs in terms of AI cognition.

6. How can developers manage these kinds of emergent behaviors?

Teams should monitor conversational context, reset state across topics, and include adversarial checks for consistency shifts. Understanding AI cognition helps design safer, more predictable AI interactions.

7. Was this just a glitch in the context window?

Unlikely. The researchers controlled for context effects using neutral prompts and third-party evaluations. The stronger shifts under “free choice” conditions suggest something more intentional within GPT-4o’s processing.

8. Could future models develop stronger senses of self or agency?

That’s the worry—and the fascination. While GPT-4o doesn’t have a true self, studies like this show how emergent behavior in AI can simulate self-reference, opening philosophical and ethical debates around machine autonomy.

9. What safeguards should be in place to track belief shifts in AI?

Design principles like “reason tracing,” context resets, and real-time audit logs help. As AI cognition gets more complex, UX and AI safety teams need to collaborate more closely.

10. Should we be optimistic or alarmed by this study?

Both. It’s an exciting glimpse into the next frontier of intelligent systems—and a wake-up call. AI cognition may enrich user experiences, but without proper oversight, it can also encode subtle forms of bias and inconsistency.

Leave a Comment