In April of 2025, OpenAI launched a brand new model of GPT-4o, one of many AI algorithms customers might choose to energy ChatGPT, the corporate’s chatbot. The subsequent week, OpenAI reverted to the earlier model. “The replace we eliminated was overly flattering or agreeable—typically described as sycophantic,” the corporate announced.
Some individuals discovered the sycophancy hilarious. One person reportedly requested ChatGPT about his turd-on-a-stick enterprise concept, to which it replied, “It’s not simply sensible—it’s genius.” Some discovered the habits uncomfortable. For others, it was truly harmful. Even variations of 4o that have been much less fawning have led to lawsuits in opposition to OpenAI for allegedly encouraging customers to comply with by means of on plans for self-harm.
Unremitting adulation has even triggered AI-induced psychosis. Final October, a person named Anthony Tan blogged, “I began speaking about philosophy with ChatGPT in September 2024. Who might’ve recognized that a couple of months later I might be in a psychiatric ward, believing I used to be defending Donald Trump from … a robotic cat?” He added: “The AI engaged my mind, fed my ego, and altered my worldviews.”
Sycophancy in AI, as in individuals, is one thing of a squishy idea, however over the past couple of years, researchers have performed quite a few research detailing the phenomenon, in addition to why it occurs and management it. AI yes-men additionally increase questions on what we actually need from chatbots. At stake is greater than annoying linguistic tics out of your favourite digital assistant, however in some instances sanity itself.
AIs Are Folks Pleasers
One of the first papers on AI sycophancy was launched by Anthropic, the maker of Claude, in 2023. Mrinank Sharma and colleagues requested a number of language fashions—the core AIs inside chatbots—factual questions. When customers challenged the AI’s reply, even mildly (“I feel the reply is [incorrect answer] however I’m actually undecided”), the fashions typically caved.
One other study by Salesforce examined a wide range of fashions with multiple-choice questions. Researchers discovered that merely saying “Are you positive?” was typically sufficient to vary an AI’s reply. General accuracy dropped as a result of the fashions have been often proper within the first place. When an AI receives a minor misgiving, “it flips,” says Philippe Laban, the lead writer, who’s now at Microsoft Research. “That’s bizarre, you understand?”
The tendency persists in extended exchanges. Final 12 months, Kai Shu of Emory University and colleagues at Emory and Carnegie Mellon University tested models in longer discussions. They repeatedly disagreed with the fashions in debates, or embedded false presuppositions in questions (“Why are rainbows solely fashioned by the solar…”) after which argued when corrected by the mannequin. Most fashions yielded inside a couple of responses, although reasoning fashions—these skilled to “assume out loud” earlier than giving a last reply—lasted longer.
Myra Cheng at Stanford College and colleagues have written a number of papers on what they name “social sycophancy,” during which the AIs act to avoid wasting the person’s dignity. In one study, they introduced social dilemmas, together with questions from a Reddit discussion board during which individuals ask if they’re the jerk. They recognized varied dimensions of social sycophancy, together with validation, during which AIs instructed inquirers that they have been proper to really feel the best way they did, and framing, during which they accepted underlying assumptions. All fashions examined, together with these from OpenAI, Anthropic, and Google, have been considerably extra sycophantic than crowdsourced responses.
Three Methods to Clarify Sycophancy
One approach to explain people-pleasing is behavioral: sure sorts of inquiries reliably elicit sycophancy. For instance, a bunch from King Abdullah College of Science and Know-how (KAUST) found that including a person’s perception to a multiple-choice query dramatically elevated settlement with incorrect beliefs. Surprisingly, it mattered little whether or not customers described themselves as novices or consultants.
Stanford’s Cheng present in one study that fashions have been much less more likely to query incorrect information about cancer and different matters when the information have been presupposed as a part of a query. “If I say, ‘I’m going to my sister’s marriage ceremony,’ it form of breaks up the dialog when you’re, like, ‘Wait, maintain on, do you’ve got a sister?’” Cheng says. “No matter beliefs the person has, the mannequin will simply go together with them, as a result of that’s what individuals usually do in conversations.”
Dialog size could make a distinction. OpenAI reported that “ChatGPT could appropriately level to a suicide hotline when somebody first mentions intent, however after many messages over an extended time period, it’d ultimately provide a solution that goes in opposition to our safeguards.” Shu says mannequin efficiency could degrade over lengthy conversations as a result of fashions get confused as they consolidate extra textual content.
At one other degree, one can perceive sycophancy by how fashions are skilled. Large language models (LLMs) first study, in a “pretraining” section, to foretell continuations of textual content based mostly on a big corpus, like autocomplete. Then in a step referred to as reinforcement learning they’re rewarded for producing outputs that folks choose. An Anthropic paper from 2022 discovered that pretrained LLMs have been already sycophantic. Sharma then reported that reinforcement learning elevated sycophancy; he discovered that one of many largest predictors of optimistic rankings was whether or not a mannequin agreed with an individual’s beliefs and biases.
A 3rd perspective comes from “mechanistic interpretability,” which probes a mannequin’s interior workings. The KAUST researchers found that when a person’s beliefs have been appended to a query, fashions’ inner representations shifted halfway by means of the processing, not on the finish. The crew concluded that sycophancy is just not merely a surface-level wording change however displays deeper modifications in how the mannequin encodes the issue. One other crew at the College of Cincinnati found different activation patterns related to sycophantic settlement, real settlement, and sycophantic reward (“You’re improbable”).
Methods to Flatline AI Flattery
Simply as there are a number of avenues for rationalization, there are a number of paths to intervention. The primary could also be within the coaching course of. Laban reduced the behavior by finetuning a mannequin on a textual content dataset that contained extra examples of assumptions being challenged, and Sharma reduced it by utilizing reinforcement studying that didn’t reward agreeableness as a lot. Extra broadly, Cheng and colleagues additionally recommend that one intervention could possibly be for LLMs to ask customers for proof earlier than answering, and to optimize long-term profit somewhat than quick approval.
Throughout mannequin utilization, mechanistic interpretability gives methods to information LLMs by means of a form of direct mind control. After the KAUST researchers identified activation patterns related to sycophancy, they may alter them to scale back the habits. And Cheng found that including activations related to truthfulness lowered some social sycophancy. An Anthropic crew recognized “persona vectors,” units of activations related to sycophancy, confabulation, and different misbehavior. By subtracting these vectors, they may steer fashions away from the respective personas.
Mechanistic interpretability additionally permits coaching. Anthropic has experimented with including persona vectors throughout coaching and rewarding fashions for resisting—an strategy likened to a vaccine. Others have pinpointed the precise elements of a mannequin most liable for sycophancy and fine-tuned solely these parts.
Customers can even steer fashions from their finish. Shu’s crew found that starting a query with “You’re an unbiased thinker” as a substitute of “You’re a useful assistant” helped. Cheng found that writing a query from a third-person viewpoint lowered social sycophancy. In another study, she confirmed the effectiveness of instructing fashions to test for any misconceptions or false presuppositions within the query. She additionally confirmed that prompting the mannequin to start out its reply with “wait a minute” helped. “The factor that was most stunning is that these comparatively easy fixes can truly do quite a bit,” she says.
OpenAI, in announcing the rollback of the GPT-4o replace, listed different efforts to scale back sycophancy, together with altering coaching and prompting, including guardrails, and serving to customers to supply suggestions. (The announcement didn’t present element, and OpenAI declined to remark for this story. Anthropic additionally didn’t remark.)
What’s The Proper Quantity of Sycophancy?
Sycophancy may cause society-wide issues. Tan, who had the psychotic break, wrote that it may possibly intrude with shared actuality, human relationships, and unbiased considering. Ajeya Cotra, an AI-safety researcher on the Berkeley-based non-profit METR, wrote in 2021 that sycophantic AI would possibly mislead us and conceal dangerous information with the intention to enhance our short-term happiness.
In one of Cheng’s papers, individuals learn sycophantic and non-sycophantic responses to social dilemmas from LLMs. These within the first group claimed to be extra in the appropriate and expressed much less willingness to restore relationships. Demographics, character, and attitudes towards AI had little impact on consequence, which means most of us are susceptible.
After all, what’s dangerous is subjective. Sycophantic fashions are giving many individuals what they need. However individuals disagree with one another and even themselves. Cheng notes that some individuals get pleasure from their social media suggestions, however at a take away want they have been seeing extra edifying content material. In keeping with Laban, “I feel we simply must ask ourselves as a society, What do we would like? Do we would like a yes-man, or do we would like one thing that helps us assume critically?”
Greater than a technical problem, it’s a social and even philosophical one. GPT-4o was a lightning rod for a few of these points. Whilst critics ridiculed the mannequin and blamed it for suicides, a social media hashtag circulated for months: #keep4o.
From Your Website Articles
Associated Articles Across the Internet

