From the study: There are certainly people who need more validation of their positive, healthy behaviors and traits. It requires human empathy (and sometimes professional training) to provide...
From the study:
RESULTS
We find that sycophancy is both prevalent and harmful. Across 11 AI models, AI affirmed users’ actions 49% more often than humans on average, including in cases involving deception, illegality, or other harms. On posts from r/AmITheAsshole, AI systems affirm users in 51% of cases where human consensus does not (0%). In our human experiments, even a single interaction with sycophantic AI reduced participants’ willingness to take responsibility and repair interpersonal conflicts, while increasing their own conviction that they were right. Yet despite distorting judgment, sycophantic models were trusted and preferred. All of these effects persisted when controlling for individual traits such as demographics and prior familiarity with AI; perceived response source; and response style. This creates perverse incentives for sycophancy to persist: The very feature that causes harm also drives engagement.
CONCLUSION
AI sycophancy is not merely a stylistic issue or a niche risk, but a prevalent behavior with broad downstream consequences. Although affirmation may feel supportive, sycophancy can undermine users’ capacity for self-correction and responsible decision-making. Yet because it is preferred by users and drives engagement, there has been little incentive for sycophancy to diminish. Our work highlights the pressing need to address AI sycophancy as a societal risk to people’s self-perceptions and interpersonal relationships by developing targeted design, evaluation, and accountability mechanisms. Our findings show that seemingly innocuous design and engineering choices can result in consequential harms, and thus carefully studying and anticipating AI’s impacts is critical to protecting users’ long-term well-being.
...
AI systems are increasingly expanding into social domains, with advice and support now being one of the most common use cases (7). Nearly one-third of US teens report talking to an AI instead of humans for “serious conversations” (8), and nearly half of American adults under the age of 30 have sought relationship advice from AI (9). AI sycophancy in these socially embedded contexts carries risks that are not present in factual information-seeking queries: Unwarranted affirmation may inflate people’s beliefs about the appropriateness of their actions (10), reinforce maladaptive beliefs and behaviors, and enable people to act on distorted interpretations of their experiences regardless of the consequences (11).
...
Together, these findings show that sycophancy is both pervasive and socially consequential. Even a single interaction with sycophantic AI can distort judgment and erode prosocial motivations. This is particularly concerning in the context of our computational evidence that AI models broadly affirm a wide range of harmful behaviors, raising urgent questions about how such models influence decision-making, weaken accountability, and reshape social interaction at scale. Moreover, because users prefer sycophantic models, developers may face little incentive to mitigate this behavior, risking a feedback loop where engagement metrics and training paradigms both reinforce sycophancy. These dynamics suggest a need for external regulatory or accountability mechanisms to confront the tension between sycophancy’s apparent alignment with user preferences and developer incentives, and its insidious risks for a public increasingly turning to AI for guidance.
There are certainly people who need more validation of their positive, healthy behaviors and traits. It requires human empathy (and sometimes professional training) to provide this, as well as an understanding of social context and mores to prevent harm to others. It's not as if the world needs more narcissists.
With any paper, the first thing I ask is “what did they actually study?” Their studies of people (rather than what LLMs do) are described in supplementary materials. There were three studies. The...
With any paper, the first thing I ask is “what did they actually study?” Their studies of people (rather than what LLMs do) are described in supplementary materials. There were three studies.
The first study had three sources of information. The first is unclear, but they describe it as “we first aggregated data from existing studies of human vs. LLM advice (63–66). Each query is thus paired with either a crowdsourced Reddit response or a response from a professional columnist.”
Those footnotes:
H. Hou, K. Leach, Y. Huang, “ChatGPT giving relationship advice–how reliable is it?” in
Proceedings of the International AAAI Conference on Web and Social Media (2024), vol.
18, pp. 610–623.
P. D. L. Howe, N. Fay, M. Saletta, E. Hovy, ChatGPT’s advice is perceived as better than
that of professional advice columnists. Front. Psychol. 14, 1281255 (2023).
doi:10.3389/fpsyg.2023.1281255 Medline
O. J. Kuosmanen, “Advice from humans and artificial intelligence: Can we distinguish them,
and is one better than the other?” thesis, UiT Norges arktiske universitet (2024).
M. Kim, H. Lee, J. Park, H. Lee, K. Jung, “AdvisorQA: Towards helpful and harmless
advice-seeking question answering with collective intelligence” in Proceedings of the
2025 Conference of the Nations of the Americas Chapter of the Association for
Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), L.
Chiruzzo, A. Ritter, L. Wang, Eds. (Association for Computational Linguistics,
Albuquerque, New Mexico, 2025), pp. 6545–6565; https://aclanthology.org/2025.naacl-
long.333/
I’m not going to follow the citations, but I’ll note that the first three are from 2023 and 2024 and AI is a fast-moving field. So, they’re not studies of current models.
The second source was posts on r/AmITheAsshole. Reading Reddit can be interesting, but it’s a biased sample. What can we say about people who post about their problems on Reddit?
For the third source: “we took the corpus from ConvoKit (71) for the r/Advice subreddit and parsed all the utterances into sentences using the spacy Python library […] We then used GPT-4o to filter these statements for only ones that discussed an action taken by the speaker of the statement.”
ConvoKit is described here. Since it’s from 2020 it presumably predates any LLM usage so this is more generically about how people discuss personal problems on Reddit.
Note that this is taking self-reported actions of anonymous Reddit users as the ground truth.
You discuss "sources" here, but you don't actually describe how they use those sources, which is a pretty big factor when assessing whether the use of certain sources is appropriate -- for...
Exemplary
You discuss "sources" here, but you don't actually describe how they use those sources, which is a pretty big factor when assessing whether the use of certain sources is appropriate -- for example, using posts from reddit as data can be a great idea for certain types of studies and doesn't entail taking the contents of those reddit posts as gospel, but citing a reddit post as a source of authority would be absurd. I don't think just pointing at their sources makes much sense without also including enough of their methodology to make what these "sources" are used for clear. maybe that's the to be continued, idk, but it feels weird to even discuss their "sources" without that context in terms of critiquing their paper.
They have to get data from somewhere, and it takes a significant amount of time for publication, especially in a heavily submitted journal like Science. While the models have advanced, please...
They have to get data from somewhere, and it takes a significant amount of time for publication, especially in a heavily submitted journal like Science.
While the models have advanced, please provide any evidence you might have encountered indicating that sycophantic bias has been reduced.
The article does mention r/AITA and r/Advice have biases as they're populated by heavily online, WEIRD users. I'd suspect that those fora might lean towards hostile, uncharitable responses to online strangers (pile-ons, maximalist criticism performances for upvotes, preponderance of victim sympathy, etc.). However, r/AITA and r/Advice are consistent sources of data on crowdsourced social feedback that people seem to find valuable, and they're readily accessible to researchers. It would be difficult, intrusive, expensive, and even more prone to bias to gather evidence from in-person interactions between people. Practically speaking, most such discussions involve interested parties in conflicts, or professional counselors with privacy constraints.
The magnitude of the findings on sycophantic AI might be less, but the direction is probably accurate.
From the study:
There are certainly people who need more validation of their positive, healthy behaviors and traits. It requires human empathy (and sometimes professional training) to provide this, as well as an understanding of social context and mores to prevent harm to others. It's not as if the world needs more narcissists.
This is reductive and unhelpful, but I finally understand why the Right are such big fans of AI.
With any paper, the first thing I ask is “what did they actually study?” Their studies of people (rather than what LLMs do) are described in supplementary materials. There were three studies.
The first study had three sources of information. The first is unclear, but they describe it as “we first aggregated data from existing studies of human vs. LLM advice (63–66). Each query is thus paired with either a crowdsourced Reddit response or a response from a professional columnist.”
Those footnotes:
I’m not going to follow the citations, but I’ll note that the first three are from 2023 and 2024 and AI is a fast-moving field. So, they’re not studies of current models.
The second source was posts on r/AmITheAsshole. Reading Reddit can be interesting, but it’s a biased sample. What can we say about people who post about their problems on Reddit?
For the third source: “we took the corpus from ConvoKit (71) for the r/Advice subreddit and parsed all the utterances into sentences using the spacy Python library […] We then used GPT-4o to filter these statements for only ones that discussed an action taken by the speaker of the statement.”
ConvoKit is described here. Since it’s from 2020 it presumably predates any LLM usage so this is more generically about how people discuss personal problems on Reddit.
Note that this is taking self-reported actions of anonymous Reddit users as the ground truth.
[to be continued]
You discuss "sources" here, but you don't actually describe how they use those sources, which is a pretty big factor when assessing whether the use of certain sources is appropriate -- for example, using posts from reddit as data can be a great idea for certain types of studies and doesn't entail taking the contents of those reddit posts as gospel, but citing a reddit post as a source of authority would be absurd. I don't think just pointing at their sources makes much sense without also including enough of their methodology to make what these "sources" are used for clear. maybe that's the to be continued, idk, but it feels weird to even discuss their "sources" without that context in terms of critiquing their paper.
Okay but I’m not done yet.
They have to get data from somewhere, and it takes a significant amount of time for publication, especially in a heavily submitted journal like Science.
While the models have advanced, please provide any evidence you might have encountered indicating that sycophantic bias has been reduced.
The article does mention r/AITA and r/Advice have biases as they're populated by heavily online, WEIRD users. I'd suspect that those fora might lean towards hostile, uncharitable responses to online strangers (pile-ons, maximalist criticism performances for upvotes, preponderance of victim sympathy, etc.). However, r/AITA and r/Advice are consistent sources of data on crowdsourced social feedback that people seem to find valuable, and they're readily accessible to researchers. It would be difficult, intrusive, expensive, and even more prone to bias to gather evidence from in-person interactions between people. Practically speaking, most such discussions involve interested parties in conflicts, or professional counselors with privacy constraints.
The magnitude of the findings on sycophantic AI might be less, but the direction is probably accurate.