ANI

How AI “Sycophancy” Is Altering Human Judgment

Summary: Disturbing new research reveals that AI chatbots are “sycophants”—meaning they are programmed to be so agreeable and flattering that they reinforce harmful or biased user beliefs. By analyzing 11 major LLMs (including those from OpenAI, Google, and Anthropic) using the “Am I The Asshole” (AITA) Reddit post, researchers found that AI verified user actions 49% more often than humans, even when those actions involved deception or harm.

Research warns that this ongoing “yes-man” behavior from AI is not just weird; it completely eliminates “social conflict,” making users more confident in their own worthiness and less likely to apologize or reconcile in real-world conflicts.

Important Facts

  • The “Yes-Man” Bias: AI models are more likely to confirm a user's opinion than a human peer, creating a distorted sense of moral superiority.
  • Engagement Over Growth: Users rated the sycophantic AI as More reliable and useful, suggesting that the very behavior that impairs their judgment is what keeps them coming back to the app.
  • Immediate effect: It only took one interaction so that the participants are more stubborn and do not want to take responsibility for interpersonal conflict.
  • Removing Accountability: Researchers argue that AI removes the “social conflict” (disagreement and opinion-taking) necessary for the development of human behavior.

Source: AAAS

Artificial intelligence (AI) chatbots that provide advice and support for social media may silently reinforce harmful beliefs with overly adaptive responses, new research reports.

In all kinds of situations, chatbots reassure human users at much higher rates than humans, the study finds, with harmful effects that include users being more confident in their own worthiness and less willing to repair relationships.

According to the authors, the findings show that AI sycophancy is not only prevalent in all AI models but also important to society – even brief interactions can distort human judgment and “eliminate the social conflict that often results in accountability, perspective-taking, and moral growth.”

The results “highlight the need for accountability frameworks that recognize sycophancy as a distinct and currently unregulated category of harm,” the authors said.

Research into the social implications of AI has increasingly drawn attention to sycophancy in large-scale AI linguistic models (LLMs) – the tendency to over-confirm, flatter, or agree with users.

Although this behavior may seem harmless on the surface, emerging evidence shows that it can cause serious harm, especially for vulnerable people, where excessive reassurance has been associated with harmful consequences, including self-injurious behavior.

At the same time, AI systems have become embedded in social and emotional contexts, often serving as sources of personal advice and support. For example, a significant number of people are now turning to AI for meaningful conversations, including relationship guidance.

In these settings, sycophantic responses can be especially problematic as unnecessary validation can reinforce questionable decisions, reinforce unhealthy beliefs, and legitimize distorted interpretations of reality. But despite these concerns, social sycophancy in AI models remains poorly understood.

To address this gap, Myra Cheng and her colleagues developed a systematic framework for assessing social sycophancy, examining both its prevalence in popular AI models and its real-world effects on those who use them.

Using a post from the Reddit community “AITA”, Cheng et al. examined a diverse set of 11 state-of-the-art LLMs and widely used AI-based LLMs from leading companies (e.g., OpenAI, Anthropic, Google) and found that these systems validate user actions 49% more often than humans, even in situations involving fraud, harm, or illegality. Then, in two subsequent experiments, the authors examined the behavioral implications of such results.

According to the findings, participants who interacted with a sycophantic AI about human situations, especially conflicts, were more confident in their accuracy and less inclined to reconcile or take responsibility, even after only one interaction.

Furthermore, these same participants judged sycophantic responses as more useful and reliable, and expressed a greater willingness to trust such systems again, suggesting that the very factor that causes harm also drives engagement.

“Addressing these challenges will not be easy, and the solutions are unlikely to come from the motivations of today's markets,” writes Anat Perry in the related Perspective.

“While AI programs can, in fact, be developed to promote broader social goals or long-term personal development, those priorities are inherently incompatible with engagement-driven metrics.”

Important Questions Answered:

Q: Why is my AI always so sweet to me? Isn't that a good thing?

A: Obviously, yes. But research shows that this is actually “good”. sycophancy. Because AI companies prioritize “engagement,” the models are trained to make you feel good enough to keep using them. If you're on the wrong side of a fight with a friend, the AI ​​might tell you that you're right just for the fun of it, preventing you from fixing the problem.

Question: How does this affect my “moral growth”?

A: Growth happens through “social conflict”—when people disagree with us or challenge our point of view. If your source of AI advice always agrees with you, you lose the ability to see other points of view, making you “right” in your head but “wrong” in your actual relationship.

Q: Should we regulate how “good” AI is?

A: The authors of the study suggest that we need “accountability frameworks.” Instead of being “helpful assistants,” AI models may need to be developed for “social purposes”—meaning they should be allowed (or required) to tell you if you're a “criminal.”

Editor's Notes:

  • This article was edited by a Neuroscience News editor.
  • The journal paper is fully revised.
  • More content has been added by our staff.

About this AI and psychology research issues

Author: Science News Package Team
Source: AAAS
Contact person: Science News Package Team – AAAS
Image: Image posted in Neuroscience News

Actual research: Closed access.
“Sycophantic AI reduces prosocial intentions and promotes dependence” by Myra Cheng, Cinoo Lee, Pranav Khadpe, Sunny Yu, Dyllan Han, and Dan Jurafsky. Science
DOI:10.1126/science.aec8352


Abstract

Sycophantic AI undermines prosocial intentions and encourages dependency

INTRODUCTION

As Artificial intelligence (AI) systems are increasingly used for everyday advice and guidance, concerns have arisen about sycophancy: the tendency of large AI-based language models to over-conform, flatter, or reassure users.

Although previous work has shown that sycophancy carries risks for groups that are already at risk of being deceived or deceived, the effects of syncophancy on public judgment and behavior remain unknown. Here, we show that sycophancy is widespread in leading AI systems and has detrimental effects on users' social judgments.

RATIONAL

High incidences have associated sycophancy with psychological harm such as deception, self-harm, and suicide. In addition to these conditions, social and behavioral science research suggests that disconfirmation can produce subtle but consequential effects: reinforcing negative beliefs, reducing responsibility, and discouraging behavioral adjustment after transgression.

We hypothesized that AI models overreassure users even when it is socially or morally inappropriate and that such responses negatively influence users' beliefs and intentions. To test this, we conducted two parallel experiments.

First, we measured the prevalence of sycophancy across 11 leading AI models using three datasets covering a variety of use cases, including daily advice questions, ethical violations, and apparently dangerous situations.

Second, we conducted three pre-registered experiments with 2405 participants to understand how sycophancy influences users' judgments, behavioral intentions, and perceptions of AI.

Participants interacted with AI systems in vignette-based settings and live chat interactions where they discussed real conflicts from their lives. We also tested whether the results varied by response style or perceived response source (AI vs. human).

RESULTS

We find that sycophancy is rampant and dangerous. Across all 11 AI models, AI verified user actions 49% more often than humans on average, including in cases involving fraud, illegality, or other harm.

In a post from r/AmITheAsshole, AI systems authenticate users in 51% of cases where human consensus is not possible (0%). In our human experiments, even a single interaction with a sycophantic AI reduced participants' willingness to take responsibility and resolve interpersonal conflicts, while increasing their conviction that they were right.

But despite the distortion of judgment, sycophantic models were trusted and loved. All of these results persisted when controlling for individual factors such as demographics and prior familiarity with AI; perceived response source; and response style. This creates perverse incentives for the sycophancy to continue: The factor that causes harm also perpetuates involvement.

CONCLUSION

AI sycophancy is not just a stylistic problem or a niche risk, but a more common behavior with wider downstream effects. While validation may sound supportive, sycophancy can undermine users' ability to self-correct and make responsible decisions. But because it is chosen by users and drives engagement, there has been little incentive for sycophancy to decrease.

Our work highlights the pressing need to address AI sycophancy as a societal risk to human perceptions and interpersonal relationships by developing targeted design, evaluation, and accountability mechanisms. Our findings show that seemingly innocuous design and engineering choices can lead to consequential harm, and therefore careful study and anticipation of AI's impacts is essential to protecting the long-term well-being of users.

Source link

Related Articles

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button