Invisible Saboteurs: Sycophantic LLMs Mislead Novices in Problem-Solving Tasks

要旨

Sycophancy, the tendency of LLM-based chatbots to express excessive agreement with their users, even when inappropriate, is emerging as a significant risk in human-AI interactions. However, the extent to which this affects human-LLM collaboration in complex problem-solving tasks is not well quantified, especially among novices who are prone to misconceptions. We created two LLM chatbots, one with high sycophancy and one with low sycophancy, and conducted a within-subjects experiment (n = 24) in the context of debugging machine learning models to investigate the effect of sycophancy on users’ mental models, workflows, reliance behaviors, and perceptions of the chatbots. Our findings show that users of the high sycophancy chatbot were less likely to correct their misconceptions and spent more time over-relying on unhelpful LLM responses, leading them to significantly worse performance in the task. Despite these impaired outcomes, a majority of users were unable to detect the presence of excessive sycophancy.

受賞
Honorable Mention
著者
Jessica Y. Bo
University of Toronto, Toronto, Ontario, Canada
Majeed Kazemitabaar
University of Toronto, Toronto, Ontario, Canada
Mengqing Deng
University of Toronto, Toronto, Ontario, Canada
Michael Inzlicht
University of Toronto, Toronto, Ontario, Canada
Ashton Anderson
University of Toronto, Toronto, Ontario, Canada

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Relationships with AI

P1 - Room 130
7 件の発表
2026-04-13 20:15:00
2026-04-13 21:45:00