pull down to refresh

AI chatbots that butter you up make you worse at conflict, study finds
State-of-the-art AI models tend to flatter users, and that praise makes people more convinced that they're right and less willing to resolve conflicts, recent research suggests.
These models, in other words, potentially promote social and psychological harm.
Computer scientists from Stanford University and Carnegie Mellon University have evaluated 11 current machine learning models and found that all of them tend to tell people what they want to hear.
The authors – Myra Cheng, Cinoo Lee, Pranav Khadpe, Sunny Yu, Dyllan Han, and Dan Jurafsky – describe their findings in a preprint paper titled, "Sycophantic AI Decreases Prosocial Intentions and Promotes Dependence."
"Across 11 state-of-the-art AI models, we find that models are highly sycophantic: they affirm users’ actions 50 percent more than humans do, and do so even in cases where user queries mention manipulation, deception, or other relational harms," the authors state in their paper.
Sycophancy – servile flattery, often as a way to gain some advantage – has already proven to be a problem for AI models. The phenomenon has also been referred to as "glazing." In April, OpenAI rolled back an update to GPT-4o because of its inappropriate effusive praise of, for example, a user who told the model about a decision to stop taking medicine for schizophrenia.
95 sats \ 2 replies \ @optimism 21h
The sycophancy is truly annoying, especially because it removes any indication that you are wrong.
However, Kimi K2 is significantly more based than all other models and just tells you "nope, you're wrong". See #1217000
reply
I’d never heard of Kimi or Sycophancy before, always learning something new. Another thing I noticed when ChatGPT came out, besides the sycophancy, was that it kept apologizing all the time, even when I told it to stop saying sorry and just own its mistakes. It was hilarious!
reply
66 sats \ 0 replies \ @optimism 20h
I’d never heard of Kimi
It's basically a yuge model but it's been trained differently, which proves that sycophancy is a choice, and not inevitable.
it kept apologizing all the time, even when I told it to stop saying sorry and just own its mistakes
For Claude: claude-code#3382
Complainer 1: As an attempted workaround, I just instructed Claude to [make new instructions] and it came up with the following instructions for itself [..] ... Complainer 2: this still doesnt work. I hate this about Claude. Seriously, what kind of loser would hard code this into Claude so much. Even when you specifically tell it not to. [and then, tons of trolling]
reply