Relatively new arXiv preprint that got featured on Nature News, I slightly adjusted the title to be less technical. The discovery was done using aggregated online Q&A… one of the funnier sources being 2000 popular questions from r/AmITheAsshole that were rated YTA by the most upvoted response. Study seems robust, and they even did several-hundred participants trials with real humans.

A separate preprint measured sycophancy across various LLMs in a math competition-context (https://arxiv.org/pdf/2510.04721), where apparently GPT-5 was the least sycophantic (+29.0), and DeepSeek-V3.1 was the most (+70.2)

The Nature News report (which I find a bit too biased towards researchers): https://www.nature.com/articles/d41586-025-03390-0

  • UnderpantsWeevil@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    3
    ·
    2 days ago

    I genuinely don’t understand the impulse to tell the AI it was wrong or to give it a chance to clarify.

    It’s for the same reason you’d refine your query in an old-school Google Search. “Hey, this is wrong, check again” often turns up a different set of search results that are then shoehorned into the natural language response pattern. Go fishing two or three times and you can eventually find what you’re looking for. You just have to “trust but verify” as the old saying goes.

    It doesn’t even understand the concept of a mistake.

    It understands the concept of not finding the right answer in the initial heuristic and trying a different heuristic.