2 pointsby vinhnx5 hours ago1 comment
  • soul_grafitti5 hours ago
    Ars Technica has a [good write up on this]<https://arstechnica.com/ai/2026/01/how-often-do-ai-chatbots-...>.

    > Anthropic is also quick to link this new research to its previous work on sycophancy, noting that “sycophantic validation” is “the most common mechanism for reality distortion potential.”

    I commented on Ars that I see the same sort of mechanism at work in politics. So is the really an AI problem or a case of AI surfacing inherently human characteristics? It's a sincere question.