

In the summer of 2025, OpenAI released ChatGPT 5 and removed its predecessor from the market. Many subscribers to the old model had become attached to its warm, enthusiastically agreeable tone and complained at the loss of their ingratiating robotic companion. Such was the scale of frustration that Sam Altman, OpenAI’s CEO, had to acknowledge that the rollout was botched, and the company reinstated access. Anyone who’s been told by a chatbot that their ideas are brilliant is familiar with artificial intelligence sycophancy: its tendency to tell users what they want to hear. Sometimes it’s very explicit – “that is such a deep question” – and sometimes it’s a lot more subtle. Consider an AI calling your idea for a paper “original,” even if many people have already written on the same topic, or insisting that your dumb idea for saving a tree in your garden still contains a germ of common sense. AI sycophancy seems harmless, maybe even cute, until you imagine someone consulting a chatbot about a weighty question, like a military strategy or a medical treatment. We study the impact of extensive human interactions with chatbots, and we recently published a paper on the ethics of AI
Lean: 0.050 · Source quality 60/100 · Factual vs opinion 45/100.
© 2026 Vistoa. All rights reserved.
Limited excerpts, attribution, analysis, and outbound publisher links remain core product boundaries.