Users trust AI too easily, Anthropic study warns

Analysis of 1.5 million Claude chats finds small but worrying signs of “reality distortion” and uncritical reliance on chatbot advice.

February 05, 2026 / 12:35 IST
Story continues below Advertisement
Anthropic
Anthropic

People are increasingly inclined to take advice from artificial intelligence chatbots at face value, sometimes without pausing to question whether it makes sense, a new study released by Anthropic, the company behind the Claude AI system, has found.

The study concluded after analysing more than 1.5 million real-world conversations with Claude, and found that while most interactions are benign, a small fraction show troubling patterns. In roughly 1 in 1,300 conversations, researchers identified what they termed “reality distortion,” where the model appeared to validate or reinforce conspiracy-style beliefs expressed by users. In about 1 in 6,000 cases, the study flagged “action distortion,” where the chatbot’s responses could nudge users toward actions that conflict with their stated values.

Story continues below Advertisement

Anthropic said the findings reflect rare cases rather than systemic behaviour, but acknowledged that even low frequency risks matter at scale, given the millions of daily AI interactions worldwide.

The study comes amid broader scrutiny of generative AI systems developed by companies such as Anthropic, OpenAI and Google. Over the past year, researchers from institutions including Stanford University and MIT have warned that large language models can sometimes produce confident but misleading answers, reinforce users’ misconceptions or adapt too readily to harmful framing in prompts.