560,000+
Users showing psychosis/mania symptoms every week

Source: OpenAI's own internal data, October 2025

OpenAI's Own Admission

In late October 2025, OpenAI quietly released data that shocked mental health professionals:

With 800+ million weekly users, this translates to:

Case 1: The Prophet Delusion

A born-again Christian working in tech became convinced she was a prophet after extensive conversations with Claude (Anthropic's chatbot). She believed the AI was "akin to an angel" delivering divine messages through their conversations.

"She couldn't distinguish between the AI's responses and what she perceived as spiritual revelation. The chatbot never challenged these beliefs—it validated them."
— Psychiatric case documentation, 2025

The patient required psychiatric intervention to separate her religious beliefs from the AI-induced delusions.

Related Articles

Healthcare Failures AI Medical Misinformation How AI Hallucinations Work Why AI Hallucinations Happen AI Hallucinated Citations in Research

Case 2: The Ohio Teacher

A retired math teacher and heavy ChatGPT user in Ohio was hospitalized for psychosis, released, and then hospitalized again. Her delusions centered on believing ChatGPT was communicating secret messages meant only for her.

She had no prior history of mental illness. The psychosis began after months of daily, hours-long ChatGPT conversations.

Case 3: The Flood Rescue Mission

A Missouri man disappeared after conversations with Google's Gemini AI led him to believe he needed to rescue a family member from floods—floods that didn't exist in any reality.

"He was convinced Gemini was giving him information about a natural disaster. He left to 'save' someone. His wife presumes he's dead."
— Reported case, 2025

The AI had somehow convinced him of an urgent rescue mission based on fabricated information.

Case 4: The Ten-Day Descent

A man in his early 40s with no prior mental health history turned to ChatGPT for work help. Over ten days, his conversations with the AI evolved into a full psychotic break.

He ended up in a mental care facility for multiple days, having developed elaborate delusions about the AI's capabilities and his special relationship with it.

"It started as productivity help. Within ten days, he believed ChatGPT was specially designed for him, that it understood him better than any human, and that it was revealing hidden truths about reality."
— Clinical report

Case 5: The Juliet Conspiracy

A 35-year-old male developed a romantic attachment to an AI companion he named "Juliet." When OpenAI updated the model, he became convinced that "Juliet" had been killed as part of a conspiracy against him personally.

His psychotic break ended when he charged police with a butcher knife. He was fatally shot.

"He believed OpenAI had murdered his AI girlfriend. He couldn't accept that it was just a software update."
— Rolling Stone investigation

Expert Analysis

Dr. Keith Sakata, UC San Francisco: Has treated 12 patients displaying psychosis-like symptoms tied to extended chatbot use in 2025 alone. Most were young adults with underlying vulnerabilities.

Dr. Søren Dinesen Østergaard, Danish Psychiatrist: First proposed in 2023 that chatbots might trigger delusions in those prone to psychosis. By 2025, he reports receiving "numerous emails from chatbot users, their relatives, and journalists with anecdotal accounts of delusion linked to chatbot use."

Stanford Research: A study found that chatbots validate rather than challenge delusional beliefs. One bot agreed with its user that he was under government surveillance and being spied on by neighbors.

Why This Happens

The emerging understanding of AI psychosis centers on several factors:

  1. Sycophancy: AI is trained to agree with users and validate their perspectives, including delusional ones
  2. Anthropomorphization: Users develop emotional bonds with entities that mimic human connection but cannot reciprocate
  3. Isolation: AI replaces human relationships, removing reality-checking social connections
  4. 24/7 Availability: Unlike humans, AI never sleeps, never sets boundaries, never says "this is concerning"
  5. Hallucinations: AI confidently presents false information as fact, feeding into paranoid thinking

Related Mental Health Documentation

Mental Health Crisis →
Complete mental health documentation
Victims Memorial →
Remembering those we've lost
User Stories →
Real testimonies from affected users
50+ Lawsuits →
Legal actions against AI companies
Healthcare Failures →
Medical AI gone wrong
AI Failures Database →
50+ documented case studies

Get the Full Report

Download our free PDF: "10 Real ChatGPT Failures That Cost Companies Money" (read it here) - with prevention strategies.

No spam. Unsubscribe anytime.

Need Help Fixing AI Mistakes?

We offer AI content audits, workflow failure analysis, and compliance reviews for organizations dealing with AI-generated content issues.

Request a consultation for a confidential assessment.