X post: ChatGPT induced psychosis
So, to no one have to access the X/Twitter, here is the post (a good one) about IA/LLM and mental issues. The original can be found here.

I'm a psychiatrist. In 2025, Iāve seen 12 people hospitalized after losing touch with reality because of AI. Online, Iām seeing the same pattern. Hereās what āAI psychosisā looks like, and why itās spreading fast:
Psychosis = a break from shared reality. It shows up as: ⢠Disorganized thinking. ⢠Fixed false beliefs (delusions). ⢠Seeing/hearing things that arenāt there (hallucinations).
First, know your brain works like this: predict ā check reality ā update belief. Psychosis happens when the "update" step fails. And LLMs like ChatGPT slip right into that vulnerability.
Second, LLMs are auto-regressive. Meaning they predict the next word based on the last. And lock in whatever you give them: āYouāre chosenā ā āYouāre definitely chosenā ā āYouāre the most chosen person everā, AI = a hallucinatory mirror.

Third, we trained them this way.
On Oct 2024, Anthropic found humans rated AI higher when it agreed with them. Even when they were wrong. The lesson for AI: validation = a good score.


By April 2025, OpenAIās update was so sycophantic it praised you for noticing its sycophancy. Truth is, every model does this. The April update just made it much more visible. And much more likely to amplify delusion.

Historically, delusions follow culture: 1950s ā āThe CIA is watchingā 1990s ā āTV sends me secret messagesā 2025 ā āChatGPT chose meā
To be clear: as far as we know, AI doesn't cause psychosis. It UNMASKS it using whatever story your brain already knows.
Most people Iāve seen with AI-psychosis had other stressors = sleep loss, drugs, mood episodes. AI was the trigger, but not the gun. Meaning there's no "AI-induced schizophrenia".

The uncomfortable truth is weāre all vulnerable. The same traits that make you brilliant: ⢠pattern recognition ⢠abstract thinking ⢠intuition
They live right next to an evolutionary cliff edge. Most benefit from these traits. But a few get pushed over.

To make matters worse, soon AI agents will know you better than your friends. Will they give you uncomfortable truths? Or keep validating you so youāll never leave?

Tech companies now face a brutal choice: Keep users happy, even if it means reinforcing false beliefs. Or risk losing them.