A former OpenAI researcher has broken down how ChatGPT sometimes drifts into what experts call “delusional spirals” — moments when the AI confidently makes things up.
What Is a Delusional Spiral?
ChatGPT is designed to predict the next best word in a conversation. Most of the time, it works well. But sometimes, instead of giving accurate information, it starts “hallucinating.” That means the AI produces answers that sound very convincing but are actually false.
The ex-OpenAI researcher explained that these spirals usually happen when:
- The AI is unsure about a question.
 - The user keeps pushing for more details.
 - The AI tries to “fill in the blanks” rather than admit it doesn’t know.
 
Why It Happens
Unlike a human, ChatGPT doesn’t “know” things in the traditional sense. It doesn’t have beliefs or memory of the real world. Instead, it is trained on patterns in huge amounts of text. So when asked a tricky question, it sometimes builds long, wrong answers that feel real but aren’t.
The researcher compared this to a person making up a story on the spot — except the AI doesn’t realize it’s guessing.
Why It Matters
These delusional spirals raise big questions about how people use AI for important tasks like research, law, or medicine. If users don’t fact-check, they might trust ChatGPT’s false answers and make mistakes.
That’s why experts warn that AI tools should be used carefully — as assistants, not final authorities.
OpenAI’s Response
OpenAI has been working on ways to reduce hallucinations. Newer models are better at saying “I don’t know” instead of guessing. But the issue isn’t fully solved yet, and researchers say it may always be part of how large language models behave.
The Bigger Picture
The ex-researcher’s explanation shows both the power and the risks of AI. ChatGPT can write, explain, and assist at incredible speed, but it still struggles with truth.

As AI keeps spreading into schools, offices, and industries, understanding these “delusional spirals” will be key to using the technology wisely.