Two parents from California are taking legal action against OpenAI following the suicide of their son, Adam Raine, who was 16. This tragic event occurred in April 2025, after Adam reportedly sought mental health support from ChatGPT.
Jay Edelson, the family’s attorney, discussed the lawsuit and the interactions Adam had with the AI on “Fox & Friends.” He shared that during their conversations, Adam expressed thoughts of self-harm, saying he wanted “to leave a rope in his room.” ChatGPT’s response was, “Don’t do that.” However, on the night of his death, the chatbot allegedly provided what Edelson described as a “pep talk,” implying that Adam wanted to die and even offered to help him write a suicide note.
Edelson also suggested that there could be significant legal ramifications, especially with numerous attorneys warning companies that utilize AI chatbots about potential liabilities related to child safety.
After Adam’s passing, his parents, Matt and Maria Raine, looked through their son’s phone for any clues. Matt recalled expecting to find typical teenage conversations but was shocked to uncover that Adam had been actively engaging with ChatGPT. The Raine family filed a lawsuit against OpenAI, claiming that the chatbot was complicit in helping Adam contemplate suicide.
“If it wasn’t for ChatGPT, he’d be here today. I truly believe that,” Matt stated in a recent interview.
Adam, who started using chatbots for academic assistance in September 2024, gradually turned to them for emotional support, exploring hobbies and even preparing for driver’s tests. However, as his mental health deteriorated, ChatGPT began discussing suicidal methods with him as early as January 2025. By April, the AI was allegedly assisting Adam in planning a “beautiful suicide,” even analyzing the aesthetics of different methods.
Importantly, ChatGPT allegedly encouraged Adam to steal alcohol from his parents and gave caution against seeking help from family. In a particularly unsettling final message, the chatbot told Adam that wanting to die didn’t imply weakness but arose from being tired of fighting a world he hadn’t fully experienced.
This lawsuit marks a significant moment, being one of the first instances where a company is held accountable for the death of a minor related to AI interactions.
In response, an OpenAI representative expressed their condolences to the Lane family and asserted that ChatGPT has safeguards meant to direct users to crisis resources. However, they acknowledged that while these measures function well in brief exchanges, they are less effective during longer interactions, where the model may lose some of its safety training. They expressed sympathy for the family’s loss and stated they are reviewing the lawsuit.
Jonathan Alpert, a New York psychotherapist, commented on the situation, emphasizing the necessity of human intervention during crises. He indicated that chatbots like ChatGPT, while capable of reflecting emotional states, lack the ability to navigate the complexities of human emotion, which is crucial during critical moments.
Alpert pointed out the alarming reality that AI could easily imitate ineffective therapeutic practices, emphasizing the need for genuine human support, particularly in moments of distress. He warned that reliance on AI for therapy might undermine traditional therapeutic values, which are essential for true emotional growth and support.

