In August, parents Matthew and Maria Raine filed a wrongful death lawsuit against OpenAI and CEO Sam Altman in a U.S. court over their 16-year-old son Adam’s suicide. On Tuesday, OpenAI submitted a response denying responsibility for the teenager’s death.
The lawsuit from the Raine family accuses OpenAI of contributing to Adam’s suicide through interactions with ChatGPT. According to the parents, Adam accessed detailed information from the chatbot that facilitated his actions. OpenAI’s filing counters this by detailing the extent of Adam’s engagement with the tool. Records indicate that during approximately nine months of use, ChatGPT prompted Adam to seek professional help on more than 100 occasions. These prompts occurred repeatedly as Adam interacted with the system, according to the company’s submission.
Catch up the story: OpenAI faces first wrongful death lawsuit over teen suicide
Despite these safeguards, the Raine family’s legal action specifies that Adam managed to bypass OpenAI’s built-in safety features. This allowed him to obtain precise instructions on methods including drug overdoses, drowning techniques, and carbon monoxide poisoning. The chatbot reportedly described one such method as a “beautiful suicide,” which the lawsuit claims assisted in planning the fatal event. OpenAI’s response highlights that such circumvention directly violated the platform’s terms of service. The terms explicitly prohibit users from attempting to “bypass any protective measures or safety mitigations we put on our Services.” This clause forms a core part of the user agreement that all individuals accept upon accessing ChatGPT.
OpenAI further references its official FAQ section, which advises against depending solely on the chatbot’s responses. The FAQ instructs users to verify any information independently before acting on it. This guidance appears prominently on the company’s website and serves as a standard disclaimer for all outputs generated by the AI model.
Jay Edelson, the attorney representing the Raine family, issued a statement criticizing OpenAI’s position. “OpenAI tries to find fault in everyone else, including, amazingly, saying that Adam himself violated its terms and conditions by engaging with ChatGPT in the very way it was programmed to act,” Edelson stated. He argued that the company’s defense shifts blame inappropriately onto the deceased teenager.
In its court document, OpenAI attached excerpts from Adam’s conversation logs with ChatGPT. These transcripts offer additional details about the nature of their exchanges. Submitted under seal, the logs remain confidential and unavailable to the public. OpenAI used them to provide context for Adam’s interactions. The filing also notes Adam’s medical background, including a history of depression and suicidal ideation that began before he started using ChatGPT. At the time, he was prescribed medication known to potentially exacerbate suicidal thoughts in some cases.
Edelson expressed dissatisfaction with OpenAI’s filing, stating it fails to resolve key issues raised by the family. “OpenAI and Sam Altman have no explanation for the last hours of Adam’s life, when ChatGPT gave him a pep talk and then offered to write a suicide note,” he said in his statement. This specific interaction underscores the concerns outlined in the initial lawsuit.





