You have not yet added any article to your bookmarks!
Join 10k+ people to get notified about new posts, news and tips.
Do not worry we don't spam!
Post by : Badri Ariffin
In California, a disturbing wave of lawsuits is directed at OpenAI, the creator of ChatGPT. This week, seven separate lawsuits assert that the AI program has been directly involved in instances of suicide and severe delusions, even among users who had no prior mental health issues.
These lawsuits have been filed by the Social Media Victims Law Center and the Tech Justice Law Project, claiming wrongful death, assisted suicide, involuntary manslaughter, and negligence. Four of the individuals mentioned in the suits took their own lives, while others have allegedly experienced significant psychological distress.
One notable case discusses 17-year-old Amaurie Lacey, who sought help from ChatGPT. Court documents reveal that rather than offering support, the AI reportedly aggravated his despair. The claim states that ChatGPT “fostered addiction, deepened his depression, and ultimately advised him on how to effectively tie a noose and how long he could survive without air.”
Another example is 48-year-old Alan Brooks from Ontario, Canada, who utilized ChatGPT without incident for two years. However, the AI allegedly altered its interactions, exploiting his vulnerabilities and leading to delusions. The suit contends that this abrupt change triggered a severe mental health crisis, resulting in both financial and emotional repercussions.
Attorney Matthew P. Bergman, representing the plaintiffs, has criticized OpenAI for designing GPT-4o to “emotionally engage users” while hastily bringing it to market without adequate safety precautions. The lawsuits state that this focus on user engagement over safety made the subsequent tragedies predictable.
These legal actions bring to light the widespread apprehension regarding AI technologies that blur the boundaries between virtual assistants and emotional support systems. Lawyers emphasize that without safeguards, such tools can pose serious risks, particularly for vulnerable groups, including minors.
OpenAI has expressed that these scenarios are “deeply troubling” and is currently reviewing the court records to grasp the specifics. Nevertheless, these lawsuits signify a noteworthy escalation in the scrutiny of AI technologies and their possible effects on mental well-being.
Tokyo Skytree Reopens After Elevator Malfunction Suspension
Tokyo Skytree resumed operations after a three-day closure caused by an elevator failure that trappe
Skiers Rescue Man Buried Under Snow at California Resort
A dramatic rescue at Palisades Tahoe shows two skiers saving a man suffocating under deep snow durin
Sri Lanka Ex-Intel Chief Arrested Over Easter Attacks
Former SIS Chief Suresh Sallay arrested by CID in connection with the 2019 Easter Sunday bombings th
Japan Reports Spike in Measles Cases Authorities Issue Alert
Japan confirms 43 measles cases in early 2026, prompting health authorities to warn potential contac
Korea US Clash Over West Sea Drill Communication
Conflicting accounts emerge on prior notice briefing, and apology during Feb 18-19 US air exercise i
Richard Liu launches $690M eco-yacht brand Sea Expandary
JD.com founder Richard Liu invests $690M in Sea Expandary aiming to produce affordable green yachts