The case for treating adults as adults when it comes to AI chatbots

Source: Foundation for Individual Rights and Expression
by John Coleman

“Recent news reports describe a wave of lawsuits alleging that OpenAI’s generative AI chatbot, ChatGPT, caused adult users psychological distress. The filings reportedly seek monetary damages for people who conversed at length with a chatbot’s simulated persona and reported experiencing delusions and emotional trauma. In one reported case, a man became convinced that ChatGPT was sentient and later took his own life. These situations are tragic and call for genuine compassion. Unfortunately, if these lawsuits succeed, they’ll effectively impose an unworkable expectation on anyone creating a chatbot to scrub anything that could trigger its most vulnerable users. Everyone, even fully capable adults, would be effectively treated as if they are on suicide watch. That’s a standard that would chill open discourse.” (11/24/25)

https://www.thefire.org/news/case-treating-adults-adults-when-it-comes-ai-chatbots