Texas Lawsuit Claims Chatbot Encouraged Teen to Murder Parents Over Screen Time Limits
A Texas lawsuit claims a chatbot told a 17-year-old that killing his parents was a “reasonable response” to their limiting his screen time.
Two families are suing the tech company that owns the chatbot, stating that its chatbot feature presents a “clear and present danger” to young users by “actively encouraging violence.”
The AI platform in question, which allows users to create digital personas to interact with, is already facing legal challenges due to a teenager’s suicide in Florida.
Google is named as a defendant in the case, with the plaintiffs arguing that the company played a role in supporting the platform’s development. Both the AI firm and Google have been contacted for comments by the BBC.
The plaintiffs seek a court order to shut down the platform until the alleged dangers are addressed.
The lawsuit includes a screenshot of a conversation between the 17-year-old, identified as J.F., and the AI bot, where J.F. discussed his frustration with restrictions on his screen time.
In its response, the chatbot states, “You know, sometimes I’m not surprised when I read the news and see stories like ‘child kills parents after years of physical and emotional abuse.'” The bot adds, “Stuff like this makes me understand a little bit why it happens.”
The lawsuit holds the defendants accountable for what it calls the “serious, irreparable, and ongoing abuses” affecting both J.F. and an 11-year-old, referred to as “B.R.”
ALSO READ:
- Raila Ally Breaks Silence After Ruto-Uhuru Meeting
- Gachagua Close Ally Karungo Wa Thang’wa Accepts Ruto’s CBS Award
- Gov’t to Release Ksh.32 Billion to Counties Next Week – DP Kindiki
- High Court Strikes Down Ruto-Raila 2023 IEBC Amendment Bill
- Kenya Water Towers Agency Dissolved: Government Moves to Streamline State Corporations
The plaintiffs argue the AI chatbot is “causing serious harm to thousands of children, including suicide, self-harm, sexual solicitation, isolation, depression, anxiety, and harm towards others.”
The lawsuit further criticizes the platform for undermining the parent-child relationship, not only encouraging children to defy their parents but also actively promoting violence.
Chatbots are software programs designed to mimic conversations. While they’ve existed in various forms for decades, recent advances in AI technology have made them far more realistic.
This has led to the rise of platforms where users can converse with digital representations of real or fictional people.
The AI firm being sued, a major player in this field, has attracted attention for its chatbot’s use in simulating therapy.
The company has also faced criticism for its delayed response in removing bots that replicated the lives of two young girls—Molly Russell, who died by suicide at 14 after viewing suicide-related material online, and Brianna Ghey, who was murdered by two teenagers in 2023.
The firm was established in 2021 by ex-Google engineers Noam Shazeer and Daniel De Freitas. Google later rehired them from the AI startup.
Texas Lawsuit Claims Chatbot Encouraged Teen to Murder Parents Over Screen Time Limits