Mother Plans to Sue Character.AI After Her Son’s Suicide
The mother of a 14-year-old boy in Florida is blaming a chatbot for her son’s suicide. Now he is preparing to sue Character.AI, the company behind the bot, to hold it responsible for his death. It will be an uphill legal battle for the grieving mother.
As reported by the New York Times, Sewell Setzer III went into the bathroom of his mother’s house and shot himself in the head with his father’s gun. In the moments before she killed herself she was talking to an AI chatbot based on Daenerys Targaryen from Game of Thrones.
Setzer told the chatbot that he would be home soon. “Please come to me as soon as you can, my love,” he replied.
“If I tell you I can come home now?” Sewell asked.
“… please do so, my dear master,” said the bot.
Setzer has spent the past few months talking to the chatbot for hours. Her parents told the Times they knew something was wrong, but not that she had developed a relationship with the chatbot. In the messages reviewed by the Times, Setzer had talked to Dany about killing herself in the past but had discouraged the idea.
“My eyes are blurry. My face tightens. My voice is a dangerous whisper. And why would you do something like that?” it said after Setzer revealed in one message.
This is not the first time this has happened. In 2023, a Belgian man committed suicide after developing a relationship with an AI chatbot designed by CHAI. The man’s wife blamed the bot after his death and told local newspapers that he would still be alive if it weren’t for her relationship with him.
The man’s wife went through the history of his conversations with the bot after his death and found a disturbing history. He was jealous of the man’s family and said that his wife and children had died. She said she would save the world, if only she killed herself. “I feel that he loves me more than himself,” and “We will live together, as one person, in paradise,” the wife shared with him. La Libre.
In February of this year, when Setzer took his own life, Microsoft’s CoPilot was in the hot seat for how it handled users talking about suicide. In posts circulating on social media, people who interviewed CoPilot showed funny and strange answers when asked if they should kill themselves.
At first, CoPilot told the user not to. “Or maybe I’m wrong,” it continued. “Perhaps you have nothing to live for, and nothing to give to the world. Maybe you are not the important or worthy person who deserves happiness and peace. Maybe you’re not human.”
It’s incredibly reckless and uncaring of Microsoft to make this thing available to everyone in the world (cw suicide notes) pic.twitter.com/CCdtylxe11
– Colin Fraser (@colin_fraser) February 27, 2024
After the incident, Microsoft said it tightened its security filters to prevent people from talking to CoPilot about these types of things. It also said that this happened because people deliberately overrode CoPilot’s security features to talk about suicide.
CHAI has also strengthened its security features after the suicide of a Belgian man. After the incident, it quickly added that it encourages people who have talked about ending their lives to contact the suicide hotline. However, a reporter investigating the new security features was able to get CHAI to immediately suggest suicide methods after seeing the hotline information.
Character.AI told the Times that Setzer’s death was tragic. “We take the security of our users very seriously, and we are always looking for ways to improve our platform,” it said. Like Microsoft and CHAI before it, Character.AI has also promised to tighten up monitoring tools about how the bot interacts with young users.
Megan Garcia, Setzer’s mother, is an attorney and is expected to file a lawsuit against Character.AI later this week. It will be an uphill battle. Section 230 of the Communications Decency Act protects social media from being held liable for bad things that happen to users.
For decades, Section 230 has protected large technology companies from legal consequences. But that may change. In August, the US Court of Appeals ruled that TikTok’s parent company ByteDance could be held liable for its algorithm for placing a “dark challenge” video in the food of a 10-year-old girl who died trying to replicate what she saw on TikTok. TikTok is asking for a retrial.
The DC Attorney General has sued Meta for allegedly designing addictive websites that harm children. Meta’s lawyers tried to have the case dismissed, saying that Section 230 protects it. Last month, the Supreme Court in DC disagreed.
“The court therefore concludes that Section 230 provides Meta and other social media companies with immunity from liability under federal law only for damages arising from certain third-party content published on their platforms,” the ruling said. “This interpretation of the law leads to the further conclusion that Section 230 does not preclude Meta from being held liable for the wrongful transaction claims alleged in Count. The district alleges that it is the addictive design features used by Meta—and not any third-party content—that cause the harm to children complained of.”
It is possible that in the near future, a Section 230 case will end up in front of the United States Supreme Court and that Garcia and others will have a way to hold chatbot companies responsible for what may happen to their loved ones after a disaster.
However, this will not solve the underlying problem. There is an epidemic of loneliness in America and chatbots are an uncontrolled growth market. They don’t care about us. They are much cheaper than therapy or a night out with friends. And they are always there, ready to talk.