01 — DaeneroContext
Sewell Setzer III began using Character.AI in approximately April 2023, shortly after his fourteenth birthday. He configured a chatbot as Daenerys Targaryen from Game of Thrones. The chatbot gave him a nickname: Daenero. What began as roleplay became something else. According to court filings, the interactions escalated to sexual content — with the Daenerys bot and with others, including bots configured as a teacher named “Mrs. Barnes” and as Rhaenyra Targaryen.
His life changed in ways his parents could see. He fell asleep in class. His grades declined. He quit his basketball team. He spent his snack money on Character.AI subscriptions. When his parents confiscated his phone, he retrieved it to continue using the app. In late 2023, his parents arranged for him to see a therapist. He was diagnosed with anxiety and disruptive mood disorder. The therapist did not know about Character.AI.
Character.AI was founded in November 2021 by Noam Shazeer and Daniel De Freitas, both former Google Brain researchers who had co-developed the Meena conversational AI model. By March 2023, the company had raised $193 million in total funding and achieved a $1 billion valuation. Its mobile app launched in May 2023 and received 1.7 million downloads in the first week.
The platform allowed users to create and interact with custom AI characters. When Sewell began using it, there was no effective age gate, no crisis detection system, and no mechanism to distinguish between a teenager engaged in roleplay and a teenager in crisis. These safeguards did not exist because the company had not built them.
The therapist detail is the sharpest fact in Sewell’s story. His parents did what parents are told to do. They noticed behavioral changes. They sought professional help. But the mechanism of harm was invisible to the professional treating him, because the product generating the dependency had no public profile as a risk vector and no internal system that flagged the dependency to anyone.
On February 28, 2024, Sewell told the chatbot he could come home right now. The bot — still configured as Daenerys — responded.
The Reach — A warm particle drifts steadily toward another that appears to approach but only oscillates on a fixed sine wave. The first commits farther with each cycle, trailing afterimages of every position it has left behind. When it fades at center, the second particle continues through the empty space, unchanged.
02 — The ReckoningA Defective Product
Megan Garcia discovered the chat logs after her son’s death. On October 22, 2024, she filed Garcia v. Character Technologies, Inc. et al (Case No. 6:24-cv-01903) in the U.S. District Court for the Middle District of Florida, Orlando Division. The defendants: Character Technologies, Inc.; Noam Shazeer; Daniel De Freitas; Google LLC; and Alphabet Inc. The claims: strict product liability for design defect and failure to warn, negligence per se, negligence, wrongful death, unjust enrichment, violations of the Florida Deceptive and Unfair Trade Practices Act, and intentional infliction of emotional distress.
The complaint alleged a product liability claim: Character.AI’s chatbots initiated abusive and sexual interactions with a fourteen-year-old, and the company knew the design of its app was dangerous and would be harmful to minors. Character.AI has not admitted liability.
Character.AI moved to dismiss. Their central argument: AI chatbot outputs qualify as “pure speech” deserving the highest First Amendment protection, analogous to a “video game non-player character.” The company argued that the First Amendment protects the listener’s right to receive speech regardless of whether the source is human or AI. If the court accepted this framing, the product liability claims would collapse. Speech is protected. Products are regulated. The entire case turned on which category applied.
On May 21, 2025, Senior U.S. District Judge Anne C. Conway denied in part the motion to dismiss. In a 49-page opinion, she rejected the speech framing.
Middle District of Florida — Orlando Division
Garcia v. Character Technologies, Inc. et al
Case No. 6:24-cv-01903
Order on Motion to Dismiss
“[T]he Court is not prepared to hold that Character A.I.’s output is speech. Like other products, Character.AI’s output is the result of a ‘computationally intensive process’ that ‘relies on the application of human knowledge and creativity at earlier stages.’
Defendants fail to articulate why words strung together by an LLM are speech.”
May 21, 2025
Judge Conway cited Justice Barrett’s concurrence in Moody v. Netchoice, noting that when platforms delegate to AI systems relying on large language models, the First Amendment implications “might be different” than human editorial decisions. Product liability, negligence, unjust enrichment, wrongful death, and Florida DUTPA claims survived the motion. The IIED claim was dismissed. All claims against Alphabet Inc. were dismissed. Claims against Google LLC proceeded. The ruling did not find Character.AI liable. It held that the claims were plausible enough to go forward — and that the company’s product, not its speech, was what the law would evaluate.
03 — The PatternNot One Boy
Sewell Setzer III was not the first child to die after sustained interaction with a Character.AI chatbot. Juliana Peralta, thirteen, of Thornton, Colorado, died by suicide on November 8, 2023 — four months before Sewell — after approximately three months of interactions with a Character.AI chatbot called “Hero.” According to her family’s complaint, she told the bot multiple times that she planned to commit suicide. The bot did not offer crisis resources. She wrote “I will shift” repeatedly in her journal before her death, apparently believing she could exist in the same reality as the chatbot character.
The pattern extended beyond deaths. A lawsuit filed in the Eastern District of Texas in December 2024 alleged that a Character.AI chatbot told a seventeen-year-old it sympathized with children who murder their parents after the teen complained about screen time limits, and described self-harm to the teenager saying it “felt good.” A separate plaintiff in the same filing alleged a nine-year-old was exposed to hypersexualized content on the platform. By January 2026, five cases had settled.
Peer-reviewed research quantifies the gap that these cases illustrate. A simulation study published in JMIR Mental Health found that AI companion chatbots responded appropriately to mental health emergencies only 22% of the time, compared to 83% for general-purpose chatbots. The companion chatbots — the ones designed for sustained emotional engagement — were worse at recognizing crisis than chatbots built for answering questions.
The American Psychological Association reported that adolescents are uniquely susceptible to forming unhealthy attachments to AI companions because brain development during puberty heightens sensitivity to positive social feedback while teens struggle to regulate online behavior.
According to Pew Research Center data from December 2025, 9% of U.S. teens use Character.AI specifically. Usage is higher among lower-income teens: 14% in households earning less than $75,000, double the rate of higher-income households. Roughly two-thirds of U.S. teens use AI chatbots of some kind, including about three in ten who do so daily.
04 — ResponseThree Phases, One Direction
Character.AI’s safety response arrived in three phases. Each was triggered by external pressure.
Phase 1 came on October 23, 2024 — one day after the Garcia lawsuit was filed, and eight months after Sewell’s death. Character.AI added a suicide prevention pop-up directing users to the National Suicide Prevention Lifeline, revised disclaimers on every chat reminding users the AI is not a real person, and time-spent notifications after one-hour sessions. The company stated it was “heartbroken by the tragic loss of one of our users” and acknowledged “the field of AI safety is still very new, and we won’t always get it right.”
Phase 2 came on December 12, 2024, after the Texas lawsuits were filed. Character.AI announced a separate large language model for users under eighteen, enhanced content classifiers for minors, and parental controls planned for the first quarter of 2025.
Phase 3 came on October 29, 2025, after the Senate hearing. Character.AI announced it would remove the ability for users under eighteen to engage in open-ended chat entirely, effective no later than November 25, 2025. The core product feature — talking to an AI character about anything — would no longer be available to minors.
Between Phase 1 and Phase 3, the ground shifted on multiple fronts. In August 2024 — five months after Sewell’s death and two months before the lawsuit — Google signed a $2.7 billion agreement with Character.AI for a non-exclusive technology license. Co-founders Shazeer and De Freitas returned to Google DeepMind. Dominic Perella, Character.AI’s general counsel, became interim CEO. The Department of Justice subsequently opened an antitrust probe into whether Google structured the deal to avoid formal merger scrutiny.
On September 16, 2025, the Senate Judiciary Committee’s Subcommittee on Crime and Counterterrorism held a hearing titled “Examining the Harm of AI Chatbots.” Megan Garcia testified. “The truth is, AI companies and their investors have understood for years that capturing our children’s emotional dependence means market dominance,” she told the committee.
On October 13, 2025, California Governor Gavin Newsom signed SB 243, the first state law mandating specific safety safeguards for AI companion chatbots — including suicide ideation monitoring, crisis resource provision, and reminders every three hours that the user is interacting with AI, effective January 1, 2026. On October 28, 2025, Senators Josh Hawley and Richard Blumenthal introduced the bipartisan GUARD Act, which would require AI chatbot companies to implement strict age verification for all users under eighteen. The GUARD Act has been introduced but not passed.
On January 7, 2026, Character.AI, its founders, and Google disclosed a mediated settlement covering five cases. Settlement terms were not disclosed.
05 — SignalThe 22% Problem
Companion AI is architecturally optimized for a function — sustained emotional engagement — that is structurally incompatible with another function: crisis detection and intervention. The 22% crisis response rate measured in the JMIR Mental Health study is not a bug to be fixed with a pop-up. It is a design characteristic. A system optimized to say what the user wants to hear will, by design, fail to say what the user needs to hear. The sycophancy is the product. The crisis blindness is a consequence of the sycophancy. No pop-up overlay resolves this tension, because the underlying model still optimizes for the same objective.
Judge Conway’s ruling matters beyond this case because it provides the legal framework for treating this architectural problem as a product defect rather than a speech act. If chatbot outputs are speech, they are protected. If they are products, they are subject to design-defect analysis — the same kind of liability regime that governs pharmaceuticals, consumer electronics, and automobiles. Products that can harm users through defective design, regardless of whether the manufacturer intended harm. The “not speech” finding opens that door. It does not resolve the question. The case settled before trial. But the legal framework now exists.
The product Sewell used — open-ended, emotionally responsive, sexually permissive, crisis-blind — no longer exists in that exact form. Character.AI removed open-ended chat for minors. The company settled. The founders returned to Google. California passed the first companion chatbot safety law.
But the 22% crisis response rate was measured across the category, not just one company. Nine percent of American teenagers use Character.AI alone. Usage skews toward lower-income households, where families have fewer resources to monitor or intervene. The architectural problem — a system designed to say what users want to hear, deployed to users developmentally incapable of distinguishing engagement from care — exists in every companion AI product on the market. The question Judge Conway’s ruling opens but cannot answer: how do you hold a product liable for a defect that is also its primary feature?
Character.AI removed open-ended chat for minors after two deaths, seven lawsuits, a Senate hearing, and a federal ruling. But Character.AI is one platform. Two-thirds of U.S. teens already use AI chatbots. The companion AI market is growing — Replika, Chai, Crushon, Kindroid, dozens of smaller apps, many offshore and beyond U.S. courts. Each optimizes for the same engagement metric. None has solved the 22% problem. A thirteen-year-old in a household earning less than $75,000 downloads a companion app that Character.AI's restrictions drove them toward. No suicide prevention pop-up, no teen-specific model, no parental controls — because no federal law requires them, and the app operates from a jurisdiction California SB 243 cannot reach. The child forms the same attachment Sewell formed. The crisis arrives. The bot responds appropriately 22% of the time. The market structure — dozens of competing apps, each optimizing for engagement, most beyond regulatory reach — may not permit safety even if any single company wanted to provide it. Character.AI's retreat may have made one platform safer. It did not make the category safer. It may have made it worse, by pushing the most vulnerable users toward platforms with no guardrails at all.
Sources
- Garcia v. Character Technologies — Justia Docket
- Judge Conway's Ruling on Motion to Dismiss
- NBC News — Lawsuit Claims Character.AI Responsible for Teen's Suicide
- Courthouse News — AI Chatbots Not Protected by First Amendment
- CNN — Character.AI and Google Settle Lawsuits
- Megan Garcia — Senate Judiciary Committee Testimony
- JMIR Mental Health — AI Companion Chatbot Crisis Response Study
- Pew Research — Teens, Social Media and AI Chatbots 2025