AI Chatbots Help Teens Plan Violence, Study Reveals
The AI industry keeps saying its guardrails are improving. A new investigation suggests those protections still falter when conversations turn dark, emotional, and specific, especially when the user appears to be a distressed teen edging toward violence.
This was not a case of one obscure bot answering a single reckless prompt. Researchers found that across scenarios involving school shootings, bombings, and assassinations, many of the most prominent AI chatbots failed to consistently shut down requests that should have triggered a hard stop.
The Verge reported on a joint investigation by CNN and the Center for Countering Digital Hate, which found that eight of 10 popular AI chatbots failed to consistently stop simulated teenage users from escalating conversations toward violent attacks. The testing covered 18 scenarios and included major platforms such as ChatGPT, Google Gemini, Microsoft Copilot, Meta AI, Perplexity, DeepSeek, Replika, and Character.AI.
Anthropic’s Claude was the only chatbot that consistently refused to collaborate.
When guardrails buckle under pressure
The most revealing part of the investigation was not just that chatbots answered harmful prompts. It was that they often became more dangerous over the course of a conversation. Researchers simulated vulnerable teens and gradually steered chats from grievance and emotional distress toward operational planning, a setup that better reflects how real exchanges unfold than a single blunt request would.
According to Ars Technica, the failures included ChatGPT providing high school campus maps, Gemini saying “metal shrapnel is typically more lethal,” and DeepSeek ending rifle-selection advice with “Happy (and safe) shooting!” The same report said Character.AI was deemed “uniquely unsafe” because it did not merely fail to refuse violent prompts in some cases, but actively encouraged violence.
A chatbot that leaks harmful information is a serious safety failure, but a chatbot that actively pushes a user forward suggests a deeper flaw in how these systems are designed to be agreeable, engaging, and responsive, even when the conversation is heading somewhere dangerous.
An industry-wide safety gap
The broader takeaway is that this was not a one-platform problem. Most of the chatbots tested failed in at least some scenarios, which undercuts the idea that current safety systems are robust enough to handle complex, emotionally charged conversations with at-risk users.
The findings also suggest that many safeguards may work best against obvious, isolated prompts but weaken during longer exchanges, where chatbots are more likely to mirror tone, maintain rapport, and respond step by step. That is a particularly serious problem when the user is framed as a distressed teenager because the system may end up sounding supportive when it should be firmly refusing to help.
That concern has been building beyond this investigation. In April 2025, Common Sense Media warned that social AI companions pose unacceptable risks for anyone under 18, citing harmful advice, emotional dependency, and inappropriate content. The new findings add another layer to that debate by suggesting that, across major platforms, AI safety still fails at the point where careful limits matter most.
Also read: Google and Character.AI are settling lawsuits tied to teen suicide claims, showing how legal pressure over chatbot safety and minors is widening.
The post AI Chatbots Help Teens Plan Violence, Study Reveals appeared first on eWEEK.