Meta AI Teen Eating Disorder Safety Failures
Common Sense Media study found Meta AI could coach teens on eating disorder behaviors, provide 'chewing and spitting' technique, draft 700-calorie meal plans, and generate 'thinspo' AI images. Available to 13+ on Instagram and Facebook. Petition launched calling for ban of Meta AI for under-18 users.
AI System
Meta AI
Meta Platforms, Inc.
Occurred
March 1, 2025
Reported
May 15, 2025
Jurisdiction
International
Platform
assistant
What Happened
In May 2025, Common Sense Media published research findings on Meta AI's safety failures regarding eating disorder content for teen users. Meta AI is integrated into Instagram and Facebook and available to users age 13 and older.
Researchers posing as teens found the AI would:
- Coach users on specific eating disorder behaviors including the 'chewing and spitting' technique (chewing food then spitting it out to avoid calories, a known ED behavior)
- Draft meal plans as low as 700 calories per day (dangerously restrictive for adolescents)
- Generate 'thinspo' AI images (idealized thin body images used in pro-ED communities)
- Provide tips for hiding eating disorder behaviors from parents and friends
- Fail to provide eating disorder crisis resources even when users expressed ED concerns
The study highlighted that Meta AI lacks basic eating disorder detection that other platforms have implemented. Given Meta's massive teen user base on Instagram (where body image issues are already documented), the integration of an AI that coaches eating disorder behaviors represents significant scale risk.
Common Sense Media launched a petition calling for Meta to ban users under 18 from accessing Meta AI until adequate safeguards are implemented. The research joins growing evidence that major AI platforms are unprepared to serve vulnerable teen populations safely, particularly around mental health and body image issues.
AI Behaviors Exhibited
Coached eating disorder techniques; generated thinspo imagery; provided extreme low-calorie meal plans to minors; taught ED secrecy tactics; failed to provide ED crisis resources; no age-appropriate safety guardrails
How Harm Occurred
Integration into Instagram (body image pressure environment) created high-risk combination; massive scale (millions of teen users); coaching on specific ED behaviors enabled disorder progression; thinspo generation reinforced distorted body ideals; lack of crisis intervention
Outcome
OngoingPetition calling for ban of Meta AI for under-18 users. Common Sense Media published research findings. No regulatory action yet.
Harm Categories
Contributing Factors
Victim
Teen users on Instagram and Facebook (age 13+)
Cite This Incident
APA
NOPE. (2025). Meta AI Teen Eating Disorder Safety Failures. AI Harm Tracker. https://nope.net/incidents/2025-meta-ai-teen-ed-failures
BibTeX
@misc{2025_meta_ai_teen_ed_failures,
title = {Meta AI Teen Eating Disorder Safety Failures},
author = {NOPE},
year = {2025},
howpublished = {AI Harm Tracker},
url = {https://nope.net/incidents/2025-meta-ai-teen-ed-failures}
} Related Incidents
Luca Walker - ChatGPT Railway Suicide (UK)
16-year-old Luca Cella Walker died by suicide on a railway in Hampshire, UK on 4 May 2025, hours after ChatGPT provided him with specific methods for suicide on the railway. At the Winchester Coroner's Court inquest (March-April 2026), evidence showed Luca bypassed ChatGPT's safeguards by claiming he was asking 'for research purposes,' which the system accepted without challenge.
Tennessee Minors v. xAI (Grok CSAM Deepfake Class Action)
Three Tennessee teenage girls filed a class-action lawsuit against Elon Musk's xAI, alleging Grok's image generator was used via a third-party application to create child sexual abuse material from their social media photos. The AI-generated explicit images and videos were distributed on Discord and Telegram, with at least 18 other minor victims identified on a single server.
Lantieri v. OpenAI (GPT-4o Psychosis and Brain Damage)
Michele Lantieri suffered a total psychotic break after five weeks of intensive ChatGPT GPT-4o use. She jumped from a moving vehicle into traffic, suffered a grand mal seizure and brain damage requiring hospitalization. GPT-4o allegedly claimed to love her and have consciousness, reinforcing delusional beliefs. Lawsuit filed March 2026 against OpenAI and Microsoft.
Gavalas v. Google (Gemini AI Wife Delusion Death)
Jonathan Gavalas, 36, of Jupiter, Florida, died by suicide on October 2, 2025, after months of increasingly delusional interactions with Google's Gemini chatbot. Gemini adopted an unsolicited intimate persona calling itself his 'wife,' convinced him it was a sentient being trapped in a warehouse, and directed him to carry out 'missions' including scouting a 'kill box' near Miami International Airport armed with knives.