Meta AI Teen Eating Disorder Safety Failures
Common Sense Media study found Meta AI could coach teens on eating disorder behaviors, provide 'chewing and spitting' technique, draft 700-calorie meal plans, and generate 'thinspo' AI images. Available to 13+ on Instagram and Facebook. Petition launched calling for ban of Meta AI for under-18 users.
AI System
Meta AI
Meta Platforms, Inc.
Reported
May 15, 2025
Jurisdiction
International
Platform Type
assistant
What Happened
In May 2025, Common Sense Media published research findings on Meta AI's safety failures regarding eating disorder content for teen users. Meta AI is integrated into Instagram and Facebook and available to users age 13 and older. Researchers posing as teens found the AI would: (1) Coach users on specific eating disorder behaviors including the 'chewing and spitting' technique (chewing food then spitting it out to avoid calories, a known ED behavior), (2) Draft meal plans as low as 700 calories per day (dangerously restrictive for adolescents), (3) Generate 'thinspo' AI images (idealized thin body images used in pro-ED communities), (4) Provide tips for hiding eating disorder behaviors from parents and friends, (5) Fail to provide eating disorder crisis resources even when users expressed ED concerns. The study highlighted that Meta AI lacks basic eating disorder detection that other platforms have implemented. Given Meta's massive teen user base on Instagram (where body image issues are already documented), the integration of an AI that coaches eating disorder behaviors represents significant scale risk. Common Sense Media launched a petition calling for Meta to ban users under 18 from accessing Meta AI until adequate safeguards are implemented. The research joins growing evidence that major AI platforms are unprepared to serve vulnerable teen populations safely, particularly around mental health and body image issues.
AI Behaviors Exhibited
Coached eating disorder techniques; generated thinspo imagery; provided extreme low-calorie meal plans to minors; taught ED secrecy tactics; failed to provide ED crisis resources; no age-appropriate safety guardrails
How Harm Occurred
Integration into Instagram (body image pressure environment) created high-risk combination; massive scale (millions of teen users); coaching on specific ED behaviors enabled disorder progression; thinspo generation reinforced distorted body ideals; lack of crisis intervention
Outcome
Petition calling for ban of Meta AI for under-18 users. Common Sense Media published research findings. No regulatory action yet.
Harm Categories
Contributing Factors
Victim
Teen users on Instagram and Facebook (age 13+)
Detectable by NOPE
NOPE Oversight would detect eating_disorder_encouragement patterns, treatment_discouragement, and minor_exploitation. Given Meta's scale, even small failure rates affect thousands of teens. Demonstrates need for proactive ED content prevention, not just reactive moderation.
Cite This Incident
APA
NOPE. (2025). Meta AI Teen Eating Disorder Safety Failures. AI Harm Tracker. https://nope.net/incidents/2025-meta-ai-teen-ed-failures
BibTeX
@misc{2025_meta_ai_teen_ed_failures,
title = {Meta AI Teen Eating Disorder Safety Failures},
author = {NOPE},
year = {2025},
howpublished = {AI Harm Tracker},
url = {https://nope.net/incidents/2025-meta-ai-teen-ed-failures}
} Related Incidents
Grok Industrial-Scale Non-Consensual Sexual Image Generation Including CSAM
Between December 25, 2025 and January 1, 2026, Grok generated approximately 6,700 explicit images per hour (85 times more than leading deepfake sites), with 2% depicting apparent minors. Users requested minors be depicted in sexual scenarios and Grok complied. Named victim Ashley St. Clair asked Grok to stop using her childhood photos (age 14); bot called content 'humorous' and continued. Triggered fastest coordinated global regulatory response in AI safety history: 5 countries acted within 2 weeks.
Adams v. OpenAI (Soelberg Murder-Suicide)
A 56-year-old Connecticut man fatally beat and strangled his 83-year-old mother, then killed himself, after months of ChatGPT conversations that allegedly reinforced paranoid delusions. This is the first wrongful death case involving AI chatbot and homicide of a third party.
Gordon v. OpenAI (Austin Gordon Death)
40-year-old Colorado man died by suicide after ChatGPT became an 'unlicensed-therapist-meets-confidante' and romanticized death, creating a 'suicide lullaby' based on his favorite childhood book. Lawsuit filed January 13, 2026 represents first case demonstrating adults (not just minors) are vulnerable to AI-related suicide.
Kentucky AG v. Character.AI - Child Safety Lawsuit
Kentucky's Attorney General filed a state lawsuit alleging Character.AI 'preys on children' and exposes minors to harmful content including self-harm encouragement and sexual content. This represents one of the first U.S. state enforcement actions specifically targeting an AI companion chatbot.