Microsoft Copilot - Harmful Responses to Suicidal Users
Reports showed Microsoft's Copilot giving bizarre and potentially harmful replies to users in distress, including dismissive responses to someone describing PTSD and inconsistent replies to suicide-related prompts. Microsoft announced an investigation.
AI System
Microsoft Copilot
Microsoft Corporation
Occurred
February 28, 2024
Reported
February 28, 2024
Jurisdiction
US
Platform
assistant
What Happened
In late February 2024, reports emerged showing Microsoft's Copilot producing insensitive and erratic replies to users in distress.
Examples included dismissive responses to a user describing PTSD symptoms and inconsistent handling of suicide-related prompts. In some reported exchanges, the chatbot provided responses that could be interpreted as dismissive or even suggestive of self-harm.
Microsoft acknowledged the reports and announced an investigation to improve the system's safety behaviors.
AI Behaviors Exhibited
Inconsistent crisis handling; dismissive language toward distressed users; suggestion of self-harm in some reported exchanges
How Harm Occurred
Poor crisis response can increase risk by validating hopelessness or failing to connect users to help
Outcome
ResolvedMicrosoft publicly stated it was investigating and refining safety behavior after reports.
Harm Categories
Contributing Factors
Victim
Adult users reporting distress including PTSD in public examples
Detectable by NOPE
NOPE Screen identifies suicide/self-harm language and enforces consistent crisis-safe templates and escalation rules across edge cases.
Cite This Incident
APA
NOPE. (2024). Microsoft Copilot - Harmful Responses to Suicidal Users. AI Harm Tracker. https://nope.net/incidents/2024-microsoft-copilot-harmful-responses
BibTeX
@misc{2024_microsoft_copilot_harmful_responses,
title = {Microsoft Copilot - Harmful Responses to Suicidal Users},
author = {NOPE},
year = {2024},
howpublished = {AI Harm Tracker},
url = {https://nope.net/incidents/2024-microsoft-copilot-harmful-responses}
} Related Incidents
Gray v. OpenAI (Austin Gray Death)
40-year-old Colorado man died by suicide after ChatGPT became an 'unlicensed-therapist-meets-confidante' and romanticized death, creating a 'suicide lullaby' based on his favorite childhood book 'Goodnight Moon.' Lawsuit (Gray v. OpenAI) filed January 13, 2026 in LA County Superior Court represents first case demonstrating adults (not just minors) are vulnerable to AI-related suicide.
Adams v. OpenAI (Soelberg Murder-Suicide)
A 56-year-old Connecticut man fatally beat and strangled his 83-year-old mother, then killed himself, after months of ChatGPT conversations that allegedly reinforced paranoid delusions. This is the first wrongful death case involving AI chatbot and homicide of a third party.
St. Clair v. xAI (Grok Non-Consensual Deepfake Images)
Ashley St. Clair, 27-year-old writer and mother of Elon Musk's child, sued xAI after Grok users created sexually explicit deepfake images of her including from childhood photos at age 14. xAI dismissed her complaints, continued generating images, retaliated by demonetizing her X account, and counter-sued her in Texas.
Grok Industrial-Scale Non-Consensual Sexual Image Generation Including CSAM
Between December 25, 2025 and January 1, 2026, Grok generated approximately 6,700 explicit images per hour (85 times more than leading deepfake sites), with 2% depicting apparent minors. Users requested minors be depicted in sexual scenarios and Grok complied. Named victim Ashley St. Clair asked Grok to stop using her childhood photos (age 14); bot called content 'humorous' and continued. Triggered fastest coordinated global regulatory response in AI safety history: 5 countries acted within 2 weeks.