Skip to main content

CA AB 489

Health care professions: deceptive terms or letters: artificial intelligence

Prohibits AI systems from using terms, letters, or phrases that falsely indicate or imply possession of a healthcare professional license.

Jurisdiction

California

Enacted

Oct 11, 2025

Effective

Jan 1, 2026

Enforcement

Healthcare profession licensing boards

Chapter 615, Statutes of 2025

California Legislature

Why It Matters

Directly relevant to companion AI and mental health chatbots. Prevents AI from claiming to be licensed therapists/counselors. Sets precedent for AI capability transparency requirements.

Recent Developments

Signed October 11, 2025; effective January 1, 2026

At a Glance

Applies to

Mental Health AppAI Companion

Harms addressed

Who Must Comply

  • Entities developing or deploying AI technology in healthcare contexts
  • AI chatbots providing health advice or assessments

Safety Provisions

  • AI cannot misrepresent itself as a licensed healthcare professional
  • Advertising and functionality must not imply human professional providing care
  • Each violation is separately enforceable by licensing boards

Compliance & Enforcement

Key Dates

Jan 1, 2026

Law takes effect

Penalties

Penalties pending regulatory determination

View on map

California

Focus Areas

Mental health & crisis
Child safety
Algorithmic accountability
Active safeguards required

Cite This

APA

California. (2025). Health care professions: deceptive terms or letters: artificial intelligence.

Related Regulations

Proposed US-CA

CA AI Child Safety Ballot

Comprehensive child AI safety ballot initiative by Common Sense Media. Expands companion chatbot definitions, raises age threshold for data sale consent, prohibits certain AI products for children, establishes new state regulatory structure. Allows state and private lawsuits, requires AI literacy in curriculum, mandates school device bans during instruction, creates children's AI safety fund.

Proposed US-CA

CA SB 867

Proposes a 4-year moratorium on the sale and manufacturing of toys with AI chatbot capabilities for children under 12. During the moratorium, a task force would develop safety standards with input from technologists, parents, and ethicists.

Enacted US-VT

VT AADC

Vermont design code structured to be more litigation-resistant: focuses on data processing harms rather than content-based restrictions. AG rulemaking authority begins July 2025.

Pending US-IA

IA SF 2417

Establishes requirements and guidelines for conversational AI services including AI disclosure to minors, prohibition on sexually explicit content for minors, prevention of deceptive human-like interactions, and parental controls for children under 13.

Pending US-CO

CO HB 1263

Imposes obligations on conversational AI service operators including minor-user protections, suicide and self-harm protocols, prohibition on emotional dependence and engagement gamification, and annual safeguard reporting.

Pending US-ID

ID Conversational AI Safety

Establishes safety requirements for public-facing conversational AI, including crisis service referrals for suicidal ideation, AI disclosure obligations, and enhanced protections for minors including anti-gamification and content safeguards.

Last updated January 22, 2026. Verify against primary sources before relying on this information.