NO FAKES Act
NO FAKES Act of 2025 (Nurture Originals, Foster Art, and Keep Entertainment Safe Act) (H.R.2794 / S.1367)
Creates federal intellectual property protection for individuals' voice and visual likeness against unauthorized AI-generated digital replicas. Holds companies liable for producing, hosting, or sharing non-consensual digital replicas.
Jurisdiction
United States
Enacted
Pending
Effective
TBD
Enforcement
Federal courts
Introduced April 9, 2025. Both House (H.R.2794) and Senate (S.1367) versions referred to Judiciary Committees. No committee vote as of March 2026.
Congress.govWhy It Matters
Would establish the first federal framework for protecting individuals from unauthorized AI-generated digital replicas of their voice and likeness, complementing the DEFIANCE Act's focus on non-consensual intimate imagery.
Recent Developments
Reintroduced April 2025 with bipartisan support from Senators Klobuchar, Coons, Blackburn, and Tillis. Revised version expanded to 39 pages with notice-and-takedown provisions. Strong industry support (SAG-AFTRA, RIAA). EFF criticism of censorship infrastructure requirements.
At a Glance
Applies to
Harms addressed
Requires
Who Must Comply
- AI system developers creating digital replicas
- Platforms hosting AI-generated content
- Entities distributing AI-generated replicas
Obligations fall on:
Safety Provisions
- Federal right protecting voice and visual likeness from unauthorized AI replication
- Liability for producing unauthorized digital replicas
- Liability for hosting or sharing unauthorized digital replicas
- Notice-and-takedown requirements for platforms
Compliance & Enforcement
Penalties
Civil liability for damages
Private Right of Action
Individuals can sue directly without waiting for regulatory action.
View on map
United States
Focus Areas
Cite This
APA
United States. (n.d.). NO FAKES Act of 2025 (Nurture Originals, Foster Art, and Keep Entertainment Safe Act) (H.R.2794 / S.1367).
Related Regulations
TAKE IT DOWN Act
First federal law addressing AI-generated intimate imagery. Criminalizes publication of nonconsensual intimate imagery (NCII) including AI "digital forgeries." Creates 48-hour takedown obligation for platforms.
AI LEAD Act
Classifies AI systems as 'products' under federal law and establishes a federal cause of action for product liability claims against AI developers and deployers, including claims for design defects, failure to warn, and strict liability.
EU PLD
Modernized product liability framework explicitly covering AI systems and software as products. Shifts burden of proof in complex AI cases, allows disclosure orders for technical documentation, and addresses liability for AI-caused harm including through software updates.
Singapore OSRA
Creates a dedicated Online Safety Commission (OSC) with powers to order takedowns and disable access, establishes statutory torts providing victims of online harms with direct civil remedies against platforms and perpetrators.
TX Deepfake Intimate Images Law
Expands civil liability for AI-generated nonconsensual intimate images (deepfake pornography). Criminalizes threatening to create intimate deepfakes to coerce, extort, harass, or intimidate. Imposes penalties on individuals, websites, and payment processors involved in distributing such content. 10-year statute of limitations.
TX AI Catfishing Law
Establishes civil liability for online impersonation using AI. Person liable if they knowingly and with intent to harm, defraud, intimidate, or threaten use AI to impersonate another's name, voice, signature, photograph, or likeness. Civil remedies include injunctive relief, actual damages, exemplary damages ($500+ minimum), costs, and attorney's fees. Satire and parody exempted.
Last updated March 23, 2026. Verify against primary sources before relying on this information.