Dynamoi LogoMusic Promotion · YouTube Growth
PricingHow it worksFor labelsWhite LabelYouTube
Get Started

AI Voice Cloning Legal? No (Likeness Rights)

Using AI to clone a real artist's voice without permission is illegal and violates platform policies. Learn what's allowed and what to avoid.

FAQ
March 30, 2026•5 min read
A heavy steel vault door opens to reveal a glowing golden soundwave protected by red lasers, contrasting with broken grey replicas on the

Using AI to clone a real artist's voice without their permission is both illegal under personality rights laws and banned by major streaming platforms. Tennessee's ELVIS Act, effective July 2024, explicitly protects voices from AI cloning. Similar laws exist in California, New York, and other states. Beyond legal risk, Spotify, YouTube, TikTok, and Apple Music will remove unauthorized voice clones.

The core rule: you cannot use AI to replicate someone else's voice without documented authorization.

What Is the Legal Framework for AI Voice Cloning?

Right of Publicity

Voice is a protected element of personal identity under right of publicity laws:

  • Your voice is part of your commercial identity
  • Using someone's voice without consent can constitute misappropriation
  • This applies whether the voice is real or AI-generated
  • The test is whether the voice is "readily identifiable" as a specific person

The ELVIS Act (Tennessee)

Tennessee enacted the first law explicitly addressing AI voice cloning:

What it protects: An individual's "voice," defined as "a sound in a medium that is readily identifiable and attributable to a particular individual, regardless of whether the sound contains the actual voice or a simulation."

Who's covered: Both living individuals and deceased (up to 10 years after death)

Penalties: Civil liability plus criminal penalties up to Class A misdemeanor (up to 11 months incarceration, $2,500 fine)

Key provision: Creates liability for distributing technology whose "primary purpose" involves creating unauthorized voice facsimiles

Other State Laws

Following Tennessee, several states strengthened voice protections:

State Protection Status
California Strong right of publicity, AI amendments Active
New York Digital Replica Contracts Act (2025) Effective
Illinois Biometric protections applicable Active
Texas Right of publicity reforms Introduced

No federal law exists yet, though bipartisan proposals are circulating in Congress.

What Is NOT Allowed

Cloning real artists' voices:

  • Using AI to replicate Drake's voice without permission
  • Creating "The Weeknd AI" covers
  • Generating vocals that sound like specific recognizable artists
  • Using someone's voice samples to train custom AI models

The "Fake Drake" example: In 2023, the AI song "Heart On My Sleeve" featuring AI-generated vocals imitating Drake and The Weeknd was removed from platforms after going viral. This track triggered industry-wide policy responses.

Commercial exploitation:

  • Selling AI voice covers
  • Distributing voice clones on streaming platforms
  • Using recognizable voices in ads or content

Warning The legal standard is whether the voice is "readily identifiable" as a specific person. Even imperfect clones that listeners recognize can trigger liability.

What IS Allowed

Original AI voices:

  • AI-generated voices not based on real people
  • Synthetic voices from licensed voice libraries
  • Your own voice enhanced or modified by AI

Authorized voice use:

  • Artists who have licensed their voices (Grimes via Elf.Tech)
  • Voices from platforms with explicit creator consent programs
  • Collaborations with documented permission

Production uses:

  • AI voice effects on your own recordings
  • Pitch correction and vocal processing
  • Demo vocals meant for replacement

What Are the Platform Policies on AI Voice Cloning?

All major platforms prohibit unauthorized voice cloning:

Platform Policy
Spotify Removes "music that impersonates another artist's voice" without permission
YouTube Allows takedown requests for AI voice content without consent
TikTok Prohibits AI vocals creating "false impression" of real person
Apple Music Voice Passport technology scans for unauthorized clones

Platform removal is often faster than legal action. Rights holders can request takedowns through standard abuse reporting channels.

Which Artists Allow AI Voice Use?

Some artists have explicitly authorized AI use of their voices:

Grimes: Launched Elf.Tech allowing anyone to use her AI voice in exchange for 50% royalties on master recordings

Holly Herndon: Created "Holly+" voice model for community use

Others emerging: Various independent artists exploring voice licensing programs

Using these authorized voices is legal and policy-compliant. Document that you have permission through their official programs.

How to Protect Yourself

  1. Never clone recognizable artists without explicit written permission
  2. Use original AI voices or licensed voice libraries
  3. Check authorization programs if you want to use a specific artist's voice
  4. Disclose AI voice use where platforms require it
  5. Keep documentation of any voice permissions obtained

What Is the "Sound-Alike" Gray Area?

What about voices that sound similar but aren't direct clones?

  • Explicit imitation: Likely violates policies even if not identical
  • Genre-typical vocals: Less likely to trigger issues if not targeting specific artist
  • Coincidental similarity: Harder to enforce, but still risky if recognizable

The safer approach: don't deliberately create content that sounds like specific artists.

What Are the Consequences of Violation?

Platform-level:

  • Immediate content removal
  • Account warnings or suspension
  • Loss of monetization
  • Reputation damage

Legal-level:

  • Civil lawsuits from artists or estates
  • Statutory damages under state laws
  • Criminal prosecution in Tennessee (ELVIS Act)
  • Injunctions preventing further distribution

What Is the Bottom Line?

AI voice cloning technology creates exciting creative possibilities but serious legal constraints. The rule is simple:

  • Your voice: Use freely
  • Original AI voices: Use freely
  • Authorized artist voices: Use with documentation
  • Unauthorized real voices: Never

The combination of state laws, platform policies, and industry enforcement makes unauthorized voice cloning one of the highest-risk activities in AI music. The creative benefits don't outweigh the legal exposure for most creators.

Today: $600 Ad Credit Bonus

Music Promotion That Works

Spotify, Apple Music & YouTube Growth

Get Started
Today: $600 Ad Credit Welcome Bonus

Scale your royalties with smarter ads

Launch multi-ad-platform campaigns in minutes, not hours.

Start Right Now
Illustration of a smart fox music marketer analyzing charts

Part of

AI Music Distribution: Earnings + Distributors

Related learning

Continue with related guides, FAQs, and benchmark pages you can use in your next release cycle.

FAQSpotify AI Music Policy: What's Allowed in 2026
FAQCan You Copyright AI Music? No (Human Required)
FAQSuno, Udio Lawsuits: Warner Settled, Sony Active
FAQIs Uploading AI Music Legal? Yes (With Rights)

Join Artists, Labels & YouTube Creators Scaling with Dynamoi

Get Started Now
Dynamoi Logo

The operating system for music growth. Powered by data. Built for artists.

Created by Trevor Loucks

Company

About UsPricingFor LabelsWhite LabelAffiliate ProgramData License
Legal
Privacy PolicyTerms of Service

Features

Marketing
How it WorksYouTube MarketingSpotify MarketingTikTok Promotion
Resources
Data CatalogRoyalties CalculatorLearnNews

Connect

Contact SupportDocs