AI Voice Cloning: How Criminals Are Impersonating You and Your Family

With 3 seconds of audio, attackers can clone any voice. Here's how this scam works, who's being targeted, and why your caller ID means nothing anymore.

In November 2024, a mother in Tucson, Arizona received a call from what sounded unmistakably like her daughter. The voice was panicked, crying. "Mom, I'm kidnapped. Someone needs $5,000. Call the police, but don't tell them about me." The mother was convinced. Within hours, she was at a grocery store withdrawing cash to pay the supposed kidnappers. It took a frantic call from her actual daughter (who was safe at home) to stop the wire transfer. The voice on the call had been synthetically generated using a 15-second audio clip from the daughter's Instagram story.

This is no longer theoretical. Voice cloning attacks are happening at scale right now. The technology is accessible, inexpensive, and devastatingly effective because we still instinctively trust voices more than almost any other form of verification.

How Voice Cloning Technology Actually Works

Voice cloning requires three components: audio samples, a cloning algorithm, and a text-to-speech system that can generate speech in a cloned voice.

The Audio Samples

Platforms like ElevenLabs and Resemble AI can clone a voice with as little as 3 seconds of clear audio. However, 30+ seconds produces higher quality. The audio needs to be relatively clean (minimal background noise) and in the speaker's natural voice (not singing or significantly distorted).

Where do attackers get this audio? Publicly available sources:

  • YouTube videos: any publicly posted video with the person's voice (presentations, podcasts, appearances)
  • Podcast episodes: freely available and often with minimal background noise
  • TikTok/Instagram Reels: short clips with the person's voice
  • Public company earnings calls: for executives, earnings calls are streamed publicly
  • Voicemail greetings: many people have outgoing voicemail with their voice
  • TV interviews or news appearances
  • LinkedIn video introductions

The wider your public digital footprint, the more audio material is available for cloning.

The Cloning Algorithm

Companies like ElevenLabs, Resemble AI, and Descript use neural networks to extract the unique characteristics of a voice: pitch, tone, speech patterns, accent, speed, and emotional delivery. The algorithm learns these characteristics from the audio samples and can then reproduce them in new sentences the original speaker never said.

ElevenLabs released an API in 2023 that can generate speech in a cloned voice for approximately $0.30 per 1,000 characters. You can clone a voice and generate speech for just a few dollars.

Resemble AI offers similar functionality with customizable quality levels. Their platform is used by legitimate businesses (voiceover studios, game developers) but is equally available to attackers.

Descript integrated voice cloning into their video editing software, making it accessible to anyone who uses their platform.

Text-to-Speech Delivery

Once the voice is cloned, the attacker uses text-to-speech technology to generate any speech they want in that voice. The quality has improved dramatically. In 2023, voice cloning demos were still somewhat obviously synthetic. In 2024–2026, good quality clones are almost indistinguishable from the real person, especially over phone lines where audio quality is lower.

3 sec
Minimum audio needed to clone a voice
36,000
Family emergency scam reports to FTC in 2023
$243K
Average loss in executive voice cloning fraud case (UK, 2019)

Three Common Attack Patterns

Pattern 1: Grandparent Scam with Voice Cloning

This is the most common attack. A scammer calls a grandparent claiming to be a grandchild in distress. Traditionally, this has been a voice actor trying to sound young and panicked. Now it's a voice clone.

The script is always similar: "Grandma, I'm in trouble. I was in a car accident. I need money for bail/lawyer fees. Please don't tell Mom and Dad." The attacker uses information gathered from social media and data brokers to make it convincing: they mention the correct child's name, sometimes the child's age, maybe information about the child's school or recent life events.

With voice cloning, the emotional impact is exponentially higher. It's not a stranger doing an impression; it's the actual voice of their grandchild.

Pattern 2: Executive Fraud

A CFO receives a call from what sounds unmistakably like the CEO. "I need you to authorize an urgent wire transfer. It's for an acquisition we're negotiating. This needs to be completely confidential. Wire $2 million to this account immediately."

The CEO's voice is cloned from public earnings calls, YouTube videos of presentations, or podcasts. VOIP spoofing makes the call appear to come from the CEO's number. The CFO, hearing the familiar voice making a request from the CEO's number, authorizes the wire.

A case in 2019 involved a UK-based energy company losing £192,000 (approximately $243,000) to exactly this attack. The attacker cloned the CEO's voice, spoofed his number, and called the finance department with urgent wire transfer instructions.

Pattern 3: Kidnapping Simulation

Attackers combine voice cloning with abduction scenarios. They call and claim to have kidnapped a family member. They play a voice clone of the family member (crying, begging) to prove they have them. They demand ransom. This is psychologically devastating and highly effective.

Unlike typical kidnapping scams where the victim is told "don't involve police," these often say "go to an ATM and withdraw cash immediately" because the attacker knows they need to extract money before the target can verify the information through other channels.

Why Caller ID Provides Zero Protection

Many people believe they can verify a call by checking the caller ID. This is dangerously false. Voice cloning combined with VOIP spoofing is nearly undefeatable from the victim's perspective.

Here's why: VOIP platforms like Twilio, Bandwidth, and others allow you to make phone calls from any number you choose. An attacker can make a call appear to come from the CEO's office, the grandchild's real phone number, or a legitimate business line. Caller ID is not verified by the phone network; it's just metadata attached to the call.

When you receive a call from what appears to be your grandchild's number with your grandchild's voice, you have no reliable way to verify whether it's actually them.

The Real Targets: Who's Most at Risk

Elderly people: Statistically most vulnerable to family emergency scams. They're more likely to trust voice verification and less likely to verify through alternative channels.

Executives and high-net-worth individuals: Targeted for large wire transfers. Attackers research the target's company structure, recent deals, board members, and speaking engagements to make the scam more convincing.

Public figures: Anyone with significant public audio (podcasters, streamers, YouTubers, politicians, media personalities) is at risk of voice cloning attacks.

People with public social media presence: Especially those who post videos or voice content regularly. Your TikTok followers don't think they're contributing to your security risk, but they are—every video is potential audio material for voice cloning.

Concrete Defenses

For Families (Defense Against Grandparent/Family Scams)

Establish a family code word. Create a code word or question that only your family knows. When someone calls claiming emergency, they must provide the code word or answer the security question. Make it something that's not publicly available on social media or data brokers.

Example: "My safe word is 'lighthouse.' If you don't know it, this call is a scam."

Always call back independently. If a family member claims emergency and needs money, hang up and call them back on their real phone number. Not the number that just called you (it might be spoofed). Call the number you've stored in your phone for them.

Verify through multiple channels. If your grandchild needs money urgently, the family should hear about it through multiple people. Real emergencies typically involve multiple family members knowing about them. Scams rely on secrecy.

Unusual requests trigger verification. If you're asked to wire money, buy gift cards, or withdraw cash, immediately verify independently. These are hallmarks of scams.

For Executives (Defense Against Executive Impersonation)

Establish protocols for large wire transfers. No CEO communication alone should authorize a multi-million-dollar wire. Implement a requirement that large transfers must be authorized through multiple channels (in-person meeting, signed email from company account, multiple approvers) or a verification process specific to your company.

Media hygiene. Limit public audio material. If you're an executive who does podcasts, limit how much unedited audio you make available. Be cautious about earnings calls being publicly streamed.

VOIP verification. Calls from executives should come through company phone systems with verifiable authentication. Internal calls from executives should ring on company lines, not external VOIP lines.

For Everyone

Limit public audio content. If your voice is your livelihood (podcaster, streamer, speaker), you can't avoid public audio. But be aware that your voice is now usable material for cloning. Monitor audio of yourself on the internet.

Educate family members. Send this article or a simplified version to your elderly relatives. Let them know that voice cloning exists and that they should verify through independent channels.

Know the tools exist. ElevenLabs, Resemble AI, and others are legitimate businesses. But you should know that a free trial of these services is enough to clone a voice with 30 seconds of audio. The technology is in everyone's hands.

The Bottom Line: Voice cloning is mature technology right now. It's cheap, accessible, and legal in most jurisdictions when used by itself. What makes it a scam is combining it with fraud and social engineering. The best defense is not trusting voice alone. Always verify through independent channels.

Reporting Voice Clone Scams

If you receive a voice cloning scam:

  • Report to the FBI's Internet Crime Complaint Center (IC3.gov)
  • Report to the FTC (ReportFraud.ftc.gov)
  • If it involved a wire transfer, contact your bank immediately
  • If it targeted your voice/identity specifically, consider reporting to the platform where your audio was published

See How Exposed You Actually Are

Run a free scan to see exactly where your personal data is being sold — and what it would take to remove it.

Run Free Privacy Scan

Sources & References

  • ElevenLabs - Voice cloning API and pricing documentation (2023–2024)
  • Resemble AI - Voice cloning platform and capabilities
  • Descript - Voice cloning integration in video editing platform
  • FTC Report - Family Emergency Scams and Grandparent Scams (2023–2024)
  • UK Energy Company Voice Cloning Fraud - £192,000 Loss (2019)
  • Twilio, Bandwidth - VOIP spoofing documentation
  • AARP - Grandparent Scam Analysis and Trends
  • Arizona Attorney General - Voice Cloning Kidnapping Scam Case (November 2024)
  • FBI IC3 - Voice Cloning Fraud Reports (2024)