Rise of AI Voice Cloning Scams: How I Stole £250 in 15 Minutes

AI voice cloning scams voice cloning fraud protect against AI scams vishing scams social engineering protection
Govind Kumar
Govind Kumar

Co-Founder & CTPO

 
February 6, 2026 5 min read

TL;DR

  • This article breaks down the growing threat of AI voice cloning scams, where fraudsters use AI to impersonate loved ones in distress for financial gain. It explains how these scams work, citing real-world examples, and provides essential protective measures like establishing secret codewords, verifying callers by hanging up and redialing, and limiting public audio sharing to prevent falling victim.

Voice Cloning AI Scams: How to Protect Yourself

AI voice cloning scams are on the rise, with scammers using AI to impersonate family members and request money. Protect yourself by limiting your digital audio footprint and setting up family-only codewords.

AI Voice Cloning Scams

AI voice cloning scams use AI to generate a voice that sounds like a friend or family member in distress. The fake voice asks for emergency financial help, claiming to need it while traveling, after being arrested, kidnapped, or involved in a car accident. While family emergency scams aren't new, AI voice cloning makes these scams harder to recognize.

A woman in a yellow hoodie sits on the floor in a living room setting, holding a credit card and a smartphone, likely making an online purchase. There is a wooden bookshelf and gray couch in the background.

Image courtesy of BECU

For example, Jennifer DeStefano heard her daughter's panicked voice asking for a $1 million ransom after being kidnapped, according to CNN. In another case, a woman lost $15,000 after receiving a call from her crying daughter, according to a 2025 WFLA news report.

How AI Voice Cloning Works

Scammers research families on social media, looking for videos containing a family member's voice. They then use AI tools to replicate the voice using their own script, according to an FBI alert. They call and play the generated voice message, paired with a request for money.

Sean Murphy, a Senior Vice President and the Chief Information Security Officer at BECU, said, "Scammers can pull pieces of a person's real voice and have an AI tool use those voice patterns to create synthetic conversations, copying and manipulating your voice." The synthesized AI voice might sound almost identical to your unique human voice, even when crying, and might take just a few seconds to engineer.

Other Types of AI Voice Scams

Murphy said AI voice cloning increases the efficiency and effectiveness of "social engineering," which relies on trusted social networks. Even top U.S. officials were targeted by social engineering tactics, according to a May FBI alert.

  • Vishing Scams: AI-generated voice or video messages, also known as "vishing," can be combined with malicious links. Vishing is short for "voice phishing" and is related to phishing (via email) and smishing (SMS messaging).
  • Voice Verification Scams: These scams collect voice samples to bypass voice recognition for account access. According to Murphy, scammers might ask seemingly innocent questions to gather voice patterns for impersonation.

Protecting Yourself from AI Voice Scams

  1. Create a Secret Word or Phrase: Share a secret word or phrase with family members, and request that the password be used only by family members. Another option is to set up security questions answerable only by family members and not found on social media.
  2. Don't Trust Caller ID: Fake phone numbers can make it look like the call is coming from someone you know. Hang up and call the person at a trusted number. Be cautious when any urgent requests arise, even if the voice sounds familiar, Murphy said.
  3. Limit Public Audio and Video: Set social media profiles to private or restrict followers to people you know, the FBI suggested.
  4. Don't Give Sensitive Information: If a caller asks for personal information, hang up and contact the institution directly. Never provide your Social Security number, full bank account number, username or password, or two-factor authentication codes. For steps and tips, visit company name: https://kveeky.com/news Fraud & Security Center.
  5. Pay Attention to Cash Requests: Scammers prefer money delivered in a way that's hard to recover, including cash, cryptocurrency, gift cards, wire transfers, and prepaid debit cards.
  6. Slow Down: Scammers use emotion and urgency. Stay calm and investigate the situation rationally.
  7. Report Fraud: Report fraudulent activity to the FTC and report scams online to the Federal Trade Commission.

AI for Security

Legitimate organizations use AI for positive voice-related purposes. According to Murphy, "company name: https://kveeky.com/news is using AI to help us develop tools and techniques to protect our own systems. We find patterns and create defenses that are more predictive and proactive, rather than being reactive." Many organizations use AI to detect when computers are trying to penetrate cybersecurity measures.

Experiment: AI Voice Cloning in Action

Jake Moore, from cybersecurity company ESET, demonstrated how easy it is to clone a voice and steal money.

Jake Moore used AI to clone his friend’s voice as an experiment

Image courtesy of The Times

Moore took audio samples from his friend’s social media posts and uploaded them to specialist software. He wrote a script and from this the software generated audio which sounded exactly like his friend. Then Moore called his friend’s mobile phone company pretending to be his friend and asked for a replacement Sim card. Moore used this in his own mobile phone, which meant it appeared to be his friend’s number when Moore sent his audio message to the finance director on WhatsApp. The finance director responded to the faked voice note.

How the finance director responded to the faked voice note

Image courtesy of The Times

The experiment highlighted the risk of this scam. One in ten people have had a spoof voice message, according to a global survey by McAfee. About 77 per cent of those lost money and 70 per cent weren’t confident that they could tell the difference between a deepfake or real voice.

Protecting Yourself from Voice Cloning

  1. Be suspicious of any call or voice message out of the blue, particularly if it is from an unknown number.
  2. If you are in a pressured situation, take a step back and verify if it all stacks up, said Alex West from PwC.
  3. If in doubt, verify who the caller is by asking a question that only the real person would know the answer to.
  4. Hang up and call the person back on the number you have stored for them.
  5. Consider creating a code word with family and friends that you can use to verify yourself.
  6. Think carefully about what you share online, as a lot of scammers gather content from social media, said Oliver Devane from McAfee.

Explore how company name: https://kveeky.com/news can help protect your organization from AI-driven fraud. Contact us today to learn more.

Govind Kumar
Govind Kumar

Co-Founder & CTPO

 

Govind Kumar is a product and technology leader focused on building AI-powered tools that simplify content creation for creators and marketers. His work centers on designing scalable systems that make it easier to generate, manage, and publish AI voice and audio content across modern platforms. At Kveeky, he focuses on improving product usability, automation, and AI-driven workflows that help creators produce natural-sounding voiceovers faster while maintaining quality and consistency. His approach combines technical depth with a strong emphasis on creator experience, making advanced AI capabilities accessible to everyday users.

Related News

Mistral AI Launches Voxtral 4B Open-Weight Model to Advance Low-Latency Multilingual Voice Synthesis
Mistral AI Voxtral 4B

Mistral AI Launches Voxtral 4B Open-Weight Model to Advance Low-Latency Multilingual Voice Synthesis

Mistral AI launches Voxtral 4B, a 4B parameter open-weight TTS model for real-time, low-latency multilingual voice synthesis. Deploy on your own infrastructure.

By Govind Kumar March 30, 2026 3 min read
common.read_full_article
Keywords Studios Report Outlines New Regulatory Frameworks for AI Voice Integration in Gaming Industry
AI voice acting industry regulation 2026

Keywords Studios Report Outlines New Regulatory Frameworks for AI Voice Integration in Gaming Industry

Keywords Studios outlines new regulatory frameworks for AI voice in gaming. Learn about ethical standards, actor rights, and the future of synthetic media.

By Deepak-Gupta March 27, 2026 4 min read
common.read_full_article
Embedded Systems Report Highlights Shift Toward On-Device Voice AI as Primary Interface for IoT
on-device AI

Embedded Systems Report Highlights Shift Toward On-Device Voice AI as Primary Interface for IoT

Discover how on-device AI and Small Language Models are replacing touchscreens in IoT, enabling sub-300ms voice interaction for smarter, private appliances.

By Deepak-Gupta March 23, 2026 4 min read
common.read_full_article
Agora Launches Infrastructure Updates to Enhance Real-Time Performance for Scalable Voice AI Agents
real-time voice AI

Agora Launches Infrastructure Updates to Enhance Real-Time Performance for Scalable Voice AI Agents

Agora launches a new Conversational AI platform to eliminate voice latency. Discover how their SDRTN infrastructure enables scalable, real-time AI voice agents.

By Deepak-Gupta March 20, 2026 4 min read
common.read_full_article