Imagine receiving a panicked message from a close friend claiming they need $10,000 immediately to avoid a severe financial crisis—only it’s not actually your friend. It’s a scammer. This alarming scenario is becoming increasingly common due to advancements in artificial intelligence (AI) technology. Scammers can now mimic voices using just a few seconds of recorded audio, posing a serious risk to individuals and businesses. As your financial partner, we're committed to keeping you safe from fraud. Let’s explore how these scams work and discuss ways to protect yourself.
An Evolving Threat
Over the past few years, AI-powered voice impersonation has become a go-to tool for criminals, helping them pull off incredibly convincing scams. According to the Federal Trade Commission, around 15% of people fell victim to voice-based phishing, or "vishing," in 2022, with a median loss of $1,400.
“AI-powered voice impersonation has really ramped up scammers' ability to pull off believable cons,” Oto R., Advia’s Information Security Officer, shared. “Vishing statistics provided by the FTC show how much of a growing trend this is and how many people are falling prey. It’s clear that we need to be defensive and address these evolving threats head-on.”
Signs of an AI Voice Cloning Scam
While voice cloning software can mimic voices incredibly well, there are red flags that can help you spot a fake call. Oto shares some warning signs of a potential AI voice scam:
- If a voice sounds familiar and tries to rush you into action without giving clear details, be cautious. This could be emotional manipulation.
- Listen for weird pauses or unnatural speech patterns that don't quite add up. This could be a scammer using AI.
- Beware if they ask for money through odd methods like cash, gift cards, or cryptocurrencies. We do not recommend giving into questionable payment requests.
- If the call sounds fuzzy or has strange echoes, it might be a sign you're dealing with a scammer using low-grade software.
Oto R., Advia’s Information Security Officer
How To Protect Yourself
Oto also identified five easy tips to protect yourself:
- Always verify the caller's identity, especially when the request is unusual or urgent. Hang up and call back at a known phone number or seek confirmation via a different communication channel.
- If you are suspicious you are talking to a bot and want to check, you can ask a random question, such as “What's your go-to karaoke song?” An AI, not programmed with relevant scripts, will struggle to respond.
- Be cautious of any unexpected requests for sensitive information, money transfers, or changes in protocol.
- Trust your instincts. If the situation feels suspicious, it probably is.
- Fill out a report on the FTC’s report fraud website. You can also contact Advia if you’re worried that you may have fallen victim to an AI voice scam. Our fraud experts are here to assist you.
“Understanding how these scams operate and spreading awareness is the first line of defense,” Oto said. “Everyone should stay vigilant and careful with unexpected communications. And remember, Advia is always here to talk if you have any concerns." You can explore more about our fraud protection on our ID Theft Protection webpage.