I know it sounds like something out of a futuristic movie, but scammers are now using artificial intelligence to create highly convincing impersonations of people you know. This includes audio and video that sounds and looks just like you.
Table of Contents
This isn’t just happening to celebrities or politicians; everyday people are being targeted, and their loved ones are falling for these schemes. The technology is so advanced that a model from Microsoft, for example, could mimic a specific voice with just a three-second audio sample over two years ago. It’s a scary thought, but there’s a surprisingly low-tech solution to protect yourself and your family.
These schemes, often called a ‘grandparent scam,’ catch people off guard because they don’t expect a phone or video call from a loved one to be a fake. They hear your voice, see your face, and their immediate reaction is to help.
This emotional response can lead them to hand over money, thinking they’re rescuing you from an emergency like an accident, a bail scam, or even a virtual kidnapping scam. But you and your loved ones can avoid falling for this scam with a pretty low-tech solution: a safe word.
🔐 How to Create Your Verbal Password
The idea is to pick a unique and even nonsensical word or phrase that no one could possibly guess. You should avoid anything obvious like a pet’s name, a childhood city, or a favorite hobby, as this information is often easily found online. You and your family agree that if you ever call them with an urgent request for money, they must first ask for this special phrase.
If the person on the other end can’t provide it, they know to hang up immediately and verify your safety through another channel. This simple step can be the difference between a laugh and a financial nightmare.
📱 Beyond the Verbal Password
In addition to a verbal password, you should be mindful of how much information you share online. Social media makes it incredibly easy for scammers to grab audio and video samples of your voice. The article suggests taking an extra step: tell your family to not just ask for the random password, but to also ask you (or “you”) to tell them something only the two of you know.
This dual-layer of security makes it even harder for an AI impersonation to succeed. By being proactive and having these conversations with your family, you can empower them to recognize and avoid these increasingly sophisticated scams.
Works Cited
Yee, Alaina. “AI can easily impersonate you. This trick helps thwart scammers.” PCWorld, Aug. 2025, pp. 87-88.