Deepfake Images, Video & Voice Scams

Deepfake scams use artificial intelligence to create fake images, videos, voices, identities and conversations that appear real enough to manipulate trust.

Key warning

Seeing or hearing someone is no longer enough proof that the message is real.

A cloned voice, fake video call, edited image or AI-generated profile can be used to pressure you into sending money, sharing information or trusting a false identity.

AI impersonation alert

“It sounded exactly like them.”

Deepfake scams can imitate a family member, celebrity, company leader, recruiter, investor, bank employee or public figure.

Common trick

“I’m in trouble. Please send money now and don’t tell anyone.”

Common fake proof

A realistic voice note, video call, photo, advert or social profile.

⚠️ Deepfake content can look convincing. Verify through a separate trusted route before acting.




🤖

What is a deepfake scam?

A deepfake scam uses AI-generated or AI-manipulated media to make a person, message or identity appear genuine. This can include fake profile photos, altered videos, synthetic voices, cloned video calls, fake adverts or realistic-looking screenshots.

The aim is usually to create trust quickly. The scammer may pretend to be someone you know, someone famous, a trusted professional, a senior manager, a romantic interest, a recruiter, a bank employee or a family member in distress.

ScamAdvisory rule

If the request is urgent, emotional or financial, verify outside the conversation.

👁️

Why deepfakes are dangerous

Deepfakes attack one of the strongest human trust signals: recognition. If a person appears to sound, look or speak like someone you trust, your guard may drop.

Scammers use this to make fake requests feel real. They may create panic, secrecy, urgency or authority so you act before checking.

Common targets

  • • Families receiving urgent calls from a cloned voice.
  • • Businesses receiving fake executive payment instructions.
  • • Investors seeing fake celebrity or expert endorsements.
  • • Dating app users speaking to AI-generated identities.
  • • Social media users trusting fake videos, images or adverts.
Deepfake scam types

Images, video and voice can all be faked

The danger is not only the technology. The danger is the trust created by the fake content.

🖼️

Deepfake images

AI-generated profile photos, fake identity documents, fake screenshots, fake lifestyle images or manipulated evidence.

🎥

Deepfake video

Fake videos showing celebrities, public figures, executives or trusted experts promoting scams or giving instructions.

🎙️

Voice cloning

A cloned voice can imitate a loved one, boss, colleague, bank worker or official during an urgent phone call.

📞

Family emergency scam

The caller sounds like a relative and claims they need money immediately for an accident, arrest, travel issue or crisis.

👔

Business impersonation

A fake voice or video appears to come from a senior manager requesting payment, data, access or secrecy.

📈

Fake investment endorsement

A deepfake advert uses a known person’s image or voice to promote a fake investment, crypto scheme or trading group.

Warning signs

Trust the request less than the face or voice

Deepfake scams are designed to make you focus on who the person appears to be, instead of what they are asking you to do.

Risk level

Critical

⏱️

Urgency

You are pushed to act immediately, especially to send money, approve access or keep quiet.

🤫

Secrecy

They ask you not to tell family, colleagues, your bank or anyone who could verify the request.

💸

Money request

The content leads to a transfer, gift card, crypto wallet, payment link or new bank details.

📲

New contact route

The person suddenly contacts you from a new number, new account, new email or unfamiliar platform.

🧩

Small inconsistencies

The voice, lighting, lip movement, wording, timing, background or behaviour feels slightly wrong.

🔐

Sensitive information

They ask for passwords, codes, bank details, documents, private images or account access.

What you should do

Verify the person through a separate trusted route.

1

End the call or pause the chat if money, secrecy, fear or urgency is involved.

2

Contact the person back using a phone number, email or method you already trust.

3

Use a family or workplace safe word for urgent money, travel, medical, access or payment requests.

4

For businesses, require two-person approval for payments, new payees, bank detail changes and urgent executive requests.

5

Do not rely on video, voice or screenshots alone. Verify the request, not just the appearance.

6

If you sent money or shared sensitive data, contact your bank, platform or organisation immediately.

ScamAdvisory

In the age of AI, verification matters more than recognition.

Deepfake images, video and voice can make fraud feel personal and believable. Slow down, use trusted contact routes, agree safe words, and never act on urgent requests without independent verification.

© 2025-2026 ScamAdvisory. All rights reserved.