What Is a Deepfake? Definition, Examples & Detection (2026)

Uncover everything you need to know about deepfakes, the dangers, and how deepfake detection works.

India’s financial system runs on digital trust. Crores of people open bank accounts, apply for loans, and complete KYC from their phones every single day. 

But that trust is now under attack.

Fraud Detected

Deepfake cases in India have jumped 550% since 2019, with projected losses hitting ₹70,000 crore. The RBI reported fraud losses of ₹21,367 crore in just the first half of FY2024–25,  an eightfold increase year-on-year. A growing share of that fraud involves AI-generated fake identities.

For banks, NBFCs, and fintechs running digital onboarding, deepfakes are no longer something to monitor from a distance. This guide answers what is a deepfake, the mechanism behind it, and how you can safeguard your business against this threat. 

TL;DR: Deepfake Threats & Detection in India (2026)

The Crisis: Deepfake-enabled fraud in India has surged by 550% since 2019, with projected financial losses reaching ₹70,000 crore. The RBI reported an eightfold year-on-year increase in fraud losses (₹21,367 crore) in the first half of FY2024–25, driven largely by AI-generated synthetic identities.

Core Technologies Used by Fraudsters:

  • GANs & Diffusion Models: Used to create hyper-realistic “synthetic identities” that pass liveness checks.
  • Real-time Face-Swaps: Used to spoof Video KYC (V-CIP) by overlaying stolen identities onto fraudsters during live calls.
  • Voice Cloning: Mimicking customer voices to bypass OTPs and authorize illegal transfers.

2026 Regulatory Update

The IT Amendment Rules 2026 (effective Feb 20, 2026) formally define “synthetic media.” Major platforms must now remove flagged deepfakes within 3 hours and non-consensual intimate content within 2 hours or risk losing safe harbor protection under Section 79 of the IT Act.

Detection & Defense: Modern BFSI defense relies on Passive Liveness Detection, which analyzes biological signals (blood flow, micro-expressions) and Injection Attack Detection to ensure video feeds are coming from a live camera, not a pre-recorded file.

How are deepfakes made?

Before we get into how deepfakes are made, here’s a quick recap on what is a deepfake. 

Imagine someone takes photos, videos, and voice recordings of a real person, feeds them into an AI model, and out comes a fake version of that person: one that looks real, sounds real, and can say or do things the real person never did. That output is a deepfake. The name combines deep learning (the AI method used) and fake (the output). 

But not all deepfakes are built the same way every time. Fraudsters pick the method that fits the attack. Let’s read about them below:

1. Generative Adversarial Networks (GANs)

GANs use two competing neural networks; one tries to create convincing fake content, the other tries to detect it. They run against each other in a loop. Over thousands of rounds, the content becomes indistinguishable from the real thing. 

This adversarial training process is behind most high-quality deepfake videos used in KYC fraud today.

In practice: A fraudster trains a GAN on someone’s social media photos. Within hours, the model can generate that person’s face at different angles, lighting conditions, and expressions, enough to pass a liveness check.

2. Diffusion models

Diffusion models work differently. They start with real images, add random noise until the image is unrecognisable, then learn how to reverse that process to reconstruct a new, photorealistic image. Tools like Stable Diffusion use this method and are available as free, open-source software.

In practice: Fraudsters use diffusion models to generate entirely new face images of people who don’t exist, paired with forged documents. These ‘synthetic identities’ are then used to open accounts or apply for credit.

3. Face-swap apps

Face-swap tools overlay one person’s face onto another’s video, in real time. These use encoder-decoder architectures that map facial features from a source face onto the geometry of a target video, matching lighting and movement automatically.

In practice: During a live video KYC call, the fraudster’s real face is replaced on-screen with a stolen identity in real time. The bank employee or automated system sees the wrong face.

4. Voice cloning

AI voice synthesis tools extract the acoustic patterns from a short audio sample and use them to generate speech in that person’s voice. 

In practice: A fraudster records 30 seconds of a customer’s voice from a previous call. They clone it and call the customer’s bank, impersonating the account holder, to extract an OTP or request a fund transfer.

Let’s quickly explore some real-life cases that depict the level of damage deepfakes can inflict. 

Deepfake real-world examples (BFSI-focused)

#CaseWhat HappenedFraud TypeBFSI Relevance
1Bengaluru, 2025A 79-year-old woman lost ₹35 lakhs after watching deepfaked videos of N.R. Narayana Murthy endorsing a fake trading platform. False profit dashboards and invented ‘financial managers’ kept her engaged until the money was gone. Celebrity impersonation + investment fraudInvestment fraud via deepfaked public figures is now a recurring pattern in India. Customers of digital lending and trading platforms are primary targets.
2Valueleaf Scam, Mumbai, 2025A China-based syndicate ran deepfaked videos of known Indian stock analysts and news anchors as paid ads on Meta platforms between July 1 and 18, 2025. Victims were redirected to WhatsApp groups where scammers posed as financial advisors. Four Valueleaf employees were arrested in October 2025 for knowingly providing advertising access to the syndicate for ₹3 crore.Deepfake ads + social engineering + mule accountsExploited India’s digital advertising infrastructure. The fraud ran on Meta’s credit-based system before detection. 
3Arup CFO Scam, Hong Kong, 2024A finance employee was called into a video conference where the company’s CFO and several senior colleagues appeared on screen. All were AI-generated. The employee authorised transfers totalling USD $25 million (~₹210-230 crore).Real-time deepfake video conference + BECThis is the most cited financial deepfake case globally. Indian banks conducting video-based credit approvals, treasury authorisations, or executive sign-offs face the same risk.

Further, let’s study how the Indian regulatory landscape protects against the deepfake threat. 

Currently, India does not have a standalone AI deepfake law yet. However, several existing statutes already cover the most common forms of deepfake-enabled harm, from identity fraud to non-consensual imagery. 

Law / SectionWhat it covers
IT Act, 2000: Section 66C (Identity theft)Penalises the fraudulent use of another person’s electronic identity, password, or biometric credentials using any computer resource or communication device. 
IT Act, 2000: Section 66D (Cheating by Personation)Covers anyone who cheats by impersonating another person through a computer resource (the most cited provision in video KYC fraud and executive impersonation cases).
IT Act, 2000 — Sections 66E, 67, 67A & 67B (Privacy & obscene content)Section 66E targets non-consensual capture or transmission of private images. Sections 67 and 67A penalise publishing obscene or sexually explicit electronic content. Section 67B specifically addresses child sexual abuse material. 
BNS, 2023: Section 356 (Defamation)Covers anyone who makes or publishes a false statement intending to damage a person’s reputation. 
BNS, 2023: Sections 351 & 77 (Criminal intimidation & privacy)Section 351 penalises threatening a person with injury to cause alarm or compel action. Section 77 specifically covers capturing or sharing images of a woman in a private act without consent.
Digital Personal Data Protection Act, 2023 (DPDP Act)Governs how personal data  (including biometric and facial data collected during KYC) is processed. Penalties up to ₹250 crore per violation.

On 10 February 2026, MeitY notified India’s IT Amendment Rules 2026, effective 20 February 2026. For the first time, Indian law formally defines synthetic media and places clear, time-bound obligations on platforms. 

Significant social media intermediaries must remove flagged deepfake content within 3 hours, and non-consensual intimate content within 2 hours. Platforms that miss these windows risk losing safe harbour protection under Section 79 of the IT Act, meaning they can be held liable as if they created the content themselves.

The threat to KYC & identity verification 

Every financial institution in India that offers digital onboarding, whether through video KYC, selfie liveness, document upload, or biometrics, relies on some form of identity verification. 

Deepfakes can compromise each of these layers. Here is how.

1. Video KYC spoofing

Around 11 lakh video KYC calls take place in India every day, and 86% face spoofing attempts. Fraudsters use real-time face-swap tools to replace their face on screen with a stolen identity during a live call. A bank agent or an automated system sees the wrong person entirely.

2. Injection attacks

More advanced attackers skip the camera entirely. They inject a pre-recorded or synthetically generated video directly into the device’s camera input, so the system receives a fake feed. Standard liveness checks that look for blinks or head movement do not catch this because there’s no real person on the other side at all.

3. Synthetic identity fraud

Fraudsters can use deepfake technology to generate a near-perfect human face that belongs to no living person and pair that with forged documents (acquired through identity theft, for instance)to build a clean synthetic identity: one that passes basic document checks and has no fraud history to flag.

4. Voice cloning for OTP bypass

India saw a 100% spike in voice scam cases between early 2024, rising from 15% to nearly 35% of all reported social engineering fraud. Cloned voices are used to call bank staff or customers: impersonating known contacts, branch managers, or family members to extract OTPs, approve transactions, or redirect fund transfers.

HyperVerge’s deepfake detection API checks for face liveness, injection signals, and GAN artefacts in real time during onboarding. 
See how it works

How does deepfake detection work?

Almost every AI-generated face leaves behind traces, most invisible to the human eye, but detectable through the right analysis. 

Detection systems work by hunting for those traces across multiple dimensions simultaneously.

  • Spotting the AI’s fingerprint

Every GAN model leaves a unique mark on the images it produces, a kind of digital fingerprint shaped by its architecture, training data, and internal settings. Detection systems are trained to recognise these fingerprints across different GAN models, even ones they haven’t encountered before.

  • Reading frequency patterns

Real photographs and AI-generated images behave differently when broken down into frequency components. Detection systems apply transforms: Fourier, wavelet, and Histogram of Oriented Gradients (HOG) to convert a face image into its underlying frequency data. Deepfakes consistently produce anomalies at this level: ringing, blurring, unnatural edges, and blending artefacts that don’t appear in genuine images.

  • Analysing colour channel relationships

In a real photo, the red, green, and blue colour channels maintain a natural correlation,  the result of how light interacts with real surfaces like human skin. When a face is synthetically generated or swapped, these correlations break down. Detection systems measure those statistical deviations to flag manipulation.

Why does layering matter?

No single method catches everything. The research underpinning modern detection shows that combining frequency analysis with colour-channel correlation examination consistently outperforms any standalone approach (in some configurations reaching 100% classification accuracy across multiple GAN models).

And besides this, active and passive liveness detection also ensures deepfakes stay at bay. Let’s read more about them below. 

Passive vs. active detection: What’s the difference?

When evaluating a deepfake detection or liveness solution, one question matters most: Does it require user action, or does it work in the background?

Active liveness detection prompts the user to blink, turn their head, or read a number aloud. The system verifies whether the response looks genuine. But here’s the problem. Real-time deepfake tools can now mirror these actions in milliseconds. An active challenge is a known prompt that fraudsters can prepare for. 

Active checks also add friction to the onboarding journey, which costs conversions in a competitive market.

Passive liveness detection runs silently during the onboarding flow. The user does nothing different. The system analyses the video stream for biological and physical signals of a real, live person. There is no challenge to reverse-engineer.

Passive detection is harder to spoof because the fraudster does not know which signals are being checked. It also creates zero friction for legitimate users.

Detection methodHow it worksBest for
Active livenessPrompts user action (blink, nod, read aloud) and checks if the response is genuineBasic anti-spoofing; lower-volume or lower-risk onboarding
Passive livenessSilently analyzes blood flow, micro-expressions, and depth during the live sessionHigh-volume digital onboarding where user experience matters
GAN artefact detectionScans pixel patterns for AI-generation fingerprints at the image levelCatching synthetic identity documents and AI-generated face photos
Injection attack detectionAnalyses camera metadata, sensor noise, and frame timing for artificial feed signalsDefending against real-time face-swap tools during video KYC
Multi-layer AI detectionCombines passive liveness + injection + GAN artefact checks simultaneouslyFull-stack KYC fraud prevention: the HyperVerge approach

Deepfake statistics & threat landscape 2026

Deepfakes are no longer a niche cybersecurity risk. In India, they are a front-line fraud tool targeting video KYC, identity onboarding, investment platforms, and executive authorisations every single day.

Take a look at some of these alarming stats: 

So, one thing is clear: deepfakes have crossed from a threat-to-watch into a threat-in-progress. 

But here’s the good news: detection technology has kept pace. Passive liveness, injection detection, and GAN artefact analysis, applied in real time, give government, businesses, and financial institutions a reliable, frictionless defence. 

But the window to act keeps narrowing. With deepfake attacks occurring every five minutes globally, every unprotected onboarding flow is an open target. But not with HyperVerge.

What Is a Deepfake? Definition, Examples & Detection in India

HyperVerge’s deepfake detection solution is built specifically for India’s KYC infrastructure: real-time, API-first, and designed for the volume India demands. 

Explore it here. 

Frequently Asked Questions

A deepfake is a fake video, audio, or image created using artificial intelligence to to make it look like a real person said or did something they never actually did. AI models analyze real photos, videos, and voice recordings of a person and then generate new content that closely mimics their face, voice, or expressions.

Some deepfakes show small inconsistencies. You might notice unnatural blinking, lips not syncing properly with speech, strange lighting on the face, or slightly blurred edges around the head. Audio may also sound slightly robotic.

India does not yet have a specific law dedicated to deepfakes, but existing laws already cover many harms caused by them. For example, Section 66C of the IT Act, 2000, deals with identity theft, while Section 66D addresses cheating through online impersonation. Sections 66E, 67, and 67A cover privacy violations and obscene digital content. The DPDP Act, 2023, also regulates how personal data, such as facial or biometric information, can be used.

Fraudsters can create deepfake videos or voices to impersonate company executives, bank officials, or customers. In some cases, deepfakes may also be used to attempt identity verification fraud.

A face swap simply replaces one person’s face with another in an existing photo or video. A deepfake is more advanced because AI generates realistic facial expressions, speech, and movements, making the person appear natural in a completely fabricated scene.

Deepfake detection systems analyze videos for digital artifacts and unusual facial patterns. They examine micro-expressions, lip movements, blinking patterns, and frame inconsistencies that often appear in AI-generated media.

Yes. Some modern detection systems, like HyperVerge, analyze live video streams and identify signs of manipulation during the interaction itself. This helps stop impersonation attempts while a video call or verification session is happening.

One major concern is fraud during the Video Customer Identification Process (V-CIP) required by the Reserve Bank of India (RBI) for remote KYC. A deepfake could be used to impersonate a customer during verification, potentially allowing criminals to open fraudulent accounts or access financial services.

Nupura Ughade

Nupura Ughade

Content Marketing Lead

LinedIn
With a strong background B2B tech marketing, Nupura brings a dynamic blend of creativity and expertise. She enjoys crafting engaging narratives for HyperVerge's global customer onboarding platform.

Related Blogs

AI Fraud Is Evolving Faster Than Most Liveness Systems: What the 2026 Liveness Benchmarks Show

AI-generated identity fraud is pushing legacy liveness systems to their limits. This...

Account Aggregator vs. Manual Financial Data Tracking: Which is Better?

Struggling with delays and data errors? This guide compares manual tracking vs...

Everything You Need To Know About Reverse Penny Drop

Wondering what is reverse penny drop? Discover everything you need to know...
×

Quick intro, then the good stuff is all yours