CircadifyCircadify
Fraud Prevention8 min read

How rPPG Detects Deepfakes: Blood Flow Analysis Explained

A research-grade analysis of how remote photoplethysmography (rPPG) detects deepfakes by measuring sub-dermal blood flow signals that synthetic media cannot replicate.

tryfacescan.com Research Team·

How rPPG Detects Deepfakes: Blood Flow Analysis Explained

Generative adversarial networks can now produce face-swaps that fool human reviewers more than 40 percent of the time, according to a 2024 study published in IEEE Transactions on Information Forensics and Security. For identity verification vendors, banks, and fintech fraud teams, the question is no longer whether deepfakes will target onboarding flows — it is how quickly defenses can evolve beyond pixel-level inspection. The answer emerging from biometrics research is that rPPG detects deepfakes through blood flow signals embedded in ordinary video, exposing a physiological layer that generative models have not learned to synthesize.

"The cardiovascular pulse imprints micro-color oscillations onto facial skin at every heartbeat. A deepfake generator would need to model hemodynamics, not just appearance, to survive rPPG analysis." — Adapted from conclusions in Frontiers in Signal Processing, Vol. 3, 2024.

How rPPG Captures the Blood Flow Signal

Remote photoplethysmography (rPPG) extracts the blood volume pulse (BVP) from video frames by isolating periodic chromatic changes in facial skin regions. When the left ventricle contracts, oxygenated hemoglobin briefly raises absorption in the green channel (approximately 540 nm). Standard RGB cameras record this fluctuation at amplitudes below conscious perception — typically 0.5 to 2 percent intensity variation per cardiac cycle.

The processing pipeline follows a well-documented chain:

  1. Face detection and ROI tracking — forehead and cheek regions provide the strongest pulse signal due to dense superficial vasculature.
  2. Spatial averaging and temporal filtering — raw RGB traces are averaged across the ROI, then bandpass-filtered to the plausible cardiac range (0.7–4.0 Hz, corresponding to 42–240 BPM).
  3. Signal decomposition — algorithms such as CHROM (de Haan & Jeanne, 2013) or POS (Wang et al., 2017) separate the pulse component from motion artifacts and illumination drift.
  4. Pulse quality scoring — the resulting waveform is evaluated for periodicity, signal-to-noise ratio, and inter-beat interval consistency.

A live human face produces a coherent, quasi-periodic BVP waveform. A deepfake — whether a GAN-generated face swap, a neural rendering, or a replayed video of a screen — either produces no detectable pulse or generates an incoherent signal that fails quality thresholds.

Why Deepfakes Fail Blood Flow Analysis

The core vulnerability of synthetic face media lies in what generative models optimize for: perceptual realism at the frame level. Training objectives like L1 reconstruction loss, perceptual loss, and adversarial loss operate on spatial appearance, not on temporal physiological coherence across multi-second windows.

Research from Hernandez-Ortega et al. (2024, Pattern Recognition, Vol. 148) demonstrated that even state-of-the-art face-swap architectures produce rPPG power spectra that are statistically distinguishable from live subjects. The key failure modes include:

Failure Mode What Happens Why It Matters
Absent pulse No dominant frequency in the cardiac band Printed photos, 3D masks, and screen replays lack hemodynamic variation entirely
Flat spectrum Energy distributed uniformly across frequencies GAN outputs introduce stochastic noise rather than periodic biological signals
Frequency mismatch Dominant frequency drifts or falls outside normal cardiac range Frame interpolation and temporal blending corrupt pulse timing
Spatial incoherence Different facial regions yield conflicting pulse phases Face-swap boundary artifacts break the spatially uniform blood flow pattern
Missing harmonics Fundamental frequency present but second/third harmonics absent Authentic cardiac waveforms contain harmonic structure from arterial compliance
Low signal-to-noise ratio Pulse amplitude well below physiological norms Compression artifacts and rendering quantization suppress micro-color variations

A 2023 analysis by Nowara, Stampfer, and McDuff published through NeurIPS Workshop on Synthetic Realities confirmed that rPPG-based classifiers achieved area-under-curve (AUC) scores above 0.97 on multiple deepfake benchmarks, including FaceForensics++ and Celeb-DF v2, without requiring any deepfake-specific training data — the classifier simply measured whether a physiologically plausible pulse was present.

Applications in Identity Verification Pipelines

For KYC providers and bank fraud teams, rPPG-based liveness detection slots into existing onboarding workflows at a specific decision point: the selfie-video capture step. Rather than asking the user to perform challenge-response actions (blink, smile, turn head), the system passively analyzes the video stream for blood flow evidence during natural behavior.

This approach addresses several operational pain points:

Presentation attack resistance — the 2024 update to ISO/IEC 30107-3 (biometric presentation attack detection) explicitly recognizes physiological signal analysis as a detection mechanism. Blood flow verification covers a broad attack surface: printed photos, digital screen replays, silicone masks, and AI-generated face swaps all lack authentic cardiovascular signals.

Reduced friction — passive analysis eliminates the need for users to follow animated prompts. For mobile onboarding flows where drop-off rates already run 20–35 percent (Signicat "The Battle to Onboard" report, 2024), removing challenge-response steps directly impacts conversion.

Injection attack detection — when combined with camera integrity checks, rPPG analysis can flag virtual camera injections where a pre-recorded or AI-generated video is fed into the verification pipeline. Injected video lacks the noise profile and subtle temporal characteristics of a live camera feed, making the rPPG signal anomalous even if the injected face appears realistic.

Research Foundations

The scientific basis for rPPG-based deepfake detection rests on more than a decade of remote physiological sensing research:

  • Verkruysse et al. (2008) — first demonstrated that BVP could be recovered from ambient-light face video, published in Optics Express. This foundational work established that commodity cameras contain sufficient spectral sensitivity to capture hemodynamic fluctuations.
  • Li, Yang, Liao, et al. (2016) — among the first to propose using rPPG signals specifically for face anti-spoofing, showing that 2D and 3D presentation attacks lack the cardiac rhythm present in live captures.
  • Ciftci, Demir, and Yin (2020) — published "FakeCatcher" in IEEE TPAMI, a dedicated deepfake detection system that extracts PPG maps from facial regions and achieves high separability between authentic and manipulated videos.
  • Yu, Zhao, et al. (2023) — introduced TransrPPG, a transformer-based architecture that captures long-range temporal dependencies in the rPPG signal, improving robustness under compression and low-resolution conditions commonly encountered in mobile onboarding.
  • Hernandez-Ortega et al. (2024) — conducted a comprehensive evaluation across multiple deepfake generation methods, confirming that rPPG-based features remain discriminative even against the latest diffusion-model-based face synthesis techniques.

The Future of Physiological Liveness Signals

Three developments are poised to expand the role of rPPG in fraud prevention over the next several years.

Multi-signal fusion — researchers at ETH Zurich and the Chinese Academy of Sciences have demonstrated that combining rPPG with micro-expression timing and pupillary light reflex analysis produces complementary liveness evidence. A fused approach raises the bar for attackers from replicating one physiological channel to replicating several simultaneously.

Edge deployment — neural-network-based rPPG extraction models are being optimized for on-device inference. MediaPipe-style architectures running on mobile NPUs can process rPPG in real time without transmitting raw video to a server, addressing data residency and latency concerns that regulated industries face.

Adversarial robustness — as awareness of rPPG-based detection grows, researchers anticipate adversarial attempts to inject fake pulse signals into synthetic video. Preemptive work by Hou et al. (2024, ACM Computing Surveys) on physiological signal authentication — including pulse waveform shape analysis and heart rate variability metrics — is building defenses against this next generation of attacks.

Frequently Asked Questions

How does rPPG differ from contact-based photoplethysmography?

Contact PPG uses a dedicated sensor (such as a pulse oximeter clip) pressed against the skin. rPPG performs the same measurement remotely using reflected light captured by a standard camera. For identity verification, this distinction is critical: rPPG requires no additional hardware beyond the device camera already used for selfie capture.

Can a high-quality deepfake replicate a real blood flow signal?

Current generative models do not optimize for temporal physiological coherence. A deepfake would need to model not just per-frame appearance but also the sub-pixel chromatic oscillations caused by cardiac pulsation across the entire face, synchronized with realistic heart rate variability. No publicly documented generation pipeline achieves this as of early 2026.

Does lighting or skin tone affect rPPG signal quality?

Signal strength varies with illumination level and melanin concentration. However, multi-wavelength decomposition algorithms (CHROM, POS) are designed to compensate for these variables. Research by Nowara et al. (2020, IEEE FG) showed that rPPG-based liveness detection maintained performance across Fitzpatrick skin types I through VI when sufficient ambient illumination was present.

What video duration is needed for reliable analysis?

Most rPPG algorithms require a minimum observation window of 3–5 seconds to capture multiple cardiac cycles. This aligns with typical selfie-video capture durations in KYC workflows, meaning no additional user burden is introduced.

How does rPPG handle video compression artifacts?

Lossy compression (H.264, H.265) attenuates the micro-color variations that carry the pulse signal. Robust rPPG pipelines address this through spatial averaging over larger ROIs, temporal filtering tuned to the cardiac band, and learned denoising networks trained on compressed video datasets.


Physiological liveness detection through rPPG represents a measurable shift in how identity verification systems distinguish real faces from synthetic ones. For fraud teams evaluating their deepfake defenses, blood flow analysis offers a detection layer that operates on biology rather than pixel statistics.

Explore how Circadify integrates rPPG-based liveness detection into fraud prevention pipelines.

Request Enterprise Demo