Open Access iconOpen Access

ARTICLE

crossmark

Real-Time Deepfake Detection via Gaze and Blink Patterns: A Transformer Framework

Muhammad Javed1, Zhaohui Zhang1,*, Fida Hussain Dahri2, Asif Ali Laghari3,*, Martin Krajčík4, Ahmad Almadhor5

1 Department of Computer Science and Technology, College of Computer Science, Donghua University, Shanghai, 200022, China
2 School of Computer Science and Engineering, Southeast University, Nanjing, 211189, China
3 Software College, Shenyang Normal University, Shenyang, 110136, China
4 Department of Information Management and Business Systems, Faculty of Management, Comenius University, Bratislava Odbojárov 10, Bratislava, 82005, Slovakia
5 Department of Computer Engineering and Networks, College of Computer and Information Sciences, Jouf University, Sakaka, 72388, Saudi Arabia

* Corresponding Authors: Zhaohui Zhang. Email: email; Asif Ali Laghari. Email: email

Computers, Materials & Continua 2025, 85(1), 1457-1493. https://doi.org/10.32604/cmc.2025.062954

Abstract

Recent advances in artificial intelligence and the availability of large-scale benchmarks have made deepfake video generation and manipulation easier. Therefore, developing reliable and robust deepfake video detection mechanisms is paramount. This research introduces a novel real-time deepfake video detection framework by analyzing gaze and blink patterns, addressing the spatial-temporal challenges unique to gaze and blink anomalies using the TimeSformer and hybrid Transformer-CNN models. The TimeSformer architecture leverages spatial-temporal attention mechanisms to capture fine-grained blinking intervals and gaze direction anomalies. Compared to state-of-the-art traditional convolutional models like MesoNet and EfficientNet, which primarily focus on global facial features, our approach emphasizes localized eye-region analysis, significantly enhancing detection accuracy. We evaluate our framework on four standard datasets: FaceForensics, CelebDF-V2, DFDC, and FakeAVCeleb. The proposed framework results reveal higher accuracy, with the TimeSformer model achieving accuracies of 97.5%, 96.3%, 95.8%, and 97.1%, and with the hybrid Transformer-CNN model demonstrating accuracies of 92.8%, 91.5%, 90.9%, and 93.2%, on FaceForensics, CelebDF-V2, DFDC, and FakeAVCeleb datasets, respectively, showing robustness in distinguishing manipulated from authentic videos. Our research provides a robust state-of-the-art framework for real-time deepfake video detection. This novel study significantly contributes to video forensics, presenting scalable and accurate real-world application solutions.

Keywords

Deepfake detection; deep learning; video forensics; gaze and blink patterns; transformers; TimeSformer; MesoNet4

Cite This Article

APA Style
Javed, M., Zhang, Z., Dahri, F.H., Laghari, A.A., Krajčík, M. et al. (2025). Real-Time Deepfake Detection via Gaze and Blink Patterns: A Transformer Framework. Computers, Materials & Continua, 85(1), 1457–1493. https://doi.org/10.32604/cmc.2025.062954
Vancouver Style
Javed M, Zhang Z, Dahri FH, Laghari AA, Krajčík M, Almadhor A. Real-Time Deepfake Detection via Gaze and Blink Patterns: A Transformer Framework. Comput Mater Contin. 2025;85(1):1457–1493. https://doi.org/10.32604/cmc.2025.062954
IEEE Style
M. Javed, Z. Zhang, F. H. Dahri, A. A. Laghari, M. Krajčík, and A. Almadhor, “Real-Time Deepfake Detection via Gaze and Blink Patterns: A Transformer Framework,” Comput. Mater. Contin., vol. 85, no. 1, pp. 1457–1493, 2025. https://doi.org/10.32604/cmc.2025.062954



cc Copyright © 2025 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 2760

    View

  • 2086

    Download

  • 0

    Like

Share Link