Deepfake Detection Technology Integrating Spatial Domain and Frequency Domain

Authors

  • Haoqi Geng
  • Tianliang Lu
  • Wanxin Huang
  • Bowen Ding

DOI:

https://doi.org/10.54097/yrahtw96

Keywords:

Deepfake Detection, Spatial Domain, Frequency Domain, Fecam

Abstract

 In recent years, with the rapid development of deep forgery technology, its verisimilitude is increasing day by day, and its social impact is becoming more and more serious. However, although a variety of face deep forgery video detection algorithms have been proposed and have shown certain detection capabilities on open source data sets, in the face of increasingly sophisticated deep forgery technology, the differences between genuine and fake videos are gradually difficult to be captured by the naked eye, and existing detection methods generally have problems such as low cross-compressibility detection and poor robustness. Therefore, in order to improve the detection accuracy and model robustness, a deep forged video detection method named SFDT is proposed. In this scheme, the framework structure of fusion of air frequency domain is adopted. Firstly, feature extraction is enhanced by improving MVIT in the airspace and using ASFF adaptive module; secondly, frequency domain features are extracted by dynamic filter in the frequency domain; then FECAM is used to reduce the loss caused by frequency domain information transmission; finally, multi-mode fusion module is used for feature fusion. This air-frequency fusion detection scheme can not only improve the cross-compression detection performance of the model, but also effectively deal with various interference in the transmission of video, and improve the robustness of the model.

Downloads

Download data is not yet available.

References

[1] Mohammad Rami Koujan, Michail Christos Doukas, Anastasios Roussos, and Stefanos Zafeiriou. 2020. Head2head: Video-based neural head synthesis. arXiv preprint arXiv: 2005. 10954 (2020). Mohammad Rami Koujan, Michail Christos Doukas, Anastasios Roussos, and Stefanos Zafeiriou. 2020. Head2head: Video-based neural head synthesis. arXiv preprint arXiv:2005.10954.

[2] Goodfellow I, Pouget-Abadie J, Mirza M, et al. Generative adversarial networks[J]. Communications of the ACM, 2020, 63(11): 139-144.

[3] Rana M S, Nobi M N, Murali B, et al. Deepfake detection: A systematic literature review[J]. IEEE access, 2022, 10: 25494-25513.

[4] Suthaharan S, Suthaharan S. Support vector machine[J]. Machine learning models and algorithms for big data classification: thinking with examples for effective learning, 2016: 207-235.

[5] Afchar D, Nozick V, Yamagishi J, et al. Mesonet: a compact facial video forgery detection network[C]//2018 IEEE international workshop on information forensics and security (WIFS). IEEE, 2018: 1-7.

[6] Lu X, Firoozeh Abolhasani Zadeh Y A. Deep Learning‐Based Classification for Melanoma Detection Using XceptionNet [J]. Journal of Healthcare Engineering, 2022, 2022(1): 2196096.

[7] Sun B, Liu G, Yuan Y. F3-Net: Multiview scene matching for drone-based geo-localization[J]. IEEE Transactions on Geoscience and Remote Sensing, 2023, 61: 1-11.

[8] Yu Y, Si X, Hu C, et al. A review of recurrent neural networks: LSTM cells and network architectures[J]. Neural computation, 2019, 31(7): 1235-1270.

[9] Fan H, Xiong B, Mangalam K, et al. Multiscale vision transformers[C]//Proceedings of the IEEE/CVF international conference on computer vision. 2021: 6824-6835.

[10] Chen M, Peng H, Fu J, et al. Autoformer: Searching transformers for visual recognition[C]//Proceedings of the IEEE/CVF international conference on computer vision. 2021: 12270-12280.

[11] Wen G, Li Z, Azizzadenesheli K, et al. U-FNO—An enhanced Fourier neural operator-based deep-learning model for multiphase flow[J]. Advances in Water Resources, 2022, 163: 104180.

[12] Men Y, Yao Y, Cui M, et al. Dct-net: domain-calibrated translation for portrait stylization[J]. ACM Transactions on Graphics (TOG), 2022, 41(4): 1-9.

[13] Jiang M, Zeng P, Wang K, et al. FECAM: Frequency enhanced channel attention mechanism for time series forecasting[J]. Advanced Engineering Informatics, 2023, 58: 102158.

[14] Testa R L, Machado-Lima A, Nunes F L S. Deepfake detection on videos based on ratio images[C]//2022 12th International Congress on Advanced Applied Informatics (IIAI-AAI). IEEE, 2022: 403-408.

[15] Rossler A, Cozzolino D, Verdoliva L, et al. Faceforensics++: Learning to detect manipulated facial images[C]// Proceedings of the IEEE/CVF international conference on computer vision. 2019: 1-11.

[16] Li, Yuezun, et al. "Celeb-df: A large-scale challenging dataset for deepfake forensics." Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2020.

[17] Jiang, Liming, et al. "Deeperforensics-1.0: A large-scale dataset for real-world face forgery detection." Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2020.

[18] Li, Lingzhi, et al. "Faceshifter: Towards high fidelity and occlusion aware face swapping." arXiv preprint arXiv: 1912. 13457 (2019).

[19] Croitoru F A, Hiji A I, Hondru V, et al. Deepfake Media Generation and Detection in the Generative AI Era: A Survey and Outlook[J]. arXiv preprint arXiv:2411.19537, 2024.

[20] Wan Da. Research on face forgery image detection method based on deep learning [D]. People's Public Security University of China,2024.DOI:10. 27634/d. cnki. gzrgu. 2024. 000268.

[21] Nguyen D, Mejri N, Singh I P, et al. Laa-net: Localized artifact attention network for quality-agnostic and generalizable deepfake detection[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024: 17395-17405.

[22] Selvaraju R R, Cogswell M, Das A, et al. Grad-cam: Visual explanations from deep networks via gradient-based localization[C]//Proceedings of the IEEE international conference on computer vision. 2017: 618-626.

Downloads

Published

27-03-2025

Issue

Section

Articles

How to Cite

Geng, H., Lu, T., Huang, W., & Ding, B. (2025). Deepfake Detection Technology Integrating Spatial Domain and Frequency Domain. Frontiers in Computing and Intelligent Systems, 11(3), 54-62. https://doi.org/10.54097/yrahtw96