Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2025, Vol. 46 ›› Issue (1): 104-113.DOI: 10.11996/JG.j.2095-302X.2025010104

• Image Processing and Computer Vision • Previous Articles     Next Articles

Deepfake detection method based on multi-feature fusion of frequency domain and spatial domain

DONG Jiale1(), DENG Zhengjie1,2(), LI Xiyan1, WANG Shiyun1   

  1. 1. College of Information Science and Technology, Hainan Normal University, Haikou Hainan 571127, China
    2. Guangxi Key Laboratory of Image Processing and Intelligent Analysis, Nanning Guangxi 541004, China
  • Received:2024-08-22 Accepted:2024-11-21 Online:2025-02-28 Published:2025-02-14
  • Contact: DENG Zhengjie
  • About author:First author contact:

    DONG Jiale (1999-), master student. Her main research interest covers security of artificial intelligence systems. E-mail:dongjiale1107@163.com

  • Supported by:
    Hainan Provincial Natural Science Foundation(623QN236);Haikou Science and Technology Plan Project(2022-007);Open Funds from Guilin University of Electronic Technology, Guangxi Key Laboratory of Image and Graphic Intelligent Processing(GIIP2012)

Abstract:

In today's society, the rapid advancement of facial forgery technology has posed a substantial challenge to social security, especially in the context where deep learning techniques have been widely employed to generate realistic fake videos. These high-quality forged contents not only threaten personal privacy but can also be utilized for illegal activities. Faced with this challenge, traditional forgery detection methods based on single features have become inadequate to meet detection demands. To address this issue, a deepfake detection method based on multi-feature fusion in both frequency and spatial domains was proposed to enhance the detection accuracy and generalization capability for facial forgeries. The frequency domain was dynamically divided into three bands to extract forgery artifacts that cannot be mined in the spatial domain. The spatial domain employed the EfficientNet_b4 network and Transformer architecture to segment image blocks at multiple scales, calculate differences between different blocks, perform detection based on consistency information between upper and lower image blocks, and capture more detailed forgery feature information. Finally, a fusion block using a query-key-value mechanism integrated the methods from the frequency and spatial domains, thereby more comprehensively mining feature information from both domains to enhance the accuracy and transferability of forgery detection. Extensive experimental results confirmed the effectiveness of the proposed method, demonstrating significantly superior performance compared to traditional deepfake detection methods.

Key words: deepfake detection, EfficientNet_b4 network, frequency domain features, spatial domain features, feature fusion

CLC Number: