{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T01:43:10Z","timestamp":1760060590729,"version":"build-2065373602"},"reference-count":42,"publisher":"MDPI AG","issue":"9","license":[{"start":{"date-parts":[[2025,9,12]],"date-time":"2025-09-12T00:00:00Z","timestamp":1757635200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["J. Imaging"],"abstract":"<jats:p>Currently, deepfake detection has garnered widespread attention as a key defense mechanism against the misuse of deepfake technology. However, existing deepfake detection networks still face challenges such as insufficient robustness, limited generalization capabilities, and a single feature extraction domain (e.g., using only spatial domain features) when confronted with evolving algorithms or diverse datasets, which severely limits their application capabilities. To address these issues, this study proposes a deepfake detection network named EFIMD-Net, which enhances performance by strengthening feature interaction and integrating spatial and frequency domain features. The proposed network integrates a Cross-feature Interaction Enhancement module (CFIE) based on cosine similarity, which achieves adaptive interaction between spatial domain features (RGB stream) and frequency domain features (SRM, Spatial Rich Model stream) through a channel attention mechanism, effectively fusing macro-semantic information with high-frequency artifact information. Additionally, an Enhanced Multi-scale Feature Fusion (EMFF) module is proposed, which effectively integrates multi-scale feature information from various layers of the network through adaptive feature enhancement and reorganization techniques. Experimental results show that compared to the baseline network Xception, EFIMD-Net achieves comparable or even better Area Under the Curve (AUC) on multiple datasets. Ablation experiments also validate the effectiveness of the proposed modules. Furthermore, compared to the baseline traditional two-stream network Locate and Verify, EFIMD-Net significantly improves forgery detection performance, with a 9-percentage-point increase in Area Under the Curve on the CelebDF-v1 dataset and a 7-percentage-point increase on the CelebDF-v2 dataset. These results fully demonstrate the effectiveness and generalization of EFIMD-Net in forgery detection. Potential limitations regarding real-time processing efficiency are acknowledged.<\/jats:p>","DOI":"10.3390\/jimaging11090312","type":"journal-article","created":{"date-parts":[[2025,9,12]],"date-time":"2025-09-12T12:34:03Z","timestamp":1757680443000},"page":"312","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["EFIMD-Net: Enhanced Feature Interaction and Multi-Domain Fusion Deep Forgery Detection Network"],"prefix":"10.3390","volume":"11","author":[{"given":"Hao","family":"Cheng","sequence":"first","affiliation":[{"name":"School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China"}]},{"given":"Weiye","family":"Pang","sequence":"additional","affiliation":[{"name":"School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China"}]},{"given":"Kun","family":"Li","sequence":"additional","affiliation":[{"name":"School of Information Engineering, Guilin Institute of Information Technology, Guilin 541100, China"}]},{"given":"Yongzhuang","family":"Wei","sequence":"additional","affiliation":[{"name":"School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China"}]},{"given":"Yuhang","family":"Song","sequence":"additional","affiliation":[{"name":"School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China"}]},{"given":"Ji","family":"Chen","sequence":"additional","affiliation":[{"name":"School of Mathematical and Computational Sciences, Guilin University of Electronic Technology, Guilin 541004, China"}]}],"member":"1968","published-online":{"date-parts":[[2025,9,12]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3425780","article-title":"The creation and detection of deepfakes: A survey","volume":"54","author":"Mirsky","year":"2021","journal-title":"ACM Comput. Surv."},{"key":"ref_2","first-page":"1","article-title":"A self-supervised BEiT model with a novel hierarchical patchReducer for efficient facial deepfake detection","volume":"58","author":"Fraihat","year":"2025","journal-title":"Artif. Intell. Rev."},{"key":"ref_3","unstructured":"Li, Y., and Lyu, S. (2018). Exposing DeepFake Videos By Detecting Face Warping Artifacts. arXiv."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Yang, X., Li, Y., and Lyu, S. (2019, January 12\u201317). Exposing deep fakes using inconsistent head poses. Proceedings of the ICASSP 2019-2019 IEEE International Conference On Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK.","DOI":"10.1109\/ICASSP.2019.8683164"},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1145\/3422622","article-title":"Generative adversarial networks","volume":"63","author":"Goodfellow","year":"2020","journal-title":"Commun. ACM"},{"key":"ref_6","unstructured":"Ho, J., Jain, A., and Abbeel, P. (2020, January 6\u201312). Denoising diffusion probabilistic models. Proceedings of the NIPS\u201920: 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"1039","DOI":"10.1109\/TIFS.2022.3233774","article-title":"F2Trans: High-Frequency Fine-Grained Transformer for Face Forgery Detection","volume":"18","author":"Miao","year":"2023","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Liu, H., Li, X., Zhou, W., Chen, Y., He, Y., Xue, H., Zhang, W., and Yu, N. (2021, January 20\u201325). Spatial-phase shallow learning: Rethinking face forgery detection in frequency domain. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.00083"},{"key":"ref_9","unstructured":"Li, L., Liu, J., Wang, S., Zhang, K., Lau, R.W.H., and Chen, M. (November, January 29). UMMAFormer: A Universal Multimodal-adapter Transformer Framework for Temporal Forgery Localization. Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada."},{"key":"ref_10","unstructured":"Hussein, S.A., Tirer, T., and Giryes, R. (2020, January 7\u201312). Image-adaptive GAN based reconstruction. Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA."},{"key":"ref_11","unstructured":"Guo, M., Yin, Q., Lu, W., and Luo, X. (2025). Towards Open-world Generalized Deepfake Detection: General Feature Extraction via Unsupervised Domain Adaptation. arXiv."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., and Sun, G. (2018, January 18\u201323). Squeeze-and-excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref_13","unstructured":"(2025, March 26). DeepFakes GitHub Repository. Available online: https:\/\/github.com\/deepfakes\/faceswap."},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Thies, J., Zollhofer, M., Stamminger, M., Theobalt, C., and Nie\u00dfner, M. (2016, January 27\u201330). Face2face: Real-time face capture and reenactment of rgb videos. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.262"},{"key":"ref_15","unstructured":"(2025, March 30). FaceSwap GitHub Repository. Available online: https:\/\/github.com\/marekkowalski\/FaceSwap."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., and Aila, T. (2019, January 15\u201320). A style-based generator architecture for generative adversarial networks. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00453"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., and Aila, T. (2020, January 13\u201319). Analyzing and improving the image quality of stylegan. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. (2022, January 18\u201324). High-resolution image synthesis with latent diffusion models. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref_19","unstructured":"Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., and Chen, M. (2022). Hierarchical text-conditional image generation with clip latents. arXiv."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Li, Y., Yang, X., Sun, P., Qi, H., and Lyu, S. (2020, January 13\u201319). Celeb-df: A large-scale challenging dataset for deepfake forensics. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00327"},{"key":"ref_21","unstructured":"Yang, X., Li, Y., Qi, H., and Lyu, S. Exposing GAN-synthesized faces using landmark locations. Proceedings of the ACM Workshop on Information Hiding and Multimedia Security."},{"key":"ref_22","unstructured":"Rossler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., and Nie\u00dfner, M. (November, January 27). Faceforensics++: Learning to detect manipulated facial images. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Seoul, Republic of Korea."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Chollet, F. (2017, January 21\u201326). Xception: Deep learning with depthwise separable convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.195"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Zhao, H., Zhou, W., Chen, D., Wei, T., Zhang, W., and Yu, N. (2021, January 20\u201325). Multi-attentional deepfake detection. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.00222"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Durall, R., Keuper, M., and Keuper, J. (2020, January 18\u201324). Watch your up-convolution: Cnn based generative deep neural networks are failing to reproduce spectral distributions. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPR42600.2020.00791"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Marra, F., Gragnaniello, D., Verdoliva, L., and Poggi, G. (2019, January 28\u201330). Do gans leave artificial fingerprints?. Proceedings of the 2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), San Jose, CA, USA.","DOI":"10.1109\/MIPR.2019.00103"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Li, J., Xie, H., Li, J., Wang, Z., and Zhang, Y. (2021, January 20\u201325). Frequency-aware discriminative feature learning supervised by single-center loss for face forgery detection. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.00639"},{"key":"ref_28","unstructured":"Li, Z., Tang, W., Gao, S., Wang, S., and Wang, Y. (2024). Multiple Contexts and Frequencies Aggregation Network forDeepfake Detection. arXiv."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Qian, Y., Yin, G., Sheng, L., Chen, Z., and Shao, J. (2020, January 23\u201328). Thinking in frequency: Face forgery detection by mining frequency-aware clues. Proceedings of the European Conference on Computer Vision, Glasgow, UK.","DOI":"10.1007\/978-3-030-58610-2_6"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Zhou, P., Han, X., Morariu, V.I., and Davis, L.S. (2018, January 18\u201323). Learning rich features for image manipulation detection. Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00116"},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"103087","DOI":"10.1016\/j.inffus.2025.103087","article-title":"D2Fusion: Dual-domain fusion with feature superposition for Deepfake detection","volume":"120","author":"Qiu","year":"2025","journal-title":"Inf. Fusion"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Li, L., Bao, J., Zhang, T., Yang, H., Chen, D., Wen, F., and Guo, B. (2020, January 13\u201319). Face x-ray for more general face forgery detection. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00505"},{"key":"ref_33","unstructured":"Sun, K., Yao, T., Chen, S., Ding, S., Li, J., and Ji, R. (March, January 22). Dual contrastive learning for general face forgery detection. Proceedings of the AAAI Conference on Artificial Intelligence, Online Meeting."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Zhao, T., Xu, X., Xu, M., Ding, H., Xiong, Y., and Xia, W. (2021, January 11\u201317). Learning self-consistency for deepfake detection. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCV48922.2021.01475"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Shiohara, K., and Yamasaki, T. (2022, January 18\u201324). Detecting deepfakes with self-blended images. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.01816"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Zhou, T., Wang, W., Liang, Z., and Shen, J. (2021, January 20\u201325). Face forensics in the wild. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.00572"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Li, J., Xie, H., Yu, L., and Zhang, Y. (2022, January 10\u201314). Wavelet-enhanced weakly supervised local feature learning for face forgery detection. Proceedings of the 30th ACM International Conference on Multimedia, Lisbon, Portugal.","DOI":"10.1145\/3503161.3547832"},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Bao, J., Chen, D., Zeng, M., and Wen, F. (2021, January 11\u201317). Exploring temporal coherence for more general video face forgery detection. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCV48922.2021.01477"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"2425","DOI":"10.1109\/TIFS.2022.3186803","article-title":"LiSiam: Localization Invariance Siamese Network for Deepfake Detection","volume":"17","author":"Wang","year":"2022","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Haliassos, A., Vougioukas, K., Petridis, S., and Pantic, M. (2021, January 20\u201325). Lips don\u2019t lie: A generalisable and robust approach to face forgery detection. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.00500"},{"key":"ref_41","unstructured":"Guan, J., Zhou, H., Hong, Z., Ding, E., Wang, J., Quan, C., and Zhao, Y. (December, January 28). Delving into sequential patches for deepfake detection. Proceedings of the NIPS\u201922: 36th International Conference on Neural Information Processing Systems, New Orleans, LA, USA."},{"key":"ref_42","unstructured":"Shuai, C., Zhong, J., Wu, S., Lin, F., Wang, Z., Ba, Z., Liu, Z., Cavallaro, L., and Ren, K. (November, January 29). Locate and verify: A two-stream network for improved deepfake detection. Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada."}],"container-title":["Journal of Imaging"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2313-433X\/11\/9\/312\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T18:44:41Z","timestamp":1760035481000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2313-433X\/11\/9\/312"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,12]]},"references-count":42,"journal-issue":{"issue":"9","published-online":{"date-parts":[[2025,9]]}},"alternative-id":["jimaging11090312"],"URL":"https:\/\/doi.org\/10.3390\/jimaging11090312","relation":{},"ISSN":["2313-433X"],"issn-type":[{"type":"electronic","value":"2313-433X"}],"subject":[],"published":{"date-parts":[[2025,9,12]]}}}