{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T17:58:40Z","timestamp":1775066320464,"version":"3.50.1"},"reference-count":62,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2023,8,18]],"date-time":"2023-08-18T00:00:00Z","timestamp":1692316800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,8,18]],"date-time":"2023-08-18T00:00:00Z","timestamp":1692316800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Image Video Proc."],"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>\n                    Advancements in facial manipulation technology have resulted in highly realistic and indistinguishable face and expression swap videos. However, this has also raised concerns regarding the security risks associated with deepfakes. In the field of multimedia forensics, the detection and precise localization of image forgery has become essential tasks. Current deepfake detectors perform well with high-quality faces within specific datasets, but often struggle to maintain their performance when evaluated across different datasets. To this end, we propose an attention-based multi-task approach to improve feature maps for classification and localization tasks. The encoder and the attention-based decoder of our network generate localized maps that highlight regions with information about the type of manipulation. These localized features are shared with the classification network, improving its performance. Instead of using encoded spatial features, attention-based localized features from the decoder\u2019s first layer are combined with frequency domain features to create a discriminative representation for deepfake detection. Through extensive experiments on face and expression swap datasets, we demonstrate that our method achieves competitive performance in comparison to state-of-the-art deepfake detection approaches in both in-dataset and cross-dataset scenarios. Code is available at\n                    <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" ext-link-type=\"uri\" xlink:href=\"https:\/\/github.com\/saimawaseem\/Multi-Attention-Based-Approach-for-Deepfake-Face-and-Expression-Swap-Detection-and-Localization\">https:\/\/github.com\/saimawaseem\/Multi-Attention-Based-Approach-for-Deepfake-Face-and-Expression-Swap-Detection-and-Localization<\/jats:ext-link>\n                    .\n                  <\/jats:p>","DOI":"10.1186\/s13640-023-00614-z","type":"journal-article","created":{"date-parts":[[2023,8,18]],"date-time":"2023-08-18T10:00:34Z","timestamp":1692352834000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":32,"title":["Multi-attention-based approach for deepfake face and expression swap detection and localization"],"prefix":"10.1186","volume":"2023","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7704-7059","authenticated-orcid":false,"given":"Saima","family":"Waseem","sequence":"first","affiliation":[]},{"given":"Syed Abdul Rahman Syed","family":"Abu-Bakar","sequence":"additional","affiliation":[]},{"given":"Zaid","family":"Omar","sequence":"additional","affiliation":[]},{"given":"Bilal Ashfaq","family":"Ahmed","sequence":"additional","affiliation":[]},{"given":"Saba","family":"Baloch","sequence":"additional","affiliation":[]},{"given":"Adel","family":"Hafeezallah","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,18]]},"reference":[{"key":"614_CR1","doi-asserted-by":"crossref","unstructured":"J. Thies, M. Zollhofer, M. Stamminger, C. Theobalt, M. Nie\u00dfner, Face2face: Real-time face capture and reenactment of rgb videos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2387\u20132395 (2016)","DOI":"10.1109\/CVPR.2016.262"},{"key":"614_CR2","doi-asserted-by":"crossref","unstructured":"Y. Lu, Y.-W. Tai, C.-K. Tang, Attribute-guided face generation using conditional cyclegan. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 282\u2013297 (2018)","DOI":"10.1007\/978-3-030-01258-8_18"},{"key":"614_CR3","doi-asserted-by":"publisher","DOI":"10.1145\/3197517.3201283","author":"H Kim","year":"2018","unstructured":"H. Kim, P. Garrido, A. Tewari, W. Xu, J. Thies, M. Niessner, P. P\u00e9rez, C. Richardt, M. Zollh\u00f6fer, C. Theobalt, Deep video portraits. ACM Trans. Graph. (2018). https:\/\/doi.org\/10.1145\/3197517.3201283","journal-title":"ACM Trans. Graph."},{"key":"614_CR4","unstructured":"L. Li, J. Bao, H. Yang, D. Chen, F. Wen, FaceShifter: towards high fidelity and occlusion aware face swapping (2020). arXiv:1912.13457"},{"key":"614_CR5","unstructured":"S. Lu, FaceSwap-GAN. https:\/\/github.com\/shaoanlu\/faceswap-GAN. Accessed: 2022-01-30"},{"issue":"5","key":"614_CR6","doi-asserted-by":"publisher","first-page":"497","DOI":"10.1080\/15295036.2020.1832697","volume":"37","author":"C Gosse","year":"2020","unstructured":"C. Gosse, J. Burkell, Politics and porn: how news media characterizes problems presented by deepfakes. Crit. Stud. Media Commun. 37(5), 497\u2013511 (2020). https:\/\/doi.org\/10.1080\/15295036.2020.1832697","journal-title":"Crit. Stud. Media Commun."},{"key":"614_CR7","doi-asserted-by":"crossref","unstructured":"M. Westerlund, The emergence of deepfake technology: a review. Technol. Innovat. Manag. Rev. 9(11) (2019)","DOI":"10.22215\/timreview\/1282"},{"key":"614_CR8","doi-asserted-by":"publisher","unstructured":"Y. Li, X. Yang, P. Sun, H. Qi, S. Lyu, Celeb-df: a large-scale challenging dataset for deepfake forensics. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3204\u20133213 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00327","DOI":"10.1109\/CVPR42600.2020.00327"},{"key":"614_CR9","doi-asserted-by":"publisher","first-page":"4234","DOI":"10.1109\/TIFS.2021.3102487","volume":"16","author":"J Yang","year":"2021","unstructured":"J. Yang, A. Li, S. Xiao, W. Lu, X. Gao, Mtd-net: learning to detect deepfakes images by multi-scale texture difference. IEEE Trans. Inf. Forensics Secur. 16, 4234\u20134245 (2021). https:\/\/doi.org\/10.1109\/TIFS.2021.3102487","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"614_CR10","doi-asserted-by":"publisher","unstructured":"H.H. Nguyen, J. Yamagishi, I. Echizen, Capsule-forensics: Using capsule networks to detect forged images and videos. In: ICASSP 2019\u20132019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2307\u20132311 (2019). https:\/\/doi.org\/10.1109\/ICASSP.2019.8682602","DOI":"10.1109\/ICASSP.2019.8682602"},{"key":"614_CR11","doi-asserted-by":"publisher","first-page":"667","DOI":"10.1007\/978-3-030-58571-6_39","volume-title":"Computer Vision\u2014ECCV 2020","author":"I Masi","year":"2020","unstructured":"I. Masi, A. Killekar, R.M. Mascarenhas, S.P. Gurudatt, W. AbdAlmageed, Two-branch recurrent network for isolating deepfakes in videos, in Computer Vision\u2014ECCV 2020. ed. by A. Vedaldi, H. Bischof, T. Brox, J.-M. Frahm (Springer, Cham, 2020), pp.667\u2013684"},{"key":"614_CR12","doi-asserted-by":"publisher","unstructured":"D. Afchar, V. Nozick, J. Yamagishi, I. Echizen, Mesonet: a compact facial video forgery detection network. In: 2018 IEEE International Workshop on Information Forensics and Security (WIFS), pp. 1\u20137 (2018). https:\/\/doi.org\/10.1109\/WIFS.2018.8630761","DOI":"10.1109\/WIFS.2018.8630761"},{"key":"614_CR13","doi-asserted-by":"publisher","unstructured":"R. Durall, M. Keuper, J. Keuper, Watch your up-convolution: Cnn based generative deep neural networks are failing to reproduce spectral distributions. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7887\u20137896 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00791","DOI":"10.1109\/CVPR42600.2020.00791"},{"key":"614_CR14","doi-asserted-by":"publisher","unstructured":"X. Wu, Z. Xie, Y. Gao, Y. Xiao, Sstnet: detecting manipulated faces through spatial, steganalysis and temporal features. In: ICASSP 2020\u20132020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2952\u20132956 (2020). https:\/\/doi.org\/10.1109\/ICASSP40776.2020.9053969","DOI":"10.1109\/ICASSP40776.2020.9053969"},{"key":"614_CR15","doi-asserted-by":"publisher","unstructured":"D. G\u00fcera, E.J. Delp, Deepfake video detection using recurrent neural networks. In: 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pp. 1\u20136 (2018). https:\/\/doi.org\/10.1109\/AVSS.2018.8639163","DOI":"10.1109\/AVSS.2018.8639163"},{"key":"614_CR16","doi-asserted-by":"crossref","first-page":"86","DOI":"10.1007\/978-3-030-58610-2_6","volume-title":"Computer Vision\u2014ECCV 2020","author":"Y Qian","year":"2020","unstructured":"Y. Qian, G. Yin, L. Sheng, Z. Chen, J. Shao, Thinking in frequency: face forgery detection by mining frequency-aware clues, in Computer Vision\u2014ECCV 2020. ed. by A. Vedaldi, H. Bischof, T. Brox, J.-M. Frahm (Springer, Cham, 2020), pp.86\u2013103"},{"key":"614_CR17","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.119361","volume":"215","author":"Z Guo","year":"2023","unstructured":"Z. Guo, G. Yang, D. Zhang, M. Xia, Rethinking gradient operator for exposing ai-enabled face forgeries. Expert Syst. Appl. 215, 119361 (2023). https:\/\/doi.org\/10.1016\/j.eswa.2022.119361","journal-title":"Expert Syst. Appl."},{"key":"614_CR18","doi-asserted-by":"publisher","unstructured":"F. Matern, C. Riess, M. Stamminger, Exploiting visual artifacts to expose deepfakes and face manipulations. In: 2019 IEEE Winter Applications of Computer Vision Workshops (WACVW), pp. 83\u201392 (2019). https:\/\/doi.org\/10.1109\/WACVW.2019.00020","DOI":"10.1109\/WACVW.2019.00020"},{"key":"614_CR19","doi-asserted-by":"publisher","unstructured":"X. Yang, Y. Li, S. Lyu, Exposing deep fakes using inconsistent head poses. In: ICASSP 2019\u20142019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 8261\u20138265 (2019). https:\/\/doi.org\/10.1109\/ICASSP.2019.8683164","DOI":"10.1109\/ICASSP.2019.8683164"},{"key":"614_CR20","doi-asserted-by":"crossref","unstructured":"B. Xu, J. Liu, J. Liang, W. Lu, Y. Zhang, Deepfake videos detection based on texture features. Comput. Mater. Continua 68(1), (2021)","DOI":"10.32604\/cmc.2021.016760"},{"key":"614_CR21","doi-asserted-by":"crossref","unstructured":"F. Lugstein, S. Baier, G. Bachinger, A. Uhl, Prnu-based deepfake detection. In: Proceedings of the 2021 ACM Workshop on Information Hiding and Multimedia Security, pp. 7\u201312 (2021)","DOI":"10.1145\/3437880.3460400"},{"key":"614_CR22","doi-asserted-by":"publisher","unstructured":"Y. Zhu, Q. Li, J. Wang, C. Xu, Z. Sun, One shot face swapping on megapixels. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4832\u20134842 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00480","DOI":"10.1109\/CVPR46437.2021.00480"},{"key":"614_CR23","doi-asserted-by":"publisher","first-page":"83452","DOI":"10.1109\/ACCESS.2022.3196668","volume":"10","author":"A Groshev","year":"2022","unstructured":"A. Groshev, A. Maltseva, D. Chesakov, A. Kuznetsov, D. Dimitrov, Ghost-a new face swap approach for image and video domains. IEEE Access 10, 83452\u201383462 (2022). https:\/\/doi.org\/10.1109\/ACCESS.2022.3196668","journal-title":"IEEE Access"},{"key":"614_CR24","doi-asserted-by":"publisher","unstructured":"H. Zhao, T. Wei, W. Zhou, W. Zhang, D. Chen, N. Yu, Multi-attentional deepfake detection. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2185\u20132194 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00222","DOI":"10.1109\/CVPR46437.2021.00222"},{"key":"614_CR25","doi-asserted-by":"publisher","unstructured":"J. Li, H. Xie, L. Yu, X. Gao, Y. Zhang, Discriminative feature mining based on frequency information and metric learning for face forgery detection. IEEE Transactions on Knowledge and Data Engineering, 1\u20131 (2021). https:\/\/doi.org\/10.1109\/TKDE.2021.3117003","DOI":"10.1109\/TKDE.2021.3117003"},{"issue":"11","key":"614_CR26","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, Y. Bengio, Generative adversarial networks. Commun. ACM 63(11), 139\u2013144 (2020). https:\/\/doi.org\/10.1145\/3422622","journal-title":"Commun. ACM"},{"key":"614_CR27","unstructured":"J. Frank, T. Eisenhofer, L. Sch\u00f6nherr, A. Fischer, D. Kolossa, T. Holz, Leveraging frequency analysis for deep fake image recognition. In: International Conference on Machine Learning, pp. 3247\u20133258 (2020). PMLR"},{"key":"614_CR28","doi-asserted-by":"crossref","unstructured":"A. Rossler, D. Cozzolino, L. Verdoliva, C. Riess, J. Thies, M. Nie\u00dfner, Faceforensics++: learning to detect manipulated facial images. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1\u201311 (2019)","DOI":"10.1109\/ICCV.2019.00009"},{"key":"614_CR29","unstructured":"B. Dolhansky, R. Howes, B. Pflaum, N. Baram, C.C. Ferrer, The deepfake detection challenge (dfdc) preview dataset. arXiv preprint arXiv:1910.08854 (2019)"},{"key":"614_CR30","doi-asserted-by":"publisher","unstructured":"P. Zhou, X. Han, V.I. Morariu, L.S. Davis, Two-stream neural networks for tampered face detection. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1831\u20131839 (2017). https:\/\/doi.org\/10.1109\/CVPRW.2017.229","DOI":"10.1109\/CVPRW.2017.229"},{"key":"614_CR31","doi-asserted-by":"publisher","first-page":"83144","DOI":"10.1109\/ACCESS.2020.2988660","volume":"8","author":"T Jung","year":"2020","unstructured":"T. Jung, S. Kim, K. Kim, Deepvision: Deepfakes detection using human eye blinking pattern. IEEE Access 8, 83144\u201383154 (2020). https:\/\/doi.org\/10.1109\/ACCESS.2020.2988660","journal-title":"IEEE Access"},{"key":"614_CR32","doi-asserted-by":"publisher","unstructured":"J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, L. Fei-Fei, Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 (2009). https:\/\/doi.org\/10.1109\/CVPR.2009.5206848","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"614_CR33","unstructured":"M. Tan, Q. Le, EfficientNet: Rethinking model scaling for convolutional neural networks. In: Chaudhuri, K., Salakhutdinov, R. (eds.) Proceedings of the 36th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 97, pp. 6105\u20136114. PMLR, (2019). https:\/\/proceedings.mlr.press\/v97\/tan19a.html"},{"key":"614_CR34","doi-asserted-by":"publisher","unstructured":"S. Waseem, S.R. Abu-Bakar, Z. Omar, B.A. Ahmed, S. Baloch, A multi-color spatio-temporal approach for detecting deepfake. In: 2022 12th International Conference on Pattern Recognition Systems (ICPRS), pp. 1\u20135 (2022). https:\/\/doi.org\/10.1109\/ICPRS54038.2022.9853853","DOI":"10.1109\/ICPRS54038.2022.9853853"},{"issue":"1","key":"614_CR35","first-page":"80","volume":"3","author":"E Sabir","year":"2019","unstructured":"E. Sabir, J. Cheng, A. Jaiswal, W. AbdAlmageed, I. Masi, P. Natarajan, Recurrent convolutional strategies for face manipulation detection in videos. Interfaces (GUI) 3(1), 80\u201387 (2019)","journal-title":"Interfaces (GUI)"},{"key":"614_CR36","unstructured":"R.D. Lopez, M. Keuper, F.-J. Pfreundt, J. Keuper, Unmasking DeepFakes with simple Features (2019)"},{"key":"614_CR37","doi-asserted-by":"publisher","unstructured":"L. Tian, H. Yao, M. Li, Fakepoi: A large-scale fake person of interest video detection benchmark and a strong baseline. IEEE Transactions on Circuits and Systems for Video Technology, 1\u20131 (2023). https:\/\/doi.org\/10.1109\/TCSVT.2023.3269742","DOI":"10.1109\/TCSVT.2023.3269742"},{"key":"614_CR38","doi-asserted-by":"crossref","unstructured":"K. Sun, H. Liu, Q. Ye, Y. Gao, J. Liu, L. Shao, R. Ji, Domain general face forgery detection by learning to weight. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 2638\u20132646 (2021)","DOI":"10.1609\/aaai.v35i3.16367"},{"key":"614_CR39","unstructured":"D. Wodajo, S. Atnafu, Deepfake Video Detection Using Convolutional Vision Transformer (2021). arXiv:2102.11126"},{"key":"614_CR40","doi-asserted-by":"publisher","first-page":"1668","DOI":"10.1109\/TIP.2023.3246793","volume":"32","author":"Y Hua","year":"2023","unstructured":"Y. Hua, R. Shi, P. Wang, S. Ge, Learning patch-channel correspondence for interpretable face forgery detection. IEEE Trans. Image Process. 32, 1668\u20131680 (2023). https:\/\/doi.org\/10.1109\/TIP.2023.3246793","journal-title":"IEEE Trans. Image Process."},{"key":"614_CR41","doi-asserted-by":"publisher","unstructured":"H.H. Nguyen, F. Fang, J. Yamagishi, I. Echizen, Multi-task learning for detecting and segmenting manipulated facial images and videos. In: 2019 IEEE 10th International Conference on Biometrics Theory, Applications and Systems (BTAS), pp. 1\u20138 (2019). https:\/\/doi.org\/10.1109\/BTAS46853.2019.9185974","DOI":"10.1109\/BTAS46853.2019.9185974"},{"key":"614_CR42","doi-asserted-by":"crossref","unstructured":"L. Li, J. Bao, T. Zhang, H. Yang, D. Chen, F. Wen, B. Guo, Face x-ray for more general face forgery detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5001\u20135010 (2020)","DOI":"10.1109\/CVPR42600.2020.00505"},{"key":"614_CR43","unstructured":"P. Liu, Y. Lin, Y. He, Y. Wei, L. Zhen, J.T. Zhou, R.S.M. Goh, J. Liu, Automated deepfake detection. arXiv preprint arXiv:2106.10705 (2021)"},{"key":"614_CR44","doi-asserted-by":"crossref","unstructured":"H. Dang, F. Liu, J. Stehouwer, X. Liu, A.K. Jain, On the detection of digital face manipulation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5781\u20135790 (2020)","DOI":"10.1109\/CVPR42600.2020.00582"},{"key":"614_CR45","doi-asserted-by":"crossref","first-page":"191","DOI":"10.1007\/978-3-030-58610-2_12","volume-title":"Computer Vision\u2014ECCV 2020","author":"B Niu","year":"2020","unstructured":"B. Niu, W. Wen, W. Ren, X. Zhang, L. Yang, S. Wang, K. Zhang, X. Cao, H. Shen, Single image super-resolution via a holistic attention network, in Computer Vision\u2014ECCV 2020. ed. by A. Vedaldi, H. Bischof, T. Brox, J.-M. Frahm (Springer, Cham, 2020), pp.191\u2013207"},{"key":"614_CR46","doi-asserted-by":"crossref","unstructured":"S. Woo, J. Park, J.-Y. Lee, I.S. Kweon, Cbam: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"614_CR47","unstructured":"T. Park, J.-Y. Zhu, O. Wang, J. Lu, E. Shechtman, A. Efros, R. Zhang, Swapping autoencoder for deep image manipulation. In: H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, H. Lin, (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 7198\u20137211. Curran Associates, Inc., (2020). https:\/\/proceedings.neurips.cc\/paper\/2020\/file\/50905d7b2216bfeccb5b41016357176b-Paper.pdf"},{"issue":"10","key":"614_CR48","doi-asserted-by":"publisher","first-page":"3","DOI":"10.23915\/distill.00003","volume":"1","author":"A Odena","year":"2016","unstructured":"A. Odena, V. Dumoulin, C. Olah, Deconvolution and checkerboard artifacts. Distill 1(10), 3 (2016)","journal-title":"Distill"},{"key":"614_CR49","doi-asserted-by":"publisher","unstructured":"T.-Y. Lin, A. RoyChowdhury, S. Maji, Bilinear cnn models for fine-grained visual recognition. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1449\u20131457 (2015). https:\/\/doi.org\/10.1109\/ICCV.2015.170","DOI":"10.1109\/ICCV.2015.170"},{"issue":"10","key":"614_CR50","doi-asserted-by":"publisher","first-page":"1499","DOI":"10.1109\/LSP.2016.2603342","volume":"23","author":"K Zhang","year":"2016","unstructured":"K. Zhang, Z. Zhang, Z. Li, Y. Qiao, Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 23(10), 1499\u20131503 (2016). https:\/\/doi.org\/10.1109\/LSP.2016.2603342","journal-title":"IEEE Signal Process. Lett."},{"key":"614_CR51","doi-asserted-by":"publisher","unstructured":"D. Cozzolino, G. Poggi, L. Verdoliva, Recasting residual-based local descriptors as convolutional neural networks: an application to image forgery detection. In: Proceedings of the 5th ACM Workshop on Information Hiding and Multimedia Security, pp. 159\u2013164. Association for Computing Machinery, New York, NY, USA (2017). https:\/\/doi.org\/10.1145\/3082031.3083247","DOI":"10.1145\/3082031.3083247"},{"key":"614_CR52","unstructured":"D.P. Kingma, J. Ba, Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"614_CR53","doi-asserted-by":"publisher","unstructured":"H. Liu, X. Li, W. Zhou, Y. Chen, Y. He, H. Xue, W. Zhang, N. Yu, Spatial-phase shallow learning: Rethinking face forgery detection in frequency domain. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 772\u2013781 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00083","DOI":"10.1109\/CVPR46437.2021.00083"},{"issue":"3","key":"614_CR54","doi-asserted-by":"publisher","first-page":"868","DOI":"10.1109\/TIFS.2012.2190402","volume":"7","author":"J Fridrich","year":"2012","unstructured":"J. Fridrich, J. Kodovsky, Rich models for steganalysis of digital images. IEEE Trans. Inf. Forensics Secur. 7(3), 868\u2013882 (2012). https:\/\/doi.org\/10.1109\/TIFS.2012.2190402","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"614_CR55","doi-asserted-by":"publisher","unstructured":"J.-Y. Zhu, T. Park, P. Isola, A.A. Efros, Unpaired image-to-image translation using cycle-consistent adversarial networks. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2242\u20132251 (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.244","DOI":"10.1109\/ICCV.2017.244"},{"key":"614_CR56","unstructured":"D.S. Nigeria, DeepFake-Detection Dataset. https:\/\/github.com\/DataScienceNigeria\/Fake-Detection-dataset-for-deepfake-from-Google-and-Jigsaw. Accessed: 2022-04-30"},{"key":"614_CR57","unstructured":"Y. Li, S. Lyu, Exposing deepfake videos by detecting face warping artifacts. arXiv preprint arXiv:1811.00656 (2018)"},{"key":"614_CR58","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2022.104673","volume":"110","author":"R Tolosana","year":"2022","unstructured":"R. Tolosana, S. Romero-Tapiador, R. Vera-Rodriguez, E. Gonzalez-Sosa, J. Fierrez, Deepfakes detection across generations: analysis of facial regions, fusion, and performance evaluation. Eng. Appl. Artif. Intell. 110, 104673 (2022). https:\/\/doi.org\/10.1016\/j.engappai.2022.104673","journal-title":"Eng. Appl. Artif. Intell."},{"key":"614_CR59","doi-asserted-by":"publisher","unstructured":"B. Bayar, M.C. Stamm, A deep learning approach to universal image manipulation detection using a new convolutional layer, pp. 5\u201310. Association for Computing Machinery, New York, NY, USA (2016). https:\/\/doi.org\/10.1145\/2909827.2930786","DOI":"10.1145\/2909827.2930786"},{"key":"614_CR60","doi-asserted-by":"publisher","DOI":"10.3390\/fi13040093","author":"SS Khalil","year":"2021","unstructured":"S.S. Khalil, S.M. Youssef, S.N. Saleh, icaps-dfake: an integrated capsule-based model for deepfake image and video detection. Future Internet (2021). https:\/\/doi.org\/10.3390\/fi13040093","journal-title":"Future Internet"},{"key":"614_CR61","doi-asserted-by":"crossref","unstructured":"C. Fosco, E. Josephs, A. Andonian, A. Lee, X. Wang, A. Oliva, Deepfake caricatures: amplifying attention to artifacts increases deepfake detection by humans and machines. arXiv preprint arXiv:2206.00535 (2022)","DOI":"10.1167\/jov.22.14.4079"},{"key":"614_CR62","doi-asserted-by":"publisher","unstructured":"M. Du, S. Pentyala, Y. Li, X. Hu, Towards generalizable deepfake detection with locality-aware autoencoder. In: Proceedings of the 29th ACM International Conference on Information and Knowledge Management. CIKM \u201920, pp. 325\u2013334. Association for Computing Machinery, New York, NY, USA (2020). https:\/\/doi.org\/10.1145\/3340531.3411892","DOI":"10.1145\/3340531.3411892"}],"container-title":["EURASIP Journal on Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s13640-023-00614-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1186\/s13640-023-00614-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s13640-023-00614-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T06:35:03Z","timestamp":1729924503000},"score":1,"resource":{"primary":{"URL":"https:\/\/jivp-eurasipjournals.springeropen.com\/articles\/10.1186\/s13640-023-00614-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,18]]},"references-count":62,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2023,12]]}},"alternative-id":["614"],"URL":"https:\/\/doi.org\/10.1186\/s13640-023-00614-z","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-2198055\/v1","asserted-by":"object"}]},"ISSN":["1687-5281"],"issn-type":[{"value":"1687-5281","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,8,18]]},"assertion":[{"value":"26 October 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 August 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 August 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"14"}}