{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T12:14:45Z","timestamp":1775132085897,"version":"3.50.1"},"reference-count":50,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T00:00:00Z","timestamp":1770076800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T00:00:00Z","timestamp":1770076800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"National Natural Science Foundation of China Youth Foud","award":["61902165"],"award-info":[{"award-number":["61902165"]}]},{"name":"Provincial Department of Education","award":["JYTMS20231040"],"award-info":[{"award-number":["JYTMS20231040"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s00530-025-02173-x","type":"journal-article","created":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T03:41:43Z","timestamp":1770090103000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Fast latent-feature augmentation for cross-domain face forgery detection"],"prefix":"10.1007","volume":"32","author":[{"given":"Fang","family":"Sun","sequence":"first","affiliation":[]},{"given":"Pan","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Xiaoxuan","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Wenjun","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jing","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,3]]},"reference":[{"issue":"1","key":"2173_CR1","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1109\/TII.2021.3085669","volume":"18","author":"J Li","year":"2022","unstructured":"Li, J., Chen, J., Sheng, B., Li, P., Yang, P., David, D.F., Jun, Q.: Automatic detection and classification system of domestic waste via multimodel cascaded convolutional neural network. IEEE Trans. Industr. Inf. 18(1), 163\u2013173 (2022)","journal-title":"IEEE Trans. Industr. Inf."},{"key":"2173_CR2","doi-asserted-by":"publisher","first-page":"2226","DOI":"10.1109\/TMM.2022.3144890","volume":"25","author":"N Jiang","year":"2023","unstructured":"Jiang, N., Sheng, B., Li, P., Lee, T.-Y.: Photohelper: portrait photographing guidance via deep feature retrieval and fusion. IEEE Trans. Multimedia 25, 2226\u20132238 (2023)","journal-title":"IEEE Trans. Multimedia"},{"issue":"11","key":"2173_CR3","doi-asserted-by":"crossref","first-page":"13489","DOI":"10.1109\/TPAMI.2023.3289667","volume":"45","author":"Z Chen","year":"2023","unstructured":"Chen, Z., Qiu, G., Li, P., Zhu, L., Yang, X., Sheng, B.: Mngnas: distilling adaptive combination of multiple searched networks for one-shot neural architecture search. IEEE Trans. Pattern Anal. Mach. Intell. 45(11), 13489\u201313508 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3\/4","key":"2173_CR4","doi-asserted-by":"publisher","DOI":"10.1002\/cav.2081","volume":"33","author":"CC Wong","year":"2022","unstructured":"Wong, C.C., Cai, C.E., Tsai, H.M., Liu, G.T., Wong, S.K.: Generation of cart-pulling animation in a multiagent environment using deep learning. Comput. Anim. Virt. Worlds 33(3\/4), e2081 (2022)","journal-title":"Comput. Anim. Virt. Worlds"},{"issue":"3\/4","key":"2173_CR5","doi-asserted-by":"publisher","DOI":"10.1002\/cav.2063","volume":"33","author":"Z Chen","year":"2022","unstructured":"Chen, Z., Lyu, D.: Procedural generation of virtual pavilions via a deep convolutional generative adversarial network. Comput. Anim. Virt. Worlds 33(3\/4), e2063 (2022)","journal-title":"Comput. Anim. Virt. Worlds"},{"issue":"6","key":"2173_CR6","doi-asserted-by":"publisher","DOI":"10.1007\/s11432-021-3383-y","volume":"65","author":"Y Wang","year":"2022","unstructured":"Wang, Y., Peng, J., Wang, H., Wang, M.: Progressive learning with multi-scale attention network for cross-domain vehicle re-identification. Sci. China Inf. Sci. 65(6), 106163 (2022)","journal-title":"Sci. China Inf. Sci."},{"key":"2173_CR7","unstructured":"Chow, F., Ng, C.W., Yip, C.H., Lau, H.C.: Face recognition method and system for personal identification and authentication (2008)"},{"key":"2173_CR8","unstructured":"Zuojun, F., Gao, J., Jiang, W., Liang, G.: Method for recognizing face from monitoring video data. Google Patents. US Patent 11,263,435 (2022)"},{"key":"2173_CR9","doi-asserted-by":"crossref","unstructured":"Kaur, P., Kewal, K., Sharma, S.K., Tanuj, K.: Facial-recognition algorithms: a literature review 60(2), 131\u2013139 (2020)","DOI":"10.1177\/0025802419893168"},{"issue":"3\u20134","key":"2173_CR10","doi-asserted-by":"publisher","DOI":"10.1002\/cav.2160","volume":"34","author":"A Che","year":"2023","unstructured":"Che, A., Yang, J., Guo, C., Dai, H.-N., Xie, H., Li, P.: Aegan: generating imperceptible face synthesis via autoencoder-based generative adversarial network. Comput. Anim. Virt. Worlds 34(3\u20134), e2160 (2023)","journal-title":"Comput. Anim. Virt. Worlds"},{"key":"2173_CR11","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 4396\u20134405 (2018)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"2173_CR12","doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M.-J., Kim, M.S., Ha, J.-W., Kim, S., Choo, J.: Stargan: unified generative adversarial networks for multi-domain image-to-image translation. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 8789\u20138797 (2017)","DOI":"10.1109\/CVPR.2018.00916"},{"key":"2173_CR13","doi-asserted-by":"crossref","unstructured":"Thies, J., Zollh\u00f6fer, M., Stamminger, M., Theobalt, C., Nie\u00dfner, M.: Face2face: real-time face capture and reenactment of rgb videos. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2387\u20132395 (2016)","DOI":"10.1109\/CVPR.2016.262"},{"key":"2173_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3323035","volume":"38","author":"J Thies","year":"2019","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Deferred neural rendering. ACM Trans. Graph. (TOG) 38, 1\u201312 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"2173_CR15","doi-asserted-by":"crossref","unstructured":"Qian, Y., Yin, G., Sheng, L., Chen, Z., Shao, J.: Thinking in frequency: face forgery detection by mining frequency-aware clues. arXiv: abs\/2007.09355 (2020)","DOI":"10.1007\/978-3-030-58610-2_6"},{"key":"2173_CR16","doi-asserted-by":"crossref","unstructured":"Zhou, P., Han, X., Morariu, V.I., Davis, L.S.: Two-stream neural networks for tampered face detection. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 1831\u20131839 (2017)","DOI":"10.1109\/CVPRW.2017.229"},{"key":"2173_CR17","doi-asserted-by":"crossref","unstructured":"Cozzolino, D., Poggi, G., Verdoliva, L.: Recasting residual-based local descriptors as convolutional neural networks: an application to image forgery detection. In: Proceedings of the 5th ACM Workshop on Information Hiding and Multimedia Security (2017)","DOI":"10.1145\/3082031.3083247"},{"key":"2173_CR18","doi-asserted-by":"publisher","first-page":"2691","DOI":"10.1109\/TIFS.2018.2825953","volume":"13","author":"B Bayar","year":"2018","unstructured":"Bayar, B., Stamm, M.C.: Constrained convolutional neural networks: a new approach towards general purpose image manipulation detection. IEEE Trans. Inf. Forensics Secur. 13, 2691\u20132706 (2018)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"2173_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2022.103587","volume":"226","author":"Z Guo","year":"2022","unstructured":"Guo, Z., Yang, G., Wang, D., Zhang, D.: A data augmentation framework by mining structured features for fake face image detection. Comput. Vis. Image Underst. 226, 103587 (2022)","journal-title":"Comput. Vis. Image Underst."},{"key":"2173_CR20","doi-asserted-by":"crossref","unstructured":"Shiohara, K., Yamasaki, T.: Detecting deepfakes with self-blended images. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 18699\u201318708 (2022)","DOI":"10.1109\/CVPR52688.2022.01816"},{"key":"2173_CR21","unstructured":"Khormali, A., Yuan, J.: Self-supervised graph transformer for deepfake detection. arXiv: abs\/2307.15019 (2023)"},{"key":"2173_CR22","doi-asserted-by":"crossref","unstructured":"Masi, I., Killekar, A., Mascarenhas, R.M., Gurudatt, S.P., AbdAlmageed, W.: Two-branch recurrent network for isolating deepfakes in videos. arXiv: abs\/2008.03412 (2020)","DOI":"10.1007\/978-3-030-58571-6_39"},{"key":"2173_CR23","doi-asserted-by":"crossref","unstructured":"Zhuang, W., Chu, Q., Tan, Z., Liu, Q., Yuan, H., Miao, C., Luo, Z., Yu, N.: Uia-vit: Unsupervised inconsistency-aware method based on vision transformer for face forgery detection. arXiv: abs\/2210.12752 (2022)","DOI":"10.1007\/978-3-031-20065-6_23"},{"key":"2173_CR24","doi-asserted-by":"crossref","unstructured":"Nawaz, M., Javed, A., Irtaza, A.: Resnet-swish-dense54: a deep learning approach for deepfakes detection. Vis. Comput., 1\u201322 (2022)","DOI":"10.1007\/s00371-022-02732-7"},{"key":"2173_CR25","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: Deep learning with depthwise separable convolutions. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 1800\u20131807 (2016)","DOI":"10.1109\/CVPR.2017.195"},{"key":"2173_CR26","doi-asserted-by":"crossref","unstructured":"Afchar, D., Nozick, V., Yamagishi, J., Echizen, I.: Mesonet: a compact facial video forgery detection network. In: 2018 IEEE International Workshop on Information Forensics and Security (WIFS), 1\u20137 (2018)","DOI":"10.1109\/WIFS.2018.8630761"},{"key":"2173_CR27","unstructured":"Tan, M., Le, Q.V.: Efficientnet: rethinking model scaling for convolutional neural networks. arXiv: abs\/1905.11946 (2019)"},{"key":"2173_CR28","doi-asserted-by":"crossref","unstructured":"Guo, Y., Zhen, C., Yan, P.: Controllable guide-space for generalizable face forgery detection. arXiv: abs\/2307.14039 (2023)","DOI":"10.1109\/ICCV51070.2023.01903"},{"key":"2173_CR29","unstructured":"Frank, J.C., Eisenhofer, T., Sch\u00f6nherr, L., Fischer, A., Kolossa, D., Holz, T.: Leveraging frequency analysis for deep fake image recognition. arXiv: abs\/2003.08685 (2020)"},{"key":"2173_CR30","doi-asserted-by":"crossref","unstructured":"Wang, C., Deng, W.: Representative forgery mining for fake face detection. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 14918\u201314927 (2021)","DOI":"10.1109\/CVPR46437.2021.01468"},{"key":"2173_CR31","doi-asserted-by":"crossref","unstructured":"Sun, K., Liu, H., Ye, Q., Gao, Y., Liu, J., Shao, L., Ji, R.: Domain general face forgery detection by learning to weight. In: AAAI Conference on Artificial Intelligence (2021)","DOI":"10.1609\/aaai.v35i3.16367"},{"key":"2173_CR32","doi-asserted-by":"crossref","unstructured":"Zhao, H., Zhou, W., Chen, D., Wei, T., Zhang, W., Yu, N.: Multi-attentional deepfake detection. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2185\u20132194 (2021)","DOI":"10.1109\/CVPR46437.2021.00222"},{"key":"2173_CR33","doi-asserted-by":"publisher","DOI":"10.1016\/j.cose.2022.103024","volume":"125","author":"Y Cao","year":"2022","unstructured":"Cao, Y., Chen, J., Huang, L., Huang, T., Ye, F.: Three-classification face manipulation detection using attention-based feature decomposition. Comput. Secur. 125, 103024 (2022)","journal-title":"Comput. Secur."},{"key":"2173_CR34","doi-asserted-by":"crossref","unstructured":"Chen, S., Yao, T., Chen, Y., Ding, S., Li, J., Ji, R.: Local relation learning for face forgery detection. In: AAAI Conference on Artificial Intelligence (2021)","DOI":"10.1609\/aaai.v35i2.16193"},{"key":"2173_CR35","doi-asserted-by":"crossref","unstructured":"Cao, J., Ma, C., Yao, T., Chen, S., Ding, S., Yang, X.: End-to-end reconstruction-classification learning for face forgery detection. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 4103\u20134112 (2022)","DOI":"10.1109\/CVPR52688.2022.00408"},{"key":"2173_CR36","doi-asserted-by":"crossref","unstructured":"Ju, Y., Jia, S., Ke, L., Xue, H., Nagano, K., Lyu, S.: Fusing global and local features for generalized ai-synthesized image detection. In: 2022 IEEE International Conference on Image Processing (ICIP), 3465\u20133469 (2022)","DOI":"10.1109\/ICIP46576.2022.9897820"},{"key":"2173_CR37","doi-asserted-by":"crossref","unstructured":"Sun, K., Yao, T., Chen, S., Ding, S., Jilin, L., Ji, R.: Dual contrastive learning for general face forgery detection. In: AAAI Conference on Artificial Intelligence (2021)","DOI":"10.1609\/aaai.v36i2.20130"},{"key":"2173_CR38","doi-asserted-by":"crossref","unstructured":"Luo, Y., Zhang, Y., Yan, J., Liu, W.: Generalizing face forgery detection with high-frequency features. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 16312\u201316321 (2021)","DOI":"10.1109\/CVPR46437.2021.01605"},{"key":"2173_CR39","doi-asserted-by":"crossref","unstructured":"Li, L., Bao, J., Zhang, T., Yang, H., Chen, D., Wen, F., Guo, B.: Face x-ray for more general face forgery detection. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 5000\u20135009 (2019)","DOI":"10.1109\/CVPR42600.2020.00505"},{"key":"2173_CR40","doi-asserted-by":"crossref","unstructured":"Chen, L., Zhang, Y., Song, Y., Liu, L., Wang, J.: Self-supervised learning of adversarial example: Towards good generalizations for deepfake detection. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 18689\u201318698 (2022)","DOI":"10.1109\/CVPR52688.2022.01815"},{"key":"2173_CR41","unstructured":"Yu, C., Zhang, X., Duan, Y., Yan, S., Wang, Z., Xiang, Y., Ji, S., Chen, W.: Diff-id: An explainable identity difference quantification framework for deepfake detection. arXiv: abs\/2303.18174 (2023)"},{"key":"2173_CR42","doi-asserted-by":"crossref","unstructured":"Qiu, L., Jiang, K., Liu, S., Tan, X.: Multi-level distributional discrepancy enhancement for cross domain face forgery detection. In: Chinese Conference on Pattern Recognition and Computer Vision (CVPR) (2025)","DOI":"10.1007\/978-981-97-8499-8_35"},{"key":"2173_CR43","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2024.106909","volume":"182","author":"N Bai","year":"2024","unstructured":"Bai, N., Wang, X., Han, R., Hou, J., Wang, Q., Pang, S.: Towards generalizable face forgery detection via mitigating spurious correlation. Neural Netw. 182, 106909 (2024)","journal-title":"Neural Netw."},{"key":"2173_CR44","unstructured":"Qiu, L., Jiang, K., Tan, X.: Contrastive desensitization learning for cross domain face forgery detection. arXiv preprint arXiv:2505.20675 (2025)"},{"key":"2173_CR45","doi-asserted-by":"crossref","unstructured":"Wang, Y.: Survey on deep multi-modal data analytics: collaboration, rivalry and fusion. Multimedia computing, communications and applications (2020)","DOI":"10.1145\/3408317"},{"key":"2173_CR46","doi-asserted-by":"publisher","first-page":"1055","DOI":"10.1109\/TPAMI.2022.3155499","volume":"45","author":"M Yang","year":"2022","unstructured":"Yang, M., Li, Y., Hu, P., Bai, J., Lv, J., Peng, X.: Robust multi-view clustering with incomplete information. IEEE Trans. Pattern Anal. Mach. Intell. 45, 1055\u20131069 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2173_CR47","doi-asserted-by":"crossref","unstructured":"R\u00f6ssler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Nie\u00dfner, M.: Faceforensics++: Learning to detect manipulated facial images. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), 1\u201311 (2019)","DOI":"10.1109\/ICCV.2019.00009"},{"key":"2173_CR48","doi-asserted-by":"crossref","unstructured":"Li, Y., Yang, X., Sun, P., Qi, H., Lyu, S.: Celeb-df: A large-scale challenging dataset for deepfake forensics. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 3204\u20133213 (2019)","DOI":"10.1109\/CVPR42600.2020.00327"},{"key":"2173_CR49","unstructured":"Dolhansky, B., Howes, R., Pflaum, B., Baram, N., Canton-Ferrer, C.: The deepfake detection challenge (dfdc) preview dataset. arXiv: abs\/1910.08854 (2019)"},{"key":"2173_CR50","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1016\/j.imavis.2016.01.002","volume":"47","author":"C Sagonas","year":"2016","unstructured":"Sagonas, C., Antonakos, E., Tzimiropoulos, G., Zafeiriou, S., Pantic, M.: 300 faces in-the-wild challenge: database and results. Image Vis. Comput. 47, 3\u201318 (2016)","journal-title":"Image Vis. Comput."}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02173-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-02173-x","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02173-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T11:37:58Z","timestamp":1775129878000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-02173-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,3]]},"references-count":50,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["2173"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-02173-x","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-6426336\/v1","asserted-by":"object"}]},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,3]]},"assertion":[{"value":"11 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 February 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"100"}}