{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T19:21:49Z","timestamp":1743016909020,"version":"3.40.3"},"publisher-location":"Cham","reference-count":39,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031251146"},{"type":"electronic","value":"9783031251153"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25115-3_5","type":"book-chapter","created":{"date-parts":[[2023,1,28]],"date-time":"2023-01-28T13:57:23Z","timestamp":1674914243000},"page":"68-82","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Visual Explanations for\u00a0Exposing Potential Inconsistency of\u00a0Deepfakes"],"prefix":"10.1007","author":[{"given":"Pengfei","family":"Pei","sequence":"first","affiliation":[]},{"given":"Xianfeng","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Yun","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Chengqiao","family":"Hu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,1,29]]},"reference":[{"key":"5_CR1","doi-asserted-by":"publisher","unstructured":"Arnab, A., Dehghani, M., Heigold, G., Sun, C., Lu\u010di\u0107, M., Schmid, C.: ViViT: a video vision transformer. In: 2021 IEEE\/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, pp. 6816\u20136826 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00676","DOI":"10.1109\/ICCV48922.2021.00676"},{"key":"5_CR2","doi-asserted-by":"publisher","unstructured":"Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, pp. 1800\u20131807 (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.195","DOI":"10.1109\/CVPR.2017.195"},{"key":"5_CR3","doi-asserted-by":"publisher","unstructured":"Cozzolino, D., R\u00f6ssler, A., Thies, J., Nie\u00dfner, M., Verdoliva, L.: ID-reveal: identity-aware deepfake video detection. In: 2021 IEEE\/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, pp. 15088\u201315097 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.01483","DOI":"10.1109\/ICCV48922.2021.01483"},{"key":"5_CR4","doi-asserted-by":"publisher","unstructured":"Diao, Q., Jiang, Y., Wen, B., Sun, J., Yuan, Z.: MetaFormer: a unified meta framework for fine-grained recognition. CoRR abs\/2203.02751 (2022). https:\/\/doi.org\/10.48550\/arXiv.2203.02751","DOI":"10.48550\/arXiv.2203.02751"},{"key":"5_CR5","unstructured":"Dufour, N., Gully, A.: DeepFakeDetection dataset (2019). https:\/\/ai.googleblog.com\/2019\/09\/contributing-data-to-deepfake-detection.html"},{"issue":"3","key":"5_CR6","doi-asserted-by":"publisher","first-page":"868","DOI":"10.1109\/TIFS.2012.2190402","volume":"7","author":"JJ Fridrich","year":"2012","unstructured":"Fridrich, J.J., Kodovsk\u00fd, J.: Rich models for steganalysis of digital images. IEEE Trans. Inf. Forensics Secur. 7(3), 868\u2013882 (2012). https:\/\/doi.org\/10.1109\/TIFS.2012.2190402","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"5_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"168","DOI":"10.1007\/978-3-030-69449-4_13","volume-title":"Digital Forensics and Watermarking","author":"Y Gu","year":"2021","unstructured":"Gu, Y., Zhao, X., Gong, C., Yi, X.: Deepfake video detection using audio-visual consistency. In: Zhao, X., Shi, Y.-Q., Piva, A., Kim, H.J. (eds.) IWDW 2020. LNCS, vol. 12617, pp. 168\u2013180. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-69449-4_13"},{"key":"5_CR8","doi-asserted-by":"publisher","unstructured":"Gu, Z., et al.: Spatiotemporal inconsistency learning for deepfake video detection. In: Shen, H.T., et al. (eds.) MM 2021: ACM Multimedia Conference, pp. 3473\u20133481. ACM, Virtual Event, China (2021). https:\/\/doi.org\/10.1145\/3474085.3475508","DOI":"10.1145\/3474085.3475508"},{"key":"5_CR9","doi-asserted-by":"crossref","unstructured":"Gu, Z., Chen, Y., Yao, T., Ding, S., Li, J., Ma, L.: Delving into the local: dynamic inconsistency learning for deepfake video detection. In: Thirty-Sixth AAAI Conference on Artificial Intelligence, pp. 744\u2013752. AAAI Press, Virtual Event (2022)","DOI":"10.1609\/aaai.v36i1.19955"},{"key":"5_CR10","doi-asserted-by":"crossref","unstructured":"Guo, J., Han, K., Wu, H., Xu, C., Tang, Y., Xu, C., Wang, Y.: CMT: convolutional neural networks meet vision transformers. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, Louisiana (2022)","DOI":"10.1109\/CVPR52688.2022.01186"},{"key":"5_CR11","doi-asserted-by":"crossref","unstructured":"Haliassos, A., Vougioukas, K., Petridis, S., Pantic, M.: Lips don\u2019t lie: a generalisable and robust approach to face forgery detection. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, pp. 5039\u20135049. Virtual (2021)","DOI":"10.1109\/CVPR46437.2021.00500"},{"key":"5_CR12","doi-asserted-by":"crossref","unstructured":"Hu, J., Liao, X., Liang, J., Zhou, W., Qin, Z.: FInfer: frame inference-based deepfake detection for high-visual-quality videos. In: Thirty-Sixth AAAI Conference on Artificial Intelligence, pp. 951\u2013959. AAAI Press, Virtual Event (2022)","DOI":"10.1609\/aaai.v36i1.19978"},{"key":"5_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1007\/978-3-030-95398-0_4","volume-title":"Digital Forensics and Watermarking","author":"Y Hu","year":"2022","unstructured":"Hu, Y., Zhao, H., Yu, Z., Liu, B., Yu, X.: Exposing deepfake videos with spatial, frequency and multi-scale temporal artifacts. In: Zhao, X., Piva, A., Comesa\u00f1a-Alfaro, P. (eds.) IWDW 2021. LNCS, vol. 13180, pp. 47\u201357. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-030-95398-0_4"},{"key":"5_CR14","doi-asserted-by":"publisher","unstructured":"Hu, Z., Xie, H., Wang, Y., Li, J., Wang, Z., Zhang, Y.: Dynamic inconsistency-aware deepfake video detection. In: Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, pp. 736\u2013742. Ijcai.org, Virtual Event\/Montreal, Canada (2021). https:\/\/doi.org\/10.24963\/ijcai.2021\/102","DOI":"10.24963\/ijcai.2021\/102"},{"key":"5_CR15","unstructured":"Jiang, Y., Chang, S., Wang, Z.: TransGAN: two pure transformers can make one strong GAN, and that can scale up. In: Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, pp. 14745\u201314758. Virtual (2021)"},{"issue":"2","key":"5_CR16","doi-asserted-by":"publisher","first-page":"242","DOI":"10.1109\/TSMC.1983.6313122","volume":"13","author":"CC Lee","year":"1983","unstructured":"Lee, C.C.: Elimination of redundant operations for a fast Sobel operator. IEEE Trans. Syst. Man Cybern. 13(2), 242\u2013245 (1983). https:\/\/doi.org\/10.1109\/TSMC.1983.6313122","journal-title":"IEEE Trans. Syst. Man Cybern."},{"key":"5_CR17","doi-asserted-by":"crossref","unstructured":"Li, J., Xie, H., Li, J., Wang, Z., Zhang, Y.: Frequency-aware discriminative feature learning supervised by single-center loss for face forgery detection. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, pp. 6458\u20136467. Virtual (2021)","DOI":"10.1109\/CVPR46437.2021.00639"},{"key":"5_CR18","doi-asserted-by":"publisher","unstructured":"Li, L., et al.: Face X-ray for more general face forgery detection. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, pp. 5000\u20135009 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00505","DOI":"10.1109\/CVPR42600.2020.00505"},{"key":"5_CR19","doi-asserted-by":"publisher","unstructured":"Li, Y., Yang, X., Sun, P., Qi, H., Lyu, S.: Celeb-DF: a large-scale challenging dataset for deepfake forensics. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, pp. 3204\u20133213 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00327","DOI":"10.1109\/CVPR42600.2020.00327"},{"key":"5_CR20","doi-asserted-by":"crossref","unstructured":"Liu, H., et al.: Spatial-phase shallow learning: Rethinking face forgery detection in frequency domain. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, pp. 772\u2013781. Virtual (2021)","DOI":"10.1109\/CVPR46437.2021.00083"},{"key":"5_CR21","doi-asserted-by":"publisher","unstructured":"Liu, R., et al.: FuseFormer: fusing fine-grained information in transformers for video inpainting. In: 2021 IEEE\/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, pp. 14020\u201314029 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.01378","DOI":"10.1109\/ICCV48922.2021.01378"},{"key":"5_CR22","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"5_CR23","doi-asserted-by":"publisher","unstructured":"Luo, Y., Zhang, Y., Yan, J., Liu, W.: Generalizing face forgery detection with high-frequency features. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, pp. 16317\u201316326. Computer Vision Foundation\/IEEE, Virtual (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.01605","DOI":"10.1109\/CVPR46437.2021.01605"},{"key":"5_CR24","unstructured":"Pei, P., Zhao, X., Li, J., Cao, Y., Yi, X.: Vision transformer based video hashing retrieval for tracing the source of fake videos. CoRR abs\/2112.08117 (2021). https:\/\/arxiv.org\/abs\/2112.08117"},{"key":"5_CR25","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1007\/978-3-030-58610-2_6","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Qian","year":"2020","unstructured":"Qian, Y., Yin, G., Sheng, L., Chen, Z., Shao, J.: Thinking in frequency: face forgery detection by mining frequency-aware clues. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12357, pp. 86\u2013103. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_6"},{"key":"5_CR26","doi-asserted-by":"publisher","unstructured":"R\u00f6ssler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Nie\u00dfner, M.: FaceForensics++: learning to detect manipulated facial images. In: 2019 IEEE\/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea (South), pp. 1\u201311 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00009","DOI":"10.1109\/ICCV.2019.00009"},{"key":"5_CR27","doi-asserted-by":"publisher","unstructured":"Sorkine, O., Cohen-Or, D., Lipman, Y., Alexa, M., R\u00f6ssl, C., Seidel, H.: Laplacian surface editing. In: Boissonnat, J., Alliez, P. (eds.) Second Eurographics Symposium on Geometry Processing, Nice, France, 8\u201310 July 2004. ACM International Conference Proceeding Series, Nice, France, vol.\u00a071, pp. 175\u2013184 (2004). https:\/\/doi.org\/10.2312\/SGP\/SGP04\/179-188","DOI":"10.2312\/SGP\/SGP04\/179-188"},{"key":"5_CR28","doi-asserted-by":"crossref","unstructured":"Sun, Z., Han, Y., Hua, Z., Ruan, N., Jia, W.: Improving the efficiency and robustness of deepfakes detection through precise geometric features. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, pp. 3609\u20133618. Virtual (2021)","DOI":"10.1109\/CVPR46437.2021.00361"},{"key":"5_CR29","unstructured":"Tan, M., Le, Q.V.: EfficientNet: rethinking model scaling for convolutional neural networks. In: Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, California, vol. 97, pp. 6105\u20136114 (2019)"},{"key":"5_CR30","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, pp. 5998\u20136008 (2017)"},{"key":"5_CR31","doi-asserted-by":"crossref","unstructured":"Wang, C., Deng, W.: Representative forgery mining for fake face detection. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, pp. 14923\u201314932. Virtual (2021)","DOI":"10.1109\/CVPR46437.2021.01468"},{"key":"5_CR32","unstructured":"Wang, W., Xie, E., Li, X., Fan, D.P., Shao, L.: PVTV 2: improved baselines with pyramid vision transformer. CoRR abs\/2106.13797 (2021)"},{"key":"5_CR33","doi-asserted-by":"publisher","unstructured":"Wang, W., et al.: Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. In: 2021 IEEE\/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, pp. 548\u2013558 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00061","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"5_CR34","doi-asserted-by":"publisher","first-page":"1841","DOI":"10.1109\/TIFS.2020.3045937","volume":"16","author":"C Yang","year":"2021","unstructured":"Yang, C., Ma, J., Wang, S., Liew, A.W.: Preventing deepfake attacks on speaker authentication by dynamic lip movement analysis. IEEE Trans. Inf. Forensics Secur. 16, 1841\u20131854 (2021). https:\/\/doi.org\/10.1109\/TIFS.2020.3045937","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"5_CR35","doi-asserted-by":"publisher","first-page":"4234","DOI":"10.1109\/TIFS.2021.3102487","volume":"16","author":"J Yang","year":"2021","unstructured":"Yang, J., Li, A., Xiao, S., Lu, W., Gao, X.: MTD-net: Learning to detect deepfakes images by multi-scale texture difference. IEEE Trans. Inf. Forensics Secur. 16, 4234\u20134245 (2021). https:\/\/doi.org\/10.1109\/TIFS.2021.3102487","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"5_CR36","unstructured":"Yuan, Y., et al.: HRFormer: high-resolution vision transformer for dense predict. In: Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, pp. 7281\u20137293. Virtual (2021)"},{"issue":"10","key":"5_CR37","doi-asserted-by":"publisher","first-page":"1499","DOI":"10.1109\/LSP.2016.2603342","volume":"23","author":"K Zhang","year":"2016","unstructured":"Zhang, K., Zhang, Z., Li, Z., Qiao, Y.: Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 23(10), 1499\u20131503 (2016). https:\/\/doi.org\/10.1109\/LSP.2016.2603342","journal-title":"IEEE Signal Process. Lett."},{"key":"5_CR38","doi-asserted-by":"crossref","unstructured":"Zhao, H., Zhou, W., Chen, D., Wei, T., Zhang, W., Yu, N.: Multi-attentional deepfake detection. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, pp. 2185\u20132194. Virtual (2021)","DOI":"10.1109\/CVPR46437.2021.00222"},{"key":"5_CR39","doi-asserted-by":"publisher","unstructured":"Zhao, T., Xu, X., Xu, M., Ding, H., Xiong, Y., Xia, W.: Learning self-consistency for deepfake detection. In: 2021 IEEE\/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, pp. 15003\u201315013 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.01475","DOI":"10.1109\/ICCV48922.2021.01475"}],"container-title":["Lecture Notes in Computer Science","Digital Forensics and Watermarking"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25115-3_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,28]],"date-time":"2023-01-28T13:58:38Z","timestamp":1674914318000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25115-3_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031251146","9783031251153"],"references-count":39,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25115-3_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"29 January 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"IWDW","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Digital Watermarking","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Guilin","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 November 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 November 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iwdw2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/iwdw.site\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"30","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"14","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"47% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.23","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1.43","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}