{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T10:29:06Z","timestamp":1742984946463,"version":"3.40.3"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031159336"},{"type":"electronic","value":"9783031159343"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-15934-3_9","type":"book-chapter","created":{"date-parts":[[2022,9,6]],"date-time":"2022-09-06T00:02:53Z","timestamp":1662422573000},"page":"101-112","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Deepfake Video Detection Exploiting Binocular Synchronization"],"prefix":"10.1007","author":[{"given":"Wenjie","family":"Wang","sequence":"first","affiliation":[]},{"given":"Zhongyuan","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Guangcheng","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Qin","family":"Zou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,15]]},"reference":[{"key":"9_CR1","doi-asserted-by":"publisher","unstructured":"Afchar, D., Nozick, V., Yamagishi, J., Echizen, I.: MesoNet: a compact facial video forgery detection network. In: 2018 IEEE International Workshop on Information Forensics and Security (WIFS), pp. 1\u20137 (2018). https:\/\/doi.org\/10.1109\/WIFS.2018.8630761","DOI":"10.1109\/WIFS.2018.8630761"},{"key":"9_CR2","doi-asserted-by":"publisher","unstructured":"Baltrusaitis, T., Zadeh, A., Lim, Y.C., Morency, L.P.: OpenFace 2.0: facial behavior analysis toolkit. In: 2018 13th IEEE International Conference on Automatic Face Gesture Recognition, FG 2018, pp. 59\u201366 (2018). https:\/\/doi.org\/10.1109\/FG.2018.00019","DOI":"10.1109\/FG.2018.00019"},{"key":"9_CR3","doi-asserted-by":"crossref","unstructured":"Cao, S., Zou, Q., Mao, X., Ye, D., Wang, Z.: Metric learning for anti-compression facial forgery detection. In: Proceedings of the 29th ACM International Conference on Multimedia, pp. 1929\u20131937 (2021)","DOI":"10.1145\/3474085.3475347"},{"key":"9_CR4","doi-asserted-by":"crossref","unstructured":"Chen, H.S., Rouhsedaghat, M., Ghani, H., Hu, S., You, S., Kuo, C.C.J.: DefakeHop: a light-weight high-performance DeepFake detector. In: 2021 IEEE International Conference on Multimedia and Expo (ICME), pp. 1\u20136. IEEE (2021)","DOI":"10.1109\/ICME51207.2021.9428361"},{"key":"9_CR5","doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8789\u20138797 (2018)","DOI":"10.1109\/CVPR.2018.00916"},{"key":"9_CR6","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1251\u20131258 (2017)","DOI":"10.1109\/CVPR.2017.195"},{"key":"9_CR7","doi-asserted-by":"crossref","unstructured":"Dang, H., Liu, F., Stehouwer, J., Liu, X., Jain, A.K.: On the detection of digital face manipulation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5781\u20135790 (2020)","DOI":"10.1109\/CVPR42600.2020.00582"},{"key":"9_CR8","doi-asserted-by":"crossref","unstructured":"Demir, I., Ciftci, U.A.: Where do deep fakes look? Synthetic face detection via gaze tracking. In: ACM Symposium on Eye Tracking Research and Applications, pp. 1\u201311 (2021)","DOI":"10.1145\/3448017.3457387"},{"key":"9_CR9","unstructured":"Goodfellow, I.J., et al.: Generative adversarial nets. In: Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2, NIPS 2014, vol. 2, pp. 2672\u20132680 (2014)"},{"key":"9_CR10","doi-asserted-by":"publisher","unstructured":"G\u00fcera, D., Delp, E.J.: Deepfake video detection using recurrent neural networks. In: 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pp. 1\u20136 (2018). https:\/\/doi.org\/10.1109\/AVSS.2018.8639163","DOI":"10.1109\/AVSS.2018.8639163"},{"key":"9_CR11","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"9_CR12","doi-asserted-by":"publisher","first-page":"83144","DOI":"10.1109\/ACCESS.2020.2988660","volume":"8","author":"T Jung","year":"2020","unstructured":"Jung, T., Kim, S., Kim, K.: DeepVision: deepfakes detection using human eye blinking pattern. IEEE Access 8, 83144\u201383154 (2020). https:\/\/doi.org\/10.1109\/ACCESS.2020.2988660","journal-title":"IEEE Access"},{"key":"9_CR13","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of GANs for improved quality, stability, and variation. arXiv preprint arXiv:1710.10196 (2017)"},{"key":"9_CR14","doi-asserted-by":"publisher","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4396\u20134405 (2019). https:\/\/doi.org\/10.1109\/CVPR.2019.00453","DOI":"10.1109\/CVPR.2019.00453"},{"key":"9_CR15","doi-asserted-by":"crossref","unstructured":"Li, L., et al.: Face X-ray for more general face forgery detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5001\u20135010 (2020)","DOI":"10.1109\/CVPR42600.2020.00505"},{"key":"9_CR16","doi-asserted-by":"publisher","unstructured":"Li, M., Liu, B., Hu, Y., Wang, Y.: Exposing deepfake videos by tracking eye movements. In: 2020 25th International Conference on Pattern Recognition (ICPR), pp. 5184\u20135189 (2021). https:\/\/doi.org\/10.1109\/ICPR48806.2021.9413139","DOI":"10.1109\/ICPR48806.2021.9413139"},{"key":"9_CR17","doi-asserted-by":"crossref","unstructured":"Li, X., et al.: Sharp multiple instance learning for deepfake video detection. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 1864\u20131872 (2020)","DOI":"10.1145\/3394171.3414034"},{"key":"9_CR18","doi-asserted-by":"publisher","unstructured":"Li, Y., Chang, M.C., Lyu, S.: In Ictu Oculi: exposing AI created fake videos by detecting eye blinking. In: 2018 IEEE International Workshop on Information Forensics and Security (WIFS), pp. 1\u20137 (2018). https:\/\/doi.org\/10.1109\/WIFS.2018.8630787","DOI":"10.1109\/WIFS.2018.8630787"},{"key":"9_CR19","unstructured":"Li, Y., Lyu, S.: Exposing deepfake videos by detecting face warping artifacts. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2019"},{"key":"9_CR20","unstructured":"de Lima, O., Franklin, S., Basu, S., Karwoski, B., George, A.: Deepfake detection using spatiotemporal convolutional networks. arXiv preprint arXiv:2006.14749 (2020)"},{"key":"9_CR21","doi-asserted-by":"crossref","unstructured":"Nguyen, H.H., Fang, F., Yamagishi, J., Echizen, I.: Multi-task learning for detecting and segmenting manipulated facial images and videos. arXiv preprint arXiv:1906.06876 (2019)","DOI":"10.1109\/BTAS46853.2019.9185974"},{"key":"9_CR22","doi-asserted-by":"crossref","unstructured":"Nguyen, H.H., Yamagishi, J., Echizen, I.: Capsule-forensics: using capsule networks to detect forged images and videos. In: 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), ICASSP 2019, pp. 2307\u20132311. IEEE (2019)","DOI":"10.1109\/ICASSP.2019.8682602"},{"key":"9_CR23","doi-asserted-by":"publisher","unstructured":"R\u00f6ssler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Niessner, M.: FaceForensics++: learning to detect manipulated facial images. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1\u201311 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00009","DOI":"10.1109\/ICCV.2019.00009"},{"issue":"5","key":"9_CR24","doi-asserted-by":"publisher","first-page":"683","DOI":"10.1016\/0042-6989(81)90076-6","volume":"21","author":"CM Schor","year":"1981","unstructured":"Schor, C.M., Tyler, C.W.: Spatio-temporal properties of Panum\u2019s fusional area. Vis. Res. 21(5), 683\u2013692 (1981)","journal-title":"Vis. Res."},{"issue":"2","key":"9_CR25","first-page":"339","volume":"3","author":"R Spivak","year":"2019","unstructured":"Spivak, R.: \u201cdeepfakes\u2019\u2019: the newest way to commit one of the oldest crimes. Georgetown Law Technol. Rev. 3(2), 339\u2013401 (2019)","journal-title":"Georgetown Law Technol. Rev."},{"issue":"4","key":"9_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3323035","volume":"38","author":"J Thies","year":"2019","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Deferred neural rendering: image synthesis using neural textures. ACM Trans. Graph. (TOG) 38(4), 1\u201312 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"9_CR27","doi-asserted-by":"publisher","unstructured":"Thies, J., Zollh\u00f6fer, M., Stamminger, M., Theobalt, C., Nie\u00dfner, M.: Face2Face: real-time face capture and reenactment of RGB videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2387\u20132395 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.262","DOI":"10.1109\/CVPR.2016.262"},{"key":"9_CR28","doi-asserted-by":"crossref","unstructured":"Li, Y., Yang, X., P.S.H.Q., Lyu, S.: Celeb-DF: a large-scale challenging dataset for deepfake forensics. In: IEEE Conference on Computer Vision and Patten Recognition (CVPR) (2020)","DOI":"10.1109\/CVPR42600.2020.00327"},{"key":"9_CR29","doi-asserted-by":"crossref","unstructured":"Zhao, T., Xu, X., Xu, M., Ding, H., Xiong, Y., Xia, W.: Learning self-consistency for deepfake detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 15023\u201315033 (2021)","DOI":"10.1109\/ICCV48922.2021.01475"}],"container-title":["Lecture Notes in Computer Science","Artificial Neural Networks and Machine Learning \u2013 ICANN 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-15934-3_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,14]],"date-time":"2022-09-14T06:06:34Z","timestamp":1663135594000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-15934-3_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031159336","9783031159343"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-15934-3_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"15 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Bristol","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icann2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/e-nns.org\/icann2022\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"561","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"255","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}