{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,24]],"date-time":"2026-02-24T18:08:09Z","timestamp":1771956489491,"version":"3.50.1"},"publisher-location":"Cham","reference-count":40,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030712914","type":"print"},{"value":"9783030712921","type":"electronic"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-71292-1_8","type":"book-chapter","created":{"date-parts":[[2021,3,19]],"date-time":"2021-03-19T00:16:30Z","timestamp":1616112990000},"page":"78-94","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":20,"title":["Spatio-Temporal Deepfake Detection with Deep Neural Networks"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8473-544X","authenticated-orcid":false,"given":"Andrey","family":"Sebyakin","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0338-1227","authenticated-orcid":false,"given":"Vladimir","family":"Soloviev","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1238-004X","authenticated-orcid":false,"given":"Anatoly","family":"Zolotaryuk","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,3,19]]},"reference":[{"key":"8_CR1","doi-asserted-by":"publisher","unstructured":"Sebyakin, A.S., Zolotaryuk, A.V.: Tracking emotional state of a person with artificial intelligence methods and its application to customer services. In: Proceedings of the 2019 Twelfth International Conference \u201cManagement of Large-Scale System Development\u201d (MLSD), pp. 1\u20135 (2019). https:\/\/doi.org\/10.1109\/MLSD.2019.8911054","DOI":"10.1109\/MLSD.2019.8911054"},{"key":"8_CR2","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1016\/j.inffus.2020.06.014","volume":"64","author":"R Tolosana","year":"2020","unstructured":"Tolosana, R., Vera-Rodriguez, R., Fierrez, J., Morales, A., Ortega-Garcia, J.: Deepfakes and beyond: a survey of face manipulation and fake detection. Inf. Fusion 64, 131\u2013148 (2020). https:\/\/doi.org\/10.1016\/j.inffus.2020.06.014","journal-title":"Inf. Fusion"},{"key":"8_CR3","unstructured":"Korshunov, P., Marcel, S.: DeepFakes: a new threat to face recognition? Assessment and detection. Idiap-RR-18-2018 (2018). https:\/\/arxiv.org\/pdf\/1812.08685.pdf"},{"key":"8_CR4","doi-asserted-by":"publisher","unstructured":"Nirkin, Y., Masi, I., Tuan, A.T., Hassner, T., Medioni, G.: On face segmentation, face swapping, and face perception. In: Proceedings of the 2018 13th IEEE International Conference on Automatic Face Gesture Recognition (FG 2018), pp. 98\u2013105 (2018). https:\/\/doi.org\/10.1109\/FG.2018.00024","DOI":"10.1109\/FG.2018.00024"},{"key":"8_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.compeleceng.2020.106642","volume":"86","author":"M Ashok Kumar","year":"2020","unstructured":"Ashok Kumar, M., Rajeyyagari, S.: A novel mechanism for dynamic multifarious and disturbed human face recognition using advanced stance coalition (ASC). Comput. Electr. Eng. 86, 1\u20139 (2020). https:\/\/doi.org\/10.1016\/j.compeleceng.2020.106642","journal-title":"Comput. Electr. Eng."},{"key":"8_CR6","unstructured":"Biggio, B., Korshunov, P., Mensink, T., Patrini, G., Rao, D., Sadhu, A.: Synthetic realities: deep learning for detecting audiovisual fakes. https:\/\/sites.google.com\/view\/audiovisualfakes-icml2019. Accessed 25 Dec 2020"},{"key":"8_CR7","doi-asserted-by":"publisher","unstructured":"Ding, Zh.,Guo, Y., Zhang, L., Fu, Y.: One-shot face recognition via generative learning. In: Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition, pp.1\u20137 (2018). https:\/\/doi.org\/10.1109\/FG.2018.00011","DOI":"10.1109\/FG.2018.00011"},{"key":"8_CR8","unstructured":"Canton, C., et al.: Applications of computer vision and pattern recognition to media forensics. https:\/\/sites.google.com\/view\/mediaforensics2019. Accessed 25 Dec 2020"},{"key":"8_CR9","doi-asserted-by":"publisher","unstructured":"Verdoliva, L., Bestagini, P.: Multimedia forensics. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 2701\u20132702 (2019). https:\/\/doi.org\/10.1145\/3343031.3350542","DOI":"10.1145\/3343031.3350542"},{"key":"8_CR10","unstructured":"Raja, K., et al.: Workshop on deepfakes and presentation attacks in biometrics. https:\/\/sites.google.com\/view\/wacv2020-deeppab. Accessed 25 Dec 2020"},{"key":"8_CR11","unstructured":"Barni, M., Battiato, S., Boato, G., Farid, H., Memon, N.: Multimedia forensics in the wild. https:\/\/iplab.dmi.unict.it\/mmforwild. Accessed 25 Dec 2020"},{"key":"8_CR12","doi-asserted-by":"publisher","unstructured":"Hosler, B.C., Stamm, M.C.: Detecting video speed manipulation. In: Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 2860\u20132869 (2020). https:\/\/doi.org\/10.1109\/CVPRW50498.2020.00343","DOI":"10.1109\/CVPRW50498.2020.00343"},{"issue":"6380","key":"8_CR13","doi-asserted-by":"publisher","first-page":"1094","DOI":"10.1126\/science.aao2998","volume":"359","author":"DMJ Lazer","year":"2018","unstructured":"Lazer, D.M.J., et al.: The science of fake news. Science 359(6380), 1094\u20131096 (2018). https:\/\/doi.org\/10.1126\/science.aao2998","journal-title":"Science"},{"key":"8_CR14","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of GANs for improved quality, stability, and variation. In: Proceedings of the Sixth International Conference on Learning Representations, pp. 1\u201326 (2018). https:\/\/arxiv.org\/pdf\/1710.10196.pdf"},{"key":"8_CR15","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4401\u20134410 (2019). https:\/\/arxiv.org\/pdf\/1812.04948v3.pdf"},{"key":"8_CR16","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of StyleGAN. In: Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8110\u20138119 (2020). https:\/\/arxiv.org\/pdf\/1912.04958.pdf"},{"issue":"8","key":"8_CR17","doi-asserted-by":"publisher","first-page":"2001","DOI":"10.1109\/TIFS.2018.2807791","volume":"13","author":"E Gonzalez-Sosa","year":"2018","unstructured":"Gonzalez-Sosa, E., Fierrez, J., Vera-Rodriguez, R., Alonso-Fernandez, F.: Facial soft biometrics for recognition in the wild: recent works, annotation and COTS evaluation. IEEE Trans. Inf. Forensics Secur. 13(8), 2001\u20132014 (2018). https:\/\/doi.org\/10.1109\/TIFS.2018.2807791","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"8_CR18","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8789\u20138797 (2018). https:\/\/arxiv.org\/pdf\/1711.09020.pdf"},{"key":"8_CR19","unstructured":"Liu, M., et al.: STGAN: a unified selective transfer network for arbitrary image attribute editing. In: Proceedings of the 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3673\u20133682 (2019). https:\/\/arxiv.org\/pdf\/1904.09709.pdf"},{"issue":"4","key":"8_CR20","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1145\/3306346.3323035","volume":"38","author":"J Thies","year":"2019","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Deferred neural rendering: image synthesis using neural textures. ACM Trans. Graph. 38(4), 66 (2019). https:\/\/doi.org\/10.1145\/3306346.3323035","journal-title":"ACM Trans. Graph."},{"key":"8_CR21","unstructured":"R\u00f6ssler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Nie\u00dfner, M.: FaceForensics++: learning to detect manipulated facial images. In: Proceedings of the 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, p. 11 (2019). https:\/\/arxiv.org\/pdf\/1901.08971.pdf"},{"key":"8_CR22","unstructured":"Li, Y., Yang, X., Sun, P., Qi, H., Lyu, S.: Celeb-DF: a large-scale challenging dataset for DeepFake forensics. In: Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3207\u20133216 (2020). https:\/\/arxiv.org\/pdf\/1909.12962.pdf"},{"key":"8_CR23","doi-asserted-by":"publisher","unstructured":"Cao, Q., Shen, L., Xie, W., Parkhi, O., Zisserman, A.: VGGFace2: a dataset for recognizing faces across pose and age. In: Proceedings of the 13th IEEE International Conference on Automatic Face & Gesture Recognition, pp. 67\u201374 (2018). https:\/\/doi.org\/10.1109\/FG.2018.00020","DOI":"10.1109\/FG.2018.00020"},{"key":"8_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"70","DOI":"10.1007\/978-3-030-58610-2_5","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Zhang","year":"2020","unstructured":"Zhang, Y., et al.: CelebA-spoof: large-scale face anti-spoofing dataset with rich annotations. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV. LNCS, vol. 12357, pp. 70\u201385. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_5"},{"key":"8_CR25","unstructured":"Dang, H., Liu, F., Stehouwer, J., Liu, X., Jain, A.: On the detection of digital face manipulation. In: Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5781\u20135790 (2020). https:\/\/arxiv.org\/pdf\/1910.01717.pdf"},{"issue":"5","key":"8_CR26","doi-asserted-by":"publisher","first-page":"1038","DOI":"10.1109\/JSTSP.2020.3007250","volume":"14","author":"J Neves","year":"2020","unstructured":"Neves, J., Tolosana, R., Vera-Rodriguez, R., Lopes, V., Proen\u00e7a, H., Fierrez, J.: GANprintR: improved fakes and evaluation of the state-of-the-art in face manipulation detection. IEEE J. Sel. Top. Signal Process. 14(5), 1038\u20131048 (2020). https:\/\/doi.org\/10.1109\/JSTSP.2020.3007250","journal-title":"IEEE J. Sel. Top. Signal Process."},{"key":"8_CR27","unstructured":"McCloskey, S., Albright, M.: Detecting GAN-generated imagery using color cues (2018). https:\/\/arxiv.org\/pdf\/1812.08247.pdf"},{"key":"8_CR28","unstructured":"Wang, R., Ma, L., Juefei-Xu, F., Xie, X., Wang, J., Liu, Y.: FakeSpotter: a simple baseline for spotting AI-synthesized fake faces. In: Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, pp. 3444\u20133451 (2019). https:\/\/arxiv.org\/pdf\/1909.06122v3.pdf"},{"key":"8_CR29","unstructured":"Shen, Y., Gu, J., Tang, X., Zhou, B.: Interpreting the latent space of GANs for semantic face editing. In: Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, p. 12 (2020). https:\/\/arxiv.org\/pdf\/1907.10786.pdf"},{"key":"8_CR30","unstructured":"Guarnera, L., Giudice, O., Battiato, S.: DeepFake detection by analyzing convolutional traces. In: Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern RecognitionWorkshops, p. 10 (2020). https:\/\/arxiv.org\/pdf\/2004.10448v1.pdf"},{"issue":"11","key":"8_CR31","doi-asserted-by":"publisher","first-page":"5464","DOI":"10.1109\/TIP.2019.2916751","volume":"28","author":"Z He","year":"2019","unstructured":"He, Z., Zuo, W., Kan, M., Shan, S., Chen, X.: Attgan: facial attribute editing by only changing what you want. IEEE Trans. Image Process. 28(11), 5464\u20135478 (2019). https:\/\/doi.org\/10.1109\/TIP.2019.2916751","journal-title":"IEEE Trans. Image Process."},{"key":"8_CR32","unstructured":"Cho, W., Choi, S., Park, D.K., Shin I., Choo, J.: Image-to-image translation via group-wise deep whitening-and-coloring transformation. In: Proceedings of the 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10639\u201310647 (2019). https:\/\/arxiv.org\/pdf\/1812.09912.pdf"},{"key":"8_CR33","unstructured":"Yu, N., Davis, L., Fritz M.: Attributing fake images to GANs: analyzing fingerprints in generated images. In: Proceedings of the 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 1\u201341 (2019). https:\/\/arxiv.org\/pdf\/1811.08180.pdf"},{"key":"8_CR34","unstructured":"Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y.: Spectral normalization for generative adversarial networks. In: Proceedings of the International Conference on Learning Representations, pp. 1\u201326 (2018). https:\/\/arxiv.org\/pdf\/1802.05957.pdf"},{"key":"8_CR35","unstructured":"Binkowski, M., Sutherland, D., Arbel, M., Gretton, A.: Demystifying MMD GANs. In: Proceedings of the International Conference on Learning Representations, pp. 1\u201336 (2018). https:\/\/arxiv.org\/pdf\/1801.01401.pdf"},{"key":"8_CR36","unstructured":"Dolhansky, B., Bitton, J., Pflaum, B., Lu, J., Howes, R., Wang, M., Ferrer, C.C.: The DeepFake detection challenge dataset (2020). https:\/\/arxiv.org\/pdf\/2006.07397.pdf"},{"key":"8_CR37","unstructured":"Li, L., Bao, J., Yang, H., Chen, D., Wen, F.: FaceShifter: towards high fidelity and occlusion aware face swapping (2020). https:\/\/arxiv.org\/pdf\/1912.13457.pdf"},{"key":"8_CR38","unstructured":"Tran, D., Wang, H., Torresani L., Ray J., Le Cun, Y., Paluri, M.: A closer look at spatiotemporal convolutions for action recognition. In: Proceedings of the 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6450\u20136459 (2018). https:\/\/arxiv.org\/pdf\/1711.11248.pdf"},{"key":"8_CR39","unstructured":"Lin, J., Gan, C., Han, S.: TSM: temporal shift module for efficient video understanding. In: Proceedings of the 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 7083\u20137093 (2019). https:\/\/arxiv.org\/pdf\/1811.08383.pdf"},{"key":"8_CR40","unstructured":"Ghadiyaram, D., Mahajan, D.: Large-scale weakly-supervised pretraining for video action recognition. In: Proceedings of the 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12038\u201312047 (2019). https:\/\/arxiv.org\/pdf\/1905.00561.pdf"}],"container-title":["Lecture Notes in Computer Science","Diversity, Divergence, Dialogue"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-71292-1_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,3,19]],"date-time":"2021-03-19T00:19:39Z","timestamp":1616113179000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-71292-1_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030712914","9783030712921"],"references-count":40,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-71292-1_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"19 March 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"iConference","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Information","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Beijing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 March 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31 March 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconference2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ischools.org\/iConference","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ConfTool","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"225","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"32","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"59","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"14% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}