{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:15:55Z","timestamp":1742912155026,"version":"3.40.3"},"publisher-location":"Cham","reference-count":16,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031621093"},{"type":"electronic","value":"9783031621109"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-62110-9_2","type":"book-chapter","created":{"date-parts":[[2024,6,1]],"date-time":"2024-06-01T01:03:59Z","timestamp":1717203839000},"page":"14-22","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhance Deepfake Video Detection Through Optical Flow Algorithms-Based CNN"],"prefix":"10.1007","author":[{"given":"Amani","family":"Alzahrani","sequence":"first","affiliation":[]},{"given":"Danda B.","family":"Rawat","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,1]]},"reference":[{"issue":"1","key":"2_CR1","doi-asserted-by":"publisher","first-page":"18","DOI":"10.3390\/jimaging9010018","volume":"9","author":"Z Akhtar","year":"2023","unstructured":"Akhtar, Z.: Deepfakes generation and detection: a short survey. J. Imag. 9(1), 18 (2023)","journal-title":"J. Imag."},{"key":"2_CR2","doi-asserted-by":"crossref","unstructured":"Amerini, I., Galteri, L., Caldelli, R., Del\u00a0Bimbo, A.: Deepfake video detection through optical flow based CNN. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops, pp.\u00a00\u20130 (2019)","DOI":"10.1109\/ICCVW.2019.00152"},{"key":"2_CR3","doi-asserted-by":"publisher","first-page":"31","DOI":"10.1016\/j.patrec.2021.03.005","volume":"146","author":"R Caldelli","year":"2021","unstructured":"Caldelli, R., Galteri, L., Amerini, I., Del Bimbo, A.: Optical flow based CNN for detection of unlearnt deepfake manipulations. Pattern Recogn. Lett. 146, 31\u201337 (2021)","journal-title":"Pattern Recogn. Lett."},{"key":"2_CR4","unstructured":"Dolhansky, B., et al.: The deepfake detection challenge (dfdc) dataset. arXiv preprint arXiv:2006.07397 (2020)"},{"key":"2_CR5","doi-asserted-by":"crossref","unstructured":"Fang, S., Wang, S., Ye, R.: Deepfake video detection through facial sparse optical flow based light CNN. J. Phys.: Conf. Series. 2224, 012014. IOP Publishing (2022)","DOI":"10.1088\/1742-6596\/2224\/1\/012014"},{"key":"2_CR6","doi-asserted-by":"publisher","first-page":"363","DOI":"10.1007\/3-540-45103-X_50","volume-title":"Image Anal.","author":"G Farneb\u00e4ck","year":"2003","unstructured":"Farneb\u00e4ck, G.: Two-frame motion estimation based on polynomial expansion. In: Bigun, J., Gustavsson, T. (eds.) Image Anal., pp. 363\u2013370. Springer Berlin Heidelberg, Berlin, Heidelberg (2003). https:\/\/doi.org\/10.1007\/3-540-45103-X_50"},{"issue":"8","key":"2_CR7","doi-asserted-by":"publisher","first-page":"2555","DOI":"10.1109\/TPAMI.2020.2976928","volume":"43","author":"TW Hui","year":"2020","unstructured":"Hui, T.W., Tang, X., Loy, C.C.: A lightweight optical flow CNN-revisiting data fidelity and regularization. IEEE Trans. Pattern Anal. Mach. Intell. 43(8), 2555\u20132569 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2_CR8","doi-asserted-by":"crossref","unstructured":"Lewis, J.K., et a.: Deepfake video detection based on spatial, spectral, and temporal inconsistencies using multimodal deep learning. In: 2020 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), pp.\u00a01\u20139. IEEE (2020)","DOI":"10.1109\/AIPR50011.2020.9425167"},{"issue":"20","key":"2_CR9","doi-asserted-by":"publisher","first-page":"7767","DOI":"10.3390\/s22207767","volume":"22","author":"S Li","year":"2022","unstructured":"Li, S., Dutta, V., He, X., Matsumaru, T.: Deep learning based one-class detection system for fake faces generated by GAN network. Sensors 22(20), 7767 (2022)","journal-title":"Sensors"},{"issue":"1","key":"2_CR10","doi-asserted-by":"publisher","first-page":"715","DOI":"10.32604\/cmc.2023.034963","volume":"75","author":"T Lu","year":"2023","unstructured":"Lu, T., Bao, Y., Li, L.: Deepfake video detection based on improved capsnet and temporal-spatial features. CMC-Comput. Mater. Continua 75(1), 715\u2013740 (2023)","journal-title":"CMC-Comput. Mater. Continua"},{"key":"2_CR11","doi-asserted-by":"crossref","unstructured":"Lyasheva, S., Rakhmankulov, R., Shleymovich, M.: Frame interpolation in video stream using optical flow methods. J. Phys.: Conf. Series. 1488, 012024. IOP Publishing (2020)","DOI":"10.1088\/1742-6596\/1488\/1\/012024"},{"issue":"7","key":"2_CR12","doi-asserted-by":"publisher","first-page":"2500","DOI":"10.3390\/s22072500","volume":"22","author":"AB Nassif","year":"2022","unstructured":"Nassif, A.B., Nasir, Q., Talib, M.A., Gouda, O.M.: Improved optical flow estimation method for deepfake videos. Sensors 22(7), 2500 (2022)","journal-title":"Sensors"},{"key":"2_CR13","doi-asserted-by":"crossref","unstructured":"Saikia, P., Dholaria, D., Yadav, P., Patel, V., Roy, M.: A hybrid CNN-LSTM model for video deepfake detection by leveraging optical flow features. In: 2022 International Joint Conference on Neural Networks (IJCNN), pp.\u00a01\u20137. IEEE (2022)","DOI":"10.1109\/IJCNN55064.2022.9892905"},{"issue":"8","key":"2_CR14","doi-asserted-by":"publisher","first-page":"9727","DOI":"10.1007\/s13369-022-07321-3","volume":"48","author":"S Suratkar","year":"2023","unstructured":"Suratkar, S., Kazi, F.: Deep fake video detection using transfer learning approach. Arab. J. Sci. Eng. 48(8), 9727\u20139737 (2023)","journal-title":"Arab. J. Sci. Eng."},{"key":"2_CR15","first-page":"9","volume":"72","author":"Z Tu","year":"2019","unstructured":"Tu, Z., et al.: A survey of variational and CNN-based optical flow techniques. Signal Process.: Image Commun. 72, 9\u201324 (2019)","journal-title":"Signal Process.: Image Commun."},{"key":"2_CR16","doi-asserted-by":"publisher","unstructured":"Zach, C., Pock, T., Bischof, H.: A duality based approach for realtime tv-l 1 optical flow. In: Pattern Recognition: 29th DAGM Symposium, Heidelberg, Germany, September 12-14, 2007. Proceedings 29, pp. 214\u2013223. Springer (2007). https:\/\/doi.org\/10.1007\/978-3-540-74936-3_22","DOI":"10.1007\/978-3-540-74936-3_22"}],"container-title":["Communications in Computer and Information Science","HCI International 2024 Posters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-62110-9_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,1]],"date-time":"2024-06-01T01:44:39Z","timestamp":1717206279000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-62110-9_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031621093","9783031621109"],"references-count":16,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-62110-9_2","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"1 June 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"HCII","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Human-Computer Interaction","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Washington DC","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 June 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 July 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"hcii2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/2024.hci.international\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}