{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T18:55:36Z","timestamp":1742928936014,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":27,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819714162"},{"type":"electronic","value":"9789819714179"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-1417-9_33","type":"book-chapter","created":{"date-parts":[[2024,5,21]],"date-time":"2024-05-21T07:05:03Z","timestamp":1716275103000},"page":"354-364","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Face Reenactment Based on\u00a0Motion Field Representation"],"prefix":"10.1007","author":[{"given":"Si","family":"Zheng","sequence":"first","affiliation":[]},{"given":"Junbin","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Zhijing","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Tianshui","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Yongyi","family":"Lu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,22]]},"reference":[{"key":"33_CR1","doi-asserted-by":"crossref","unstructured":"Koujan, M.R., Doukas, M.C., Roussos, A., Zafeiriou, S.: Head2head: video-based neural head synthesis. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), pp. 16\u201323. IEEE (2020)","DOI":"10.1109\/FG47880.2020.00048"},{"issue":"1","key":"33_CR2","doi-asserted-by":"publisher","first-page":"31","DOI":"10.1109\/TBIOM.2021.3049576","volume":"3","author":"MC Doukas","year":"2021","unstructured":"Doukas, M.C., Koujan, M.R., Sharmanska, V., Roussos, A., Zafeiriou, S.: Head2head++: deep facial attributes re-targeting. IEEE Trans. Biometr. Behav. Ident. Sci. 3(1), 31\u201343 (2021)","journal-title":"IEEE Trans. Biometr. Behav. Ident. Sci."},{"key":"33_CR3","doi-asserted-by":"crossref","unstructured":"Wang, T.-C., Mallya, A., Liu, M.-Y.: One-shot free-view neural talking-head synthesis for video conferencing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10\u00a0039\u201310\u00a0049 (2021)","DOI":"10.1109\/CVPR46437.2021.00991"},{"key":"33_CR4","doi-asserted-by":"crossref","unstructured":"Wiles, O.,\u00a0Koepke, A.,\u00a0Zisserman, A.: X2face: A network for controlling face generation using images, audio, and pose codes. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 670\u2013686 (2018)","DOI":"10.1007\/978-3-030-01261-8_41"},{"issue":"4","key":"33_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073640","volume":"36","author":"S Suwajanakorn","year":"2017","unstructured":"Suwajanakorn, S., Seitz, S.M., Kemelmacher-Shlizerman, I.: Synthesizing obama: learning lip sync from audio. ACM Trans. Graph. (ToG) 36(4), 1\u201313 (2017)","journal-title":"ACM Trans. Graph. (ToG)"},{"issue":"4","key":"33_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3323035","volume":"38","author":"J Thies","year":"2019","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Deferred neural rendering: image synthesis using neural textures. ACM Trans. Graph. (TOG) 38(4), 1\u201312 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"6","key":"33_CR7","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1145\/2816795.2818056","volume":"34","author":"J Thies","year":"2015","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M., Valgaerts, L., Stamminger, M., Theobalt, C.: Real-time expression transfer for facial reenactment. ACM Trans. Graph. 34(6), 183\u20131 (2015)","journal-title":"ACM Trans. Graph."},{"key":"33_CR8","doi-asserted-by":"crossref","unstructured":"Thies, J.,\u00a0Zollhofer, M.,\u00a0Stamminger, M.,\u00a0Theobalt, C.,\u00a0Nie\u00dfner, M.: Face2face: real-time face capture and reenactment of RGB videos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2387\u20132395 (2016)","DOI":"10.1109\/CVPR.2016.262"},{"key":"33_CR9","doi-asserted-by":"crossref","unstructured":"Blanz, V.,\u00a0Vetter, T.: A morphable model for the synthesis of 3D faces. In: Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, pp. 187\u2013194 (1999)","DOI":"10.1145\/311535.311556"},{"key":"33_CR10","doi-asserted-by":"crossref","unstructured":"Wu, W.,\u00a0Zhang, Y.,\u00a0Li, C.,\u00a0Qian, C., Loy, C.C.: Reenactgan: learning to reenact faces via boundary transfer. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 603\u2013619 (2018)","DOI":"10.1007\/978-3-030-01246-5_37"},{"key":"33_CR11","doi-asserted-by":"crossref","unstructured":"Bansal, A.,\u00a0Ma, S.,\u00a0Ramanan, D.,\u00a0Sheikh, Y.: Recycle-gan: unsupervised video retargeting. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 119\u2013135 (2018)","DOI":"10.1007\/978-3-030-01228-1_8"},{"key":"33_CR12","first-page":"1144","volume":"31","author":"T-C Wang","year":"2018","unstructured":"Wang, T.-C., Liu, M.-Y., Zhu, J.-Y., Liu, G., Tao, A., Kautz, J., Catanzaro, B.: Video-to-video synthesis. Adv. Neural. Inf. Process. Syst. 31, 1144\u20131156 (2018)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"4","key":"33_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3197517.3201283","volume":"37","author":"H Kim","year":"2018","unstructured":"Kim, H., et al.: Deep video portraits. ACM Trans. Graph. (TOG) 37(4), 1\u201314 (2018)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"6","key":"33_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3130800.3130818","volume":"36","author":"H Averbuch-Elor","year":"2017","unstructured":"Averbuch-Elor, H., Cohen-Or, D., Kopf, J., Cohen, M.F.: Bringing portraits to life. ACM Trans. Graph. (TOG) 36(6), 1\u201313 (2017)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"33_CR15","doi-asserted-by":"crossref","unstructured":"Burkov, E.,\u00a0Pasechnik, I.,\u00a0Grigorev, A.,\u00a0Lempitsky, V.: Neural head reenactment with latent pose descriptors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13\u00a0786\u201313\u00a0795 (2020)","DOI":"10.1109\/CVPR42600.2020.01380"},{"key":"33_CR16","doi-asserted-by":"crossref","unstructured":"Chen, L., Maddox, R.K.,\u00a0Duan, Z.,\u00a0Xu, C.: Hierarchical cross-modal talking face generation with dynamic pixel-wise loss. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7832\u20137841 (2019)","DOI":"10.1109\/CVPR.2019.00802"},{"issue":"6","key":"33_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3272127.3275043","volume":"37","author":"J Geng","year":"2018","unstructured":"Geng, J., Shao, T., Zheng, Y., Weng, Y., Zhou, K.: Warp-guided gans for single-photo facial animation. ACM Trans. Graph. (ToG) 37(6), 1\u201312 (2018)","journal-title":"ACM Trans. Graph. (ToG)"},{"issue":"4","key":"33_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3323028","volume":"38","author":"O Fried","year":"2019","unstructured":"Fried, O., et al.: Text-based editing of talking-head video. ACM Trans. Graph. (ToG) 38(4), 1\u201314 (2019)","journal-title":"ACM Trans. Graph. (ToG)"},{"issue":"6","key":"33_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3272127.3275075","volume":"37","author":"K Nagano","year":"2018","unstructured":"Nagano, K., et al.: pagan: real-time avatars using dynamic textures. ACM Trans. Graph. (ToG) 37(6), 1\u201312 (2018)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"33_CR20","doi-asserted-by":"crossref","unstructured":"Olszewski, K., et al.: Realistic dynamic facial textures from a single image using gans. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 5429\u20135438 (2017)","DOI":"10.1109\/ICCV.2017.580"},{"key":"33_CR21","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"33_CR22","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8110\u20138119 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"issue":"5","key":"33_CR23","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3395208","volume":"39","author":"B Egger","year":"2020","unstructured":"Egger, B., et al.: 3d morphable face models-past, present, and future. ACM Trans. Graph. (TOG) 39(5), 1\u201338 (2020)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"33_CR24","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M.-Y., Wang, T.-C., Zhu, J.-Y.: Semantic image synthesis with spatially-adaptive normalization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2337\u20132346 (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"key":"33_CR25","first-page":"7137","volume":"32","author":"A Siarohin","year":"2019","unstructured":"Siarohin, A., Lathuili\u00e8re, S., Tulyakov, S., Ricci, E., Sebe, N.: First order motion model for image animation. Adv. Neural. Inf. Process. Syst. 32, 7137\u20137147 (2019)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"33_CR26","doi-asserted-by":"crossref","unstructured":"Hwang, G., Hong, S., Lee, S., Park, S., Chae, G.: Discohead: audio-and-video-driven talking head generation by disentangled control of head pose and facial expressions. arXiv preprint arXiv:2303.07697 (2023)","DOI":"10.1109\/ICASSP49357.2023.10095670"},{"key":"33_CR27","unstructured":"Bochkovskiy, A., Wang, C.-Y., Liao, H.-Y.M.: Yolov4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020)"}],"container-title":["Lecture Notes in Computer Science","Advances in Brain Inspired Cognitive Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-1417-9_33","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,21]],"date-time":"2024-05-21T07:11:20Z","timestamp":1716275480000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-1417-9_33"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819714162","9789819714179"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-1417-9_33","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"22 May 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"BICS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Brain Inspired Cognitive Systems","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kuala Lumpur","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Malaysia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 August 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 August 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"bics2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"58","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"36","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"62% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}