{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,28]],"date-time":"2026-01-28T20:28:04Z","timestamp":1769632084487,"version":"3.49.0"},"publisher-location":"Cham","reference-count":62,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198380","type":"print"},{"value":"9783031198397","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19839-7_8","type":"book-chapter","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T11:40:06Z","timestamp":1666438806000},"page":"124-142","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":49,"title":["Responsive Listening Head Generation: A Benchmark Dataset and\u00a0Baseline"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3250-4978","authenticated-orcid":false,"given":"Mohan","family":"Zhou","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8416-9027","authenticated-orcid":false,"given":"Yalong","family":"Bai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1492-8286","authenticated-orcid":false,"given":"Wei","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7587-101X","authenticated-orcid":false,"given":"Ting","family":"Yao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4659-4935","authenticated-orcid":false,"given":"Tiejun","family":"Zhao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5990-7307","authenticated-orcid":false,"given":"Tao","family":"Mei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,23]]},"reference":[{"key":"8_CR1","unstructured":"Afouras, T., Chung, J.S., Zisserman, A.: Lrs3-ted: a large-scale dataset for visual speech recognition. arXiv preprint arXiv:1809.00496 (2018)"},{"key":"8_CR2","doi-asserted-by":"crossref","unstructured":"Bansal, A., Ma, S., Ramanan, D., Sheikh, Y.: Recycle-gan: unsupervised video retargeting. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 119\u2013135 (2018)","DOI":"10.1007\/978-3-030-01228-1_8"},{"key":"8_CR3","unstructured":"Barker, L.L.: Listening behavior (1971)"},{"key":"8_CR4","unstructured":"Beltagy, I., Peters, M.E., Cohan, A.: Longformer: the long-document transformer. arXiv preprint arXiv:2004.05150 (2020)"},{"key":"8_CR5","doi-asserted-by":"publisher","first-page":"415","DOI":"10.1111\/j.1460-2466.2005.tb02680.x","volume":"55","author":"CR Berger","year":"2005","unstructured":"Berger, C.R.: Interpersonal communication: theoretical perspectives, future prospects. J. Commun. 55, 415\u2013477 (2005)","journal-title":"J. Commun."},{"key":"8_CR6","doi-asserted-by":"crossref","unstructured":"Blanz, V., Vetter, T.: A morphable model for the synthesis of 3d faces. In: Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, pp. 187\u2013194 (1999)","DOI":"10.1145\/311535.311556"},{"key":"8_CR7","unstructured":"Bohr, P., Gargote, R., Vhorkate, R., Yawle, R., Bairagi, V.: A no reference image blur detection using cumulative probability blur detection (cpbd) metric. Int. J. Sci. Modern Eng. 1(5) (2013)"},{"key":"8_CR8","unstructured":"Buschmeier, H., et al.: Alico: a multimodal corpus for the study of active listening. In: LREC 2014, Ninth International Conference on Language Resources and Evaluation, Reykjavik, Iceland,, 26\u201331 May 2014, pp. 3638\u20133643 (2014)"},{"issue":"3","key":"8_CR9","first-page":"413","volume":"20","author":"C Cao","year":"2013","unstructured":"Cao, C., Weng, Y., Zhou, S., Tong, Y., Zhou, K.: Facewarehouse: a 3d facial expression database for visual computing. IEEE Trans. Vis. Comput. Graph. 20(3), 413\u2013425 (2013)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"8_CR10","doi-asserted-by":"crossref","unstructured":"Cassel, N.N.W.W.: Elements of face-to-face conversation for embodied conversational agents, embodied conversational agents (2000)","DOI":"10.7551\/mitpress\/2697.001.0001"},{"key":"8_CR11","unstructured":"Chung, J.S., Jamaludin, A., Zisserman, A.: You said that? arXiv preprint arXiv:1705.02966 (2017)"},{"key":"8_CR12","doi-asserted-by":"crossref","unstructured":"Chung, J.S., Nagrani, A., Zisserman, A.: Voxceleb2: deep speaker recognition. arXiv preprint arXiv:1806.05622 (2018)","DOI":"10.21437\/Interspeech.2018-1929"},{"key":"8_CR13","unstructured":"Chung, J., Gulcehre, C., Cho, K., Bengio, Y.: Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555 (2014)"},{"key":"8_CR14","doi-asserted-by":"crossref","unstructured":"Deng, Y., Yang, J., Xu, S., Chen, D., Jia, Y., Tong, X.: Accurate 3D face reconstruction with weakly-supervised learning: from single image to image set. In: IEEE Computer Vision and Pattern Recognition Workshops (2019)","DOI":"10.1109\/CVPRW.2019.00038"},{"issue":"3","key":"8_CR15","doi-asserted-by":"publisher","first-page":"258","DOI":"10.1016\/j.pec.2007.06.011","volume":"68","author":"T Fassaert","year":"2007","unstructured":"Fassaert, T., van Dulmen, S., Schellevis, F., Bensing, J.: Active listening in medical consultations: development of the active listening observation scale (alos-global). Patient Educ. Counsel. 68(3), 258\u2013264 (2007)","journal-title":"Patient Educ. Counsel."},{"issue":"5","key":"8_CR16","doi-asserted-by":"publisher","first-page":"579","DOI":"10.1002\/cav.267","volume":"19","author":"M Gillies","year":"2008","unstructured":"Gillies, M., Pan, X., Slater, M., Shawe-Taylor, J.: Responsive listening behavior. Comput. Anim. Virt. Worlds 19(5), 579\u2013589 (2008)","journal-title":"Comput. Anim. Virt. Worlds"},{"key":"8_CR17","doi-asserted-by":"crossref","unstructured":"Ginosar, S., Bar, A., Kohavi, G., Chan, C., Owens, A., Malik, J.: Learning individual styles of conversational gesture. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3497\u20133506 (2019)","DOI":"10.1109\/CVPR.2019.00361"},{"issue":"4","key":"8_CR18","doi-asserted-by":"publisher","first-page":"214","DOI":"10.1007\/BF00986881","volume":"9","author":"U Hadar","year":"1985","unstructured":"Hadar, U., Steiner, T.J., Rose, F.C.: Head movement during listening turns in conversation. J. Nonverbal Behav. 9(4), 214\u2013228 (1985)","journal-title":"J. Nonverbal Behav."},{"key":"8_CR19","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. Adv. Neural Inf. Processi. Syst. 30 (2017)"},{"key":"8_CR20","doi-asserted-by":"publisher","unstructured":"Heylen, D., Bevacqua, E., Pelachaud, C., Poggi, I., Gratch, J., Schr\u00f6der, M.: Generating listening behaviour. In: Emotion-Oriented Systems, pp. 321\u2013347. Springer, Heidleberg (2011). https:\/\/doi.org\/10.1007\/978-3-642-15184-2_17","DOI":"10.1007\/978-3-642-15184-2_17"},{"key":"8_CR21","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1007\/978-3-540-74997-4_14","volume-title":"Intelligent Virtual Agents","author":"D Heylen","year":"2007","unstructured":"Heylen, D., Bevacqua, E., Tellier, M., Pelachaud, C.: Searching for prototypical facial feedback signals. In: Pelachaud, C., Martin, J.-C., Andr\u00e9, E., Chollet, G., Karpouzis, K., Pel\u00e9, D. (eds.) IVA 2007. LNCS (LNAI), vol. 4722, pp. 147\u2013153. Springer, Heidelberg (2007). https:\/\/doi.org\/10.1007\/978-3-540-74997-4_14"},{"issue":"8","key":"8_CR22","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735\u20131780 (1997)","journal-title":"Neural Comput."},{"issue":"12","key":"8_CR23","doi-asserted-by":"publisher","first-page":"e0208030","DOI":"10.1371\/journal.pone.0208030","volume":"13","author":"P H\u00f6mke","year":"2018","unstructured":"H\u00f6mke, P., Holler, J., Levinson, S.C.: Eye blinks are perceived as communicative signals in human face-to-face interaction. PloS One 13(12), e0208030 (2018)","journal-title":"PloS One"},{"issue":"1","key":"8_CR24","first-page":"315","volume":"25","author":"JM Honeycutt","year":"2001","unstructured":"Honeycutt, J.M., Ford, S.G.: Mental imagery and intrapersonal communication: a review of research on imagined interactions (iis) and current developments. Ann. Int. Commun. Assoc. 25(1), 315\u2013345 (2001)","journal-title":"Ann. Int. Commun. Assoc."},{"key":"8_CR25","doi-asserted-by":"crossref","unstructured":"Huang, Z., Zhang, T., Heng, W., Shi, B., Zhou, S.: Real-time intermediate flow estimation for video frame interpolation. In: Proceedings of the European Conference on Computer Vision (ECCV) (2022)","DOI":"10.1007\/978-3-031-19781-9_36"},{"issue":"1","key":"8_CR26","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1080\/00094056.1995.10522637","volume":"72","author":"MR Jalongo","year":"1995","unstructured":"Jalongo, M.R.: Promoting active listening in the classroom. Childhood Educ. 72(1), 13\u201318 (1995)","journal-title":"Childhood Educ."},{"key":"8_CR27","doi-asserted-by":"crossref","unstructured":"Joo, H., Simon, T., Cikara, M., Sheikh, Y.: Towards social artificial intelligence: nonverbal social signal prediction in a triadic interaction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10873\u201310883 (2019)","DOI":"10.1109\/CVPR.2019.01113"},{"key":"8_CR28","doi-asserted-by":"publisher","first-page":"101","DOI":"10.1016\/0001-6918(70)90094-6","volume":"32","author":"A Kendon","year":"1970","unstructured":"Kendon, A.: Movement coordination in social interaction: some examples described. Acta Psychologica 32, 101\u2013125 (1970)","journal-title":"Acta Psychologica"},{"key":"8_CR29","unstructured":"Kendon, A., Harris, R.M., Key, M.R.: Organization of behavior in face-to-face interaction. Walter de Gruyter (2011)"},{"issue":"4","key":"8_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3197517.3201283","volume":"37","author":"H Kim","year":"2018","unstructured":"Kim, H.: Deep video portraits. ACM Trans. Graph. (TOG) 37(4), 1\u201314 (2018)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"8_CR31","doi-asserted-by":"crossref","unstructured":"Kong, L., et al.: Ifrnet: Intermediate feature refine network for efficient frame interpolation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1969\u20131978 (2022)","DOI":"10.1109\/CVPR52688.2022.00201"},{"key":"8_CR32","doi-asserted-by":"crossref","unstructured":"Li, L., et al.: Write-a-speaker: text-based emotional and rhythmic talking-head generation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 1911\u20131920 (2021)","DOI":"10.1609\/aaai.v35i3.16286"},{"key":"8_CR33","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"issue":"3","key":"8_CR34","doi-asserted-by":"publisher","first-page":"251","DOI":"10.1111\/j.1468-2885.1992.tb00042.x","volume":"2","author":"N Luhmann","year":"1992","unstructured":"Luhmann, N.: What is communication? Commun. Theory 2(3), 251\u2013259 (1992)","journal-title":"Commun. Theory"},{"key":"8_CR35","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1007\/11550617_3","volume-title":"Intelligent Virtual Agents","author":"RM Maatman","year":"2005","unstructured":"Maatman, R.M., Gratch, J., Marsella, S.: Natural behavior of a listening agent. In: Panayiotopoulos, T., Gratch, J., Aylett, R., Ballin, D., Olivier, P., Rist, T. (eds.) IVA 2005. LNCS (LNAI), vol. 3661, pp. 25\u201336. Springer, Heidelberg (2005). https:\/\/doi.org\/10.1007\/11550617_3"},{"issue":"1","key":"8_CR36","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1109\/T-AFFC.2011.20","volume":"3","author":"G McKeown","year":"2011","unstructured":"McKeown, G., Valstar, M., Cowie, R., Pantic, M., Schroder, M.: The semaine database: annotated multimodal records of emotionally colored conversations between a person and a limited agent. IEEE Trans. Affect. Comput. 3(1), 5\u201317 (2011)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"4","key":"8_CR37","doi-asserted-by":"publisher","first-page":"223","DOI":"10.1177\/0271121407311241","volume":"27","author":"D McNaughton","year":"2008","unstructured":"McNaughton, D., Hamlin, D., McCarthy, J., Head-Reeves, D., Schreiner, M.: Learning to listen: teaching an active listening strategy to preservice education professionals. Topics Early Childhood Spec. Educ. 27(4), 223\u2013231 (2008)","journal-title":"Topics Early Childhood Spec. Educ."},{"key":"8_CR38","unstructured":"Melis, G., Ko\u010disk\u1ef3, T., Blunsom, P.: Mogrifier lstm. arXiv preprint arXiv:1909.01792 (2019)"},{"issue":"2","key":"8_CR39","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1539\/joh.49.81","volume":"49","author":"S Mineyama","year":"2007","unstructured":"Mineyama, S., Tsutsumi, A., Takao, S., Nishiuchi, K., Kawakami, N.: Supervisors\u2019 attitudes and skills for active listening with regard to working conditions and psychological stress reactions among subordinate workers. J. Occup. Health 49(2), 81\u201387 (2007)","journal-title":"J. Occup. Health"},{"key":"8_CR40","doi-asserted-by":"crossref","unstructured":"Oertel, C., Jonell, P., Kontogiorgos, D., Mora, K.F., Odobez, J.M., Gustafson, J.: Towards an engagement-aware attentive artificial listener for multi-party interactions. Front. Rob. AI 189 (2021)","DOI":"10.3389\/frobt.2021.555913"},{"key":"8_CR41","doi-asserted-by":"crossref","unstructured":"Park, J., Lee, C., Kim, C.S.: Asymmetric bilateral motion estimation for video frame interpolation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14539\u201314548 (2021)","DOI":"10.1109\/ICCV48922.2021.01427"},{"issue":"5","key":"8_CR42","doi-asserted-by":"publisher","first-page":"453","DOI":"10.1136\/jamia.2000.0070453","volume":"7","author":"J Parker","year":"2000","unstructured":"Parker, J., Coiera, E.: Improving clinical communication: a view from psychology. J. Am. Med. Inf. Assoc. 7(5), 453\u2013461 (2000)","journal-title":"J. Am. Med. Inf. Assoc."},{"key":"8_CR43","doi-asserted-by":"crossref","unstructured":"Paysan, P., Knothe, R., Amberg, B., Romdhani, S., Vetter, T.: A 3D face model for pose and illumination invariant face recognition. In: 2009 Sixth IEEE International Conference on Advanced Video and Signal Based Surveillance, pp. 296\u2013301. IEEE (2009)","DOI":"10.1109\/AVSS.2009.58"},{"issue":"2","key":"8_CR44","doi-asserted-by":"publisher","first-page":"186","DOI":"10.1016\/j.imavis.2012.08.014","volume":"31","author":"S Petridis","year":"2013","unstructured":"Petridis, S., Martinez, B., Pantic, M.: The mahnob laughter database. Image Vision Comput. 31(2), 186\u2013202 (2013)","journal-title":"Image Vision Comput."},{"key":"8_CR45","doi-asserted-by":"crossref","unstructured":"Prajwal, K., Mukhopadhyay, R., Namboodiri, V.P., Jawahar, C.: A lip sync expert is all you need for speech to lip generation in the wild. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 484\u2013492 (2020)","DOI":"10.1145\/3394171.3413532"},{"key":"8_CR46","doi-asserted-by":"crossref","unstructured":"Ramamoorthi, R., Hanrahan, P.: An efficient representation for irradiance environment maps. In: Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques, pp. 497\u2013500 (2001)","DOI":"10.1145\/383259.383317"},{"key":"8_CR47","doi-asserted-by":"crossref","unstructured":"Ren, Y., Li, G., Chen, Y., Li, T.H., Liu, S.: Pirenderer: controllable portrait image generation via semantic neural rendering. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13759\u201313768 (2021)","DOI":"10.1109\/ICCV48922.2021.01350"},{"key":"8_CR48","doi-asserted-by":"crossref","unstructured":"Richard, A., Zollh\u00f6fer, M., Wen, Y., De la Torre, F., Sheikh, Y.: Meshtalk: 3D face animation from speech using cross-modality disentanglement. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1173\u20131182 (2021)","DOI":"10.1109\/ICCV48922.2021.00121"},{"key":"8_CR49","unstructured":"Robertson, K.: Active listening: more than just paying attention. Aust. Family Phys. 34(12) (2005)"},{"key":"8_CR50","unstructured":"Rogers, C.R., Farson, R.E.: Active listening (1957)"},{"key":"8_CR51","doi-asserted-by":"publisher","DOI":"10.4324\/9781315832920","volume-title":"Active Listening","author":"M Rost","year":"2013","unstructured":"Rost, M., Wilson, J.: Active Listening. Routledge, Abingdon (2013)"},{"key":"8_CR52","doi-asserted-by":"publisher","DOI":"10.4324\/9780203887011","volume-title":"An Integrated Approach to Communication Theory and Research","author":"DW Stacks","year":"2014","unstructured":"Stacks, D.W., Salwen, M.B.: An Integrated Approach to Communication Theory and Research. Routledge, Abingdon (2014)"},{"key":"8_CR53","first-page":"3104","volume":"27","author":"I Sutskever","year":"2014","unstructured":"Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. Adv. Neural Inf. Process. Syst. 27, 3104\u20133112 (2014)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"8_CR54","volume-title":"Origins of Human Communication","author":"M Tomasello","year":"2010","unstructured":"Tomasello, M.: Origins of Human Communication. MIT press, London (2010)"},{"key":"8_CR55","unstructured":"Vaswani, A., et al.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"8_CR56","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"700","DOI":"10.1007\/978-3-030-58589-1_42","volume-title":"Computer Vision \u2013 ECCV 2020","author":"K Wang","year":"2020","unstructured":"Wang, K., et al.: MEAD: a large-scale audio-visual dataset for emotional talking-face generation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12366, pp. 700\u2013717. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58589-1_42"},{"issue":"4","key":"8_CR57","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"8_CR58","doi-asserted-by":"crossref","unstructured":"Wu, W., Zhang, Y., Li, C., Qian, C., Loy, C.C.: Reenactgan: learning to reenact faces via boundary transfer. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 603\u2013619 (2018)","DOI":"10.1007\/978-3-030-01246-5_37"},{"key":"8_CR59","unstructured":"Zhang, C., Ni, S., Fan, Z., Li, H., Zeng, M., Budagavi, M., Guo, X.: 3d talking face with personalized pose dynamics. IEEE Trans. Vis. Comput. Graph. (2021)"},{"key":"8_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, C., et al.: Facial: synthesizing dynamic talking face with implicit attribute learning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3867\u20133876 (2021)","DOI":"10.1109\/ICCV48922.2021.00384"},{"issue":"10","key":"8_CR61","doi-asserted-by":"publisher","first-page":"1499","DOI":"10.1109\/LSP.2016.2603342","volume":"23","author":"K Zhang","year":"2016","unstructured":"Zhang, K., Zhang, Z., Li, Z., Qiao, Y.: Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 23(10), 1499\u20131503 (2016)","journal-title":"IEEE Signal Process. Lett."},{"key":"8_CR62","unstructured":"Zhu, H., Luo, M.D., Wang, R., Zheng, A.H., He, R.: Deep audio-visual learning: a survey. Int. J. Autom. Comput., 1\u201326 (2021)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19839-7_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T12:18:31Z","timestamp":1709813911000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19839-7_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198380","9783031198397"],"references-count":62,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19839-7_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"23 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}