{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:13:42Z","timestamp":1775578422646,"version":"3.50.1"},"publisher-location":"Cham","reference-count":56,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031200854","type":"print"},{"value":"9783031200861","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20086-1_20","type":"book-chapter","created":{"date-parts":[[2022,11,10]],"date-time":"2022-11-10T10:31:55Z","timestamp":1668076315000},"page":"345-362","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":71,"title":["Realistic One-Shot Mesh-Based Head Avatars"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4753-4811","authenticated-orcid":false,"given":"Taras","family":"Khakhulin","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8883-9972","authenticated-orcid":false,"given":"Vanessa","family":"Sklyarova","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4118-710X","authenticated-orcid":false,"given":"Victor","family":"Lempitsky","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9880-9531","authenticated-orcid":false,"given":"Egor","family":"Zakharov","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,11]]},"reference":[{"key":"20_CR1","unstructured":"AvatarSDK. https:\/\/avatarsdk.com\/"},{"key":"20_CR2","doi-asserted-by":"crossref","unstructured":"Blanz, V., Romdhani, S., Vetter, T.: Face identification across different poses and illuminations with a 3d morphable model. In: Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition, pp. 202\u2013207 (2002)","DOI":"10.1109\/AFGR.2002.1004155"},{"key":"20_CR3","doi-asserted-by":"crossref","unstructured":"Blanz, V., Vetter, T.: A morphable model for the synthesis of 3d faces. In: SIGGRAPH 1999 (1999)","DOI":"10.1145\/311535.311556"},{"key":"20_CR4","doi-asserted-by":"crossref","unstructured":"Bulat, A., Tzimiropoulos, G.: How far are we from solving the 2D & 3D face alignment problem? (and a dataset of 230,000 3D facial landmarks). In: International Conference on Computer Vision (2017)","DOI":"10.1109\/ICCV.2017.116"},{"key":"20_CR5","doi-asserted-by":"crossref","unstructured":"Cao, Q., Shen, L., Xie, W., Parkhi, O.M., Zisserman, A.: VGGFace2: a dataset for recognising faces across pose and age. In: 2018 13th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2018), pp. 67\u201374 (2018)","DOI":"10.1109\/FG.2018.00020"},{"key":"20_CR6","doi-asserted-by":"crossref","unstructured":"Chung, J.S., Nagrani, A., Zisserman, A.: VoxCeleb2: deep speaker recognition. In: INTERSPEECH (2018)","DOI":"10.21437\/Interspeech.2018-1929"},{"key":"20_CR7","unstructured":"Doukas, M.C., Zafeiriou, S., Sharmanska, V.: HeadGAN: video-and-audio-driven talking head synthesis (2021)"},{"key":"20_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3395208","volume":"39","author":"B Egger","year":"2020","unstructured":"Egger, B., et al.: 3d morphable face models-past, present, and future. ACM Trans. Graph. (TOG) 39, 1\u201338 (2020)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"20_CR9","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3414685.3417834","volume":"40","author":"Y Feng","year":"2020","unstructured":"Feng, Y., Feng, H., Black, M.J., Bolkart, T.: Learning an animatable detailed 3D face model from in-the-wild images. ACM Trans. Graph. (TOG) 40, 1\u201313 (2020)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"20_CR10","doi-asserted-by":"crossref","unstructured":"Gafni, G., Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Dynamic neural radiance fields for monocular 4D facial avatar reconstruction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2021)","DOI":"10.1109\/CVPR46437.2021.00854"},{"key":"20_CR11","unstructured":"Goodfellow, I.J., et al.: Generative adversarial nets. In: NIPS (2014)"},{"key":"20_CR12","doi-asserted-by":"crossref","unstructured":"Grassal, P.W., Prinzler, M., Leistner, T., Rother, C., Nie\u00dfner, M., Thies, J.: Neural head avatars from monocular RGB videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.01810"},{"key":"20_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"152","DOI":"10.1007\/978-3-030-58529-7_10","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Guo","year":"2020","unstructured":"Guo, J., Zhu, X., Yang, Y., Yang, F., Lei, Z., Li, S.Z.: Towards fast, accurate and stable 3D dense face alignment. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12364, pp. 152\u2013168. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58529-7_10"},{"key":"20_CR14","doi-asserted-by":"publisher","first-page":"217","DOI":"10.1137\/090771806","volume":"53","author":"N Halko","year":"2011","unstructured":"Halko, N., Martinsson, P.G., Tropp, J.A.: Finding structure with randomness: probabilistic algorithms for constructing approximate matrix decompositions. SIAM Rev. 53, 217\u2013288 (2011)","journal-title":"SIAM Rev."},{"key":"20_CR15","doi-asserted-by":"crossref","unstructured":"Hassner, T., Harel, S., Paz, E., Enbar, R.: Effective face frontalization in unconstrained images. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4295\u20134304 (2015)","DOI":"10.1109\/CVPR.2015.7299058"},{"key":"20_CR16","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"20_CR17","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local Nash equilibrium. In: Advances in Neural Information Processing Systems (2017)"},{"key":"20_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694\u2013711. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43"},{"key":"20_CR19","doi-asserted-by":"crossref","unstructured":"Kellnhofer, P., Jebe, L., Jones, A., Spicer, R.P., Pulli, K., Wetzstein, G.: Neural lumigraph rendering. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2021)","DOI":"10.1109\/CVPR46437.2021.00427"},{"issue":"4","key":"20_CR20","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1145\/3197517.3201283","volume":"37","author":"H Kim","year":"2018","unstructured":"Kim, H., et al.: Deep video portraits. ACM Trans. Graph. (TOG) 37(4), 163 (2018)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"20_CR21","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: International Conference for Learning Representations (2015)"},{"key":"20_CR22","doi-asserted-by":"crossref","unstructured":"Laine, S., Hellsten, J., Karras, T., Seol, Y., Lehtinen, J., Aila, T.: Modular primitives for high-performance differentiable rendering. ACM Trans. Graph. 39(6) (2020)","DOI":"10.1145\/3414685.3417861"},{"key":"20_CR23","first-page":"1","volume":"36","author":"T Li","year":"2017","unstructured":"Li, T., Bolkart, T., Black, M.J., Li, H., Romero, J.: Learning a model of facial shape and expression from 4D scans. ACM Trans. Graph. (TOG) 36, 1\u201317 (2017)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"20_CR24","doi-asserted-by":"crossref","unstructured":"Liu, S., Li, T., Chen, W., Li, H.: Soft rasterizer: a differentiable renderer for image-based 3D reasoning. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 7707\u20137716 (2019)","DOI":"10.1109\/ICCV.2019.00780"},{"key":"20_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3197517.3201401","volume":"37","author":"S Lombardi","year":"2018","unstructured":"Lombardi, S., Saragih, J.M., Simon, T., Sheikh, Y.: Deep appearance models for face rendering. ACM Trans. Graph. (TOG) 37, 1\u201313 (2018)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"20_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3323020","volume":"38","author":"S Lombardi","year":"2019","unstructured":"Lombardi, S., Simon, T., Saragih, J.M., Schwartz, G., Lehrmann, A.M., Sheikh, Y.: Neural volumes. ACM Trans. Graph. (TOG) 38, 1\u201314 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"20_CR27","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459863","volume":"40","author":"S Lombardi","year":"2021","unstructured":"Lombardi, S., Simon, T., Schwartz, G., Zollhoefer, M., Sheikh, Y., Saragih, J.M.: Mixture of volumetric primitives for efficient neural rendering. ACM Trans. Graph. (TOG) 40, 1\u201313 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"20_CR28","doi-asserted-by":"crossref","unstructured":"Ma, Q., Saito, S., Yang, J., Tang, S., Black, M.J.: Scale: modeling clothed humans with a surface codec of articulated local elements. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01582"},{"key":"20_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"20_CR30","doi-asserted-by":"crossref","unstructured":"Milletari, F., Navab, N., Ahmadi, S.A.: V-net: fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth International Conference on 3D Vision (3DV), pp. 565\u2013571 (2016)","DOI":"10.1109\/3DV.2016.79"},{"key":"20_CR31","doi-asserted-by":"crossref","unstructured":"Oechsle, M., Peng, S., Geiger, A.: UniSurf: unifying neural implicit surfaces and radiance fields for multi-view reconstruction. In: International Conference on Computer Vision (ICCV) (2021)","DOI":"10.1109\/ICCV48922.2021.00554"},{"key":"20_CR32","doi-asserted-by":"crossref","unstructured":"Park, J.J., Florence, P.R., Straub, J., Newcombe, R.A., Lovegrove, S.: DeepSDF: learning continuous signed distance functions for shape representation. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 165\u2013174 (2019)","DOI":"10.1109\/CVPR.2019.00025"},{"key":"20_CR33","unstructured":"Pinscreen. https:\/\/www.pinscreen.com\/"},{"key":"20_CR34","doi-asserted-by":"crossref","unstructured":"Ploumpis, S., et al.: Towards a complete 3D morphable model of the human head. IEEE Trans. Pattern Anal. Mach. Intell. (2021)","DOI":"10.1109\/TPAMI.2020.2991150"},{"key":"20_CR35","doi-asserted-by":"crossref","unstructured":"Ramon, E., et al.: H3D-Net: few-shot high-fidelity 3D head reconstruction. Proceedings of the IEEE\/CVF International Conference on Computer Vision (2021)","DOI":"10.1109\/ICCV48922.2021.00557"},{"key":"20_CR36","unstructured":"Ravi, N., et al.: Accelerating 3D deep learning with PyTorch3D. arXiv:2007.08501 (2020)"},{"key":"20_CR37","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"20_CR38","doi-asserted-by":"crossref","unstructured":"Saito, S., Simon, T., Saragih, J., Joo, H.: PifuHD: multi-level pixel-aligned implicit function for high-resolution 3D human digitization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 2020","DOI":"10.1109\/CVPR42600.2020.00016"},{"key":"20_CR39","doi-asserted-by":"crossref","unstructured":"Saito, S., Simon, T., Saragih, J.M., Joo, H.: PifuHD: multi-level pixel-aligned implicit function for high-resolution 3D human digitization. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 81\u201390 (2020)","DOI":"10.1109\/CVPR42600.2020.00016"},{"key":"20_CR40","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A.G., Zhu, M., Zhmoginov, A., Chen, L.C.: MobileNetv 2: inverted residuals and linear bottlenecks. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"20_CR41","doi-asserted-by":"crossref","unstructured":"Sanyal, S., Bolkart, T., Feng, H., Black, M.J.: Learning to regress 3D face shape and expression from an image without 3D supervision. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7755\u20137764 (2019)","DOI":"10.1109\/CVPR.2019.00795"},{"key":"20_CR42","unstructured":"Siarohin, A., Lathuili\u00e8re, S., Tulyakov, S., Ricci, E., Sebe, N.: First order motion model for image animation. In: Advances in Neural Information Processing Systems (NeurIPS) (2019)"},{"key":"20_CR43","unstructured":"Sorkine-Hornung, O.: Laplacian mesh processing. In: Eurographics (2005)"},{"key":"20_CR44","doi-asserted-by":"crossref","unstructured":"Su, S., et al.: Blindly assess image quality in the wild guided by a self-adaptive hyper network. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020)","DOI":"10.1109\/CVPR42600.2020.00372"},{"key":"20_CR45","doi-asserted-by":"crossref","unstructured":"Thies, J., Zollh\u00f6fer, M., Stamminger, M., Theobalt, C., Nie\u00dfner, M.: Face2Face: real-time face capture and reenactment of RGB videos. In: Proceedings of Computer Vision and Pattern Recognition (CVPR). IEEE (2016)","DOI":"10.1109\/CVPR.2016.262"},{"key":"20_CR46","doi-asserted-by":"crossref","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Deferred neural rendering: image synthesis using neural textures. arXiv: Computer Vision and Pattern Recognition (2019)","DOI":"10.1145\/3306346.3323035"},{"key":"20_CR47","doi-asserted-by":"crossref","unstructured":"Tran, A., Hassner, T., Masi, I., Medioni, G.G.: Regressing robust and discriminative 3D morphable models with a very deep neural network. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1493\u20131502 (2017)","DOI":"10.1109\/CVPR.2017.163"},{"key":"20_CR48","doi-asserted-by":"crossref","unstructured":"Wang, T.C., Liu, M.Y., Zhu, J.Y., Tao, A., Kautz, J., Catanzaro, B.: High-resolution image synthesis and semantic manipulation with conditional GANs. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8798\u20138807 (2018)","DOI":"10.1109\/CVPR.2018.00917"},{"key":"20_CR49","doi-asserted-by":"crossref","unstructured":"Wang, T.C., Mallya, A., Liu, M.Y.: One-shot free-view neural talking-head synthesis for video conferencing. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10034\u201310044 (2021)","DOI":"10.1109\/CVPR46437.2021.00991"},{"key":"20_CR50","doi-asserted-by":"crossref","unstructured":"Yenamandra, T., et al.: I3DMM: deep implicit 3D morphable model of human heads. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12798\u201312808 (2021)","DOI":"10.1109\/CVPR46437.2021.01261"},{"key":"20_CR51","doi-asserted-by":"crossref","unstructured":"Zakharkin, I., Mazur, K., Grigoriev, A., Lempitsky, V.S.: Point-based modeling of human clothing. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV) (2021)","DOI":"10.1109\/ICCV48922.2021.01445"},{"key":"20_CR52","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"524","DOI":"10.1007\/978-3-030-58610-2_31","volume-title":"Computer Vision \u2013 ECCV 2020","author":"E Zakharov","year":"2020","unstructured":"Zakharov, E., Ivakhnenko, A., Shysheya, A., Lempitsky, V.: Fast bi-layer neural synthesis of one-shot realistic head avatars. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12357, pp. 524\u2013540. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_31"},{"key":"20_CR53","doi-asserted-by":"crossref","unstructured":"Zakharov, E., Shysheya, A., Burkov, E., Lempitsky, V.S.: Few-shot adversarial learning of realistic neural talking head models. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00955"},{"key":"20_CR54","doi-asserted-by":"crossref","unstructured":"Zhao, J., Zhang, H.: Thin-plate spline motion model for image animation. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.00364"},{"key":"20_CR55","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Abrevaya, V.F., B\u00fchler, M.C., Chen, X., Black, M.J., Hilliges, O.: I m avatar: implicit morphable head avatars from videos. In: 2022 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.01318"},{"key":"20_CR56","doi-asserted-by":"crossref","unstructured":"Zuffi, S., Kanazawa, A., Berger-Wolf, T., Black, M.J.: Three-D safari: learning to estimate zebra pose, shape, and texture from images \u201cin the wild\u201d (2019)","DOI":"10.1109\/ICCV.2019.00546"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20086-1_20","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,7]],"date-time":"2024-10-07T22:09:52Z","timestamp":1728338992000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20086-1_20"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031200854","9783031200861"],"references-count":56,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20086-1_20","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"11 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}