{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T02:41:35Z","timestamp":1774924895770,"version":"3.50.1"},"publisher-location":"Cham","reference-count":52,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198267","type":"print"},{"value":"9783031198274","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19827-4_5","type":"book-chapter","created":{"date-parts":[[2022,11,1]],"date-time":"2022-11-01T14:42:19Z","timestamp":1667313739000},"page":"73-89","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":35,"title":["Neural Strands: Learning Hair Geometry and\u00a0Appearance from\u00a0Multi-view Images"],"prefix":"10.1007","author":[{"given":"Radu Alexandru","family":"Rosu","sequence":"first","affiliation":[]},{"given":"Shunsuke","family":"Saito","sequence":"additional","affiliation":[]},{"given":"Ziyan","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Chenglei","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Sven","family":"Behnke","sequence":"additional","affiliation":[]},{"given":"Giljoo","family":"Nam","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,2]]},"reference":[{"key":"5_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"696","DOI":"10.1007\/978-3-030-58542-6_42","volume-title":"Computer Vision \u2013 ECCV 2020","author":"K-A Aliev","year":"2020","unstructured":"Aliev, K.-A., Sevastopolsky, A., Kolos, M., Ulyanov, D., Lempitsky, V.: Neural point-based graphics. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12367, pp. 696\u2013712. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58542-6_42"},{"issue":"4","key":"5_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459850","volume":"40","author":"T Bagautdinov","year":"2021","unstructured":"Bagautdinov, T., et al.: Driving-signal aware full-body avatars. ACM Trans. Graph. (TOG) 40(4), 1\u201317 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"4","key":"5_CR3","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1145\/2185520.2185613","volume":"31","author":"T Beeler","year":"2012","unstructured":"Beeler, T., et al.: Coupled 3D reconstruction of sparse facial hair and skin. ACM Trans. Graph. (ToG) 31(4), 117 (2012)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"5_CR4","doi-asserted-by":"crossref","unstructured":"Benamira, A., Pattanaik, S.: A combined scattering and diffraction model for elliptical hair rendering. In: Computer Graphics Forum, vol. 40, pp. 163\u2013175. Wiley Online Library (2021)","DOI":"10.1111\/cgf.14349"},{"issue":"6","key":"5_CR5","doi-asserted-by":"publisher","first-page":"204","DOI":"10.1145\/2816795.2818112","volume":"34","author":"M Chai","year":"2015","unstructured":"Chai, M., Luo, L., Sunkavalli, K., Carr, N., Hadap, S., Zhou, K.: High-quality hair modeling from a single portrait photo. ACM Trans. Graph. (TOG) 34(6), 204 (2015)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"5_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"371","DOI":"10.1007\/978-3-030-58523-5_22","volume-title":"Computer Vision \u2013 ECCV 2020","author":"M Chai","year":"2020","unstructured":"Chai, M., Ren, J., Tulyakov, S.: Neural hair rendering. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12363, pp. 371\u2013388. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58523-5_22"},{"key":"5_CR7","doi-asserted-by":"crossref","unstructured":"Chai, M., Shao, T., Wu, H., Weng, Y., Zhou, K.: AutoHair: fully automatic hair modeling from a single image. ACM Trans. Graph. 35(4), 1\u201312 (2016)","DOI":"10.1145\/2897824.2925961"},{"issue":"4","key":"5_CR8","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1145\/2461912.2461990","volume":"32","author":"M Chai","year":"2013","unstructured":"Chai, M., Wang, L., Weng, Y., Jin, X., Zhou, K.: Dynamic hair manipulation in images and videos. ACM Trans. Graph. (TOG) 32(4), 75 (2013)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"4","key":"5_CR9","doi-asserted-by":"publisher","first-page":"116","DOI":"10.1145\/2185520.2185612","volume":"31","author":"M Chai","year":"2012","unstructured":"Chai, M., Wang, L., Weng, Y., Yu, Y., Guo, B., Zhou, K.: Single-view hair modeling for portrait manipulation. ACM Trans. Graph. (TOG) 31(4), 116 (2012)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"5_CR10","unstructured":"Chen, R.T., Rubanova, Y., Bettencourt, J., Duvenaud, D.: Neural ordinary differential equations. arXiv preprint arXiv:1806.07366 (2018)"},{"issue":"6","key":"5_CR11","doi-asserted-by":"publisher","first-page":"146","DOI":"10.1145\/2366145.2366165","volume":"31","author":"TL Herrera","year":"2012","unstructured":"Herrera, T.L., Zinke, A., Weber, A.: Lighting hair from the inside: a thermal approach to hair reconstruction. ACM Trans. Graph. (TOG) 31(6), 146 (2012)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"5_CR12","doi-asserted-by":"crossref","unstructured":"Hu, L., Bradley, D., Li, H., Beeler, T.: Simulation-ready hair capture. In: Computer Graphics Forum, vol. 36, pp. 281\u2013294. Wiley Online Library (2017)","DOI":"10.1111\/cgf.13126"},{"issue":"4","key":"5_CR13","doi-asserted-by":"publisher","first-page":"126","DOI":"10.1145\/2601097.2601194","volume":"33","author":"L Hu","year":"2014","unstructured":"Hu, L., Ma, C., Luo, L., Li, H.: Robust hair capture using simulated examples. ACM Trans. Graph. (TOG) 33(4), 126 (2014)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"6","key":"5_CR14","doi-asserted-by":"publisher","first-page":"225","DOI":"10.1145\/2661229.2661254","volume":"33","author":"L Hu","year":"2014","unstructured":"Hu, L., Ma, C., Luo, L., Wei, L.Y., Li, H.: Capturing braided hairstyles. ACM Trans. Graph. (TOG) 33(6), 225 (2014)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"5_CR15","doi-asserted-by":"crossref","unstructured":"Jo, Y., Park, J.: SC-FEGAN: face editing generative adversarial network with user\u2019s sketch and color. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1745\u20131753 (2019)","DOI":"10.1109\/ICCV.2019.00183"},{"key":"5_CR16","unstructured":"Karras, T., et al.: Alias-free generative adversarial networks. arXiv preprint arXiv:2106.12423 (2021)"},{"issue":"2","key":"5_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2998578","volume":"36","author":"P Khungurn","year":"2017","unstructured":"Khungurn, P., Marschner, S.: Azimuthal scattering from elliptical hair fibers. ACM Trans. Graph. (TOG) 36(2), 1\u201323 (2017)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"5_CR18","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013)"},{"issue":"6","key":"5_CR19","first-page":"1","volume":"36","author":"T Li","year":"2017","unstructured":"Li, T., Bolkart, T., Black, M.J., Li, H., Romero, J.: Learning a model of facial shape and expression from 4D scans. ACM Trans. Graph. 36(6), 1\u2013194 (2017)","journal-title":"ACM Trans. Graph."},{"key":"5_CR20","doi-asserted-by":"crossref","unstructured":"Liu, S., Li, T., Chen, W., Li, H.: Soft rasterizer: a differentiable renderer for image-based 3D reasoning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7708\u20137717 (2019)","DOI":"10.1109\/ICCV.2019.00780"},{"key":"5_CR21","doi-asserted-by":"publisher","unstructured":"Lombardi, S., Saragih, J., Simon, T., Sheikh, Y.: Deep appearance models for face rendering. ACM Trans. Graph. 37(4), 1\u201313 (2018). https:\/\/doi.org\/10.1145\/3197517.3201401","DOI":"10.1145\/3197517.3201401"},{"key":"5_CR22","doi-asserted-by":"publisher","unstructured":"Lombardi, S., Simon, T., Saragih, J., Schwartz, G., Lehrmann, A., Sheikh, Y.: Neural volumes: Learning dynamic renderable volumes from images. ACM Trans. Graph. 38(4), 1\u201314 (2019). https:\/\/doi.org\/10.1145\/3306346.3323020","DOI":"10.1145\/3306346.3323020"},{"key":"5_CR23","doi-asserted-by":"publisher","unstructured":"Lombardi, S., Simon, T., Schwartz, G., Zollhoefer, M., Sheikh, Y., Saragih, J.: Mixture of volumetric primitives for efficient neural rendering. ACM Trans. Graph. 40(4), 1\u201313 (2021). https:\/\/doi.org\/10.1145\/3450626.3459863","DOI":"10.1145\/3450626.3459863"},{"key":"5_CR24","doi-asserted-by":"crossref","unstructured":"Luo, L., Li, H., Paris, S., Weise, T., Pauly, M., Rusinkiewicz, S.: Multi-view hair capture using orientation fields. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1490\u20131497 (2012). IEEE (2012)","DOI":"10.1109\/CVPR.2012.6247838"},{"issue":"4","key":"5_CR25","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1145\/2461912.2462026","volume":"32","author":"L Luo","year":"2013","unstructured":"Luo, L., Li, H., Rusinkiewicz, S.: Structure-aware hair capture. ACM Trans. Graph. (TOG) 32(4), 76 (2013)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"5_CR26","doi-asserted-by":"crossref","unstructured":"Mehta, I., Gharbi, M., Barnes, C., Shechtman, E., Ramamoorthi, R., Chandraker, M.: Modulated periodic activations for generalizable local functional representations. arXiv preprint arXiv:2104.03960 (2021)","DOI":"10.1109\/ICCV48922.2021.01395"},{"key":"5_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"5_CR28","doi-asserted-by":"crossref","unstructured":"Nam, G., Wu, C., Kim, M.H., Sheikh, Y.: Strand-accurate multi-view hair capture. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 155\u2013164 (2019)","DOI":"10.1109\/CVPR.2019.00024"},{"key":"5_CR29","unstructured":"Nilsson, J., Akenine-M\u00f6ller, T.: Understanding SSIM. CoRR abs\/2006.13846 (2020)"},{"key":"5_CR30","doi-asserted-by":"crossref","unstructured":"Olszewski, K., et al.: Intuitive, interactive beard and hair synthesis with generative models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7446\u20137456 (2020)","DOI":"10.1109\/CVPR42600.2020.00747"},{"key":"5_CR31","doi-asserted-by":"crossref","unstructured":"Park, K., et al.: Nerfies: deformable neural radiance fields. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5865\u20135874 (2021)","DOI":"10.1109\/ICCV48922.2021.00581"},{"key":"5_CR32","unstructured":"Patel, Y., Appalaraju, S., Manmatha, R.: Deep perceptual compression. CoRR abs\/1907.08310 (2019). arxiv.org\/abs\/1907.08310"},{"key":"5_CR33","doi-asserted-by":"crossref","unstructured":"Qiu, H., Wang, C., Zhu, H., Zhu, X., Gu, J., Han, X.: Two-phase hair image synthesis by self-enhancing generative model. In: Computer Graphics Forum, vol. 38, pp. 403\u2013412. Wiley Online Library (2019)","DOI":"10.1111\/cgf.13847"},{"key":"5_CR34","doi-asserted-by":"crossref","unstructured":"R\u00fcckert, D., Franke, L., Stamminger, M.: ADOP: approximate differentiable one-pixel point rendering. arXiv preprint arXiv:2110.06635 (2021)","DOI":"10.1145\/3528223.3530122"},{"issue":"6","key":"5_CR35","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3272127.3275019","volume":"37","author":"S Saito","year":"2018","unstructured":"Saito, S., Hu, L., Ma, C., Ibayashi, H., Luo, L., Li, H.: 3D hair synthesis using volumetric variational autoencoders. ACM Trans. Graph. (TOG) 37(6), 1\u201312 (2018)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"5_CR36","doi-asserted-by":"crossref","unstructured":"Saito, S., Huang, Z., Natsume, R., Morishima, S., Kanazawa, A., Li, H.: PIFu: pixel-aligned implicit function for high-resolution clothed human digitization. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2304\u20132314 (2019)","DOI":"10.1109\/ICCV.2019.00239"},{"key":"5_CR37","unstructured":"Sitzmann, V., Martel, J., Bergman, A., Lindell, D., Wetzstein, G.: Implicit neural representations with periodic activation functions. In: Advances in Neural Information Processing Systems 33 (2020)"},{"key":"5_CR38","unstructured":"Sun, T., Nam, G., Aliaga, C., Hery, C., Ramamoorthi, R.: Human hair inverse rendering using multi-view photometric data (2021)"},{"key":"5_CR39","doi-asserted-by":"crossref","unstructured":"Tan, Z., et al.: Michigan: multi-input-conditioned hair image generation for portrait editing. arXiv preprint arXiv:2010.16417 (2020)","DOI":"10.1145\/3386569.3392488"},{"key":"5_CR40","unstructured":"Tewari, A., et al.: State of the art on neural rendering. Computer Graphics Forum (EG STAR 2020) (2020)"},{"key":"5_CR41","doi-asserted-by":"crossref","unstructured":"Tewari, A., et al.: MoFA: model-based deep convolutional face autoencoder for unsupervised monocular reconstruction. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 1274\u20131283 (2017)","DOI":"10.1109\/ICCVW.2017.153"},{"issue":"4","key":"5_CR42","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3323035","volume":"38","author":"J Thies","year":"2019","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Deferred neural rendering: image synthesis using neural textures. ACM Trans. Graph. (TOG) 38(4), 1\u201312 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"5_CR43","doi-asserted-by":"crossref","unstructured":"Tran, L., Liu, X.: Nonlinear 3D face morphable model. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7346\u20137355 (2018)","DOI":"10.1109\/CVPR.2018.00767"},{"key":"5_CR44","doi-asserted-by":"publisher","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004). https:\/\/doi.org\/10.1109\/TIP.2003.819861","DOI":"10.1109\/TIP.2003.819861"},{"key":"5_CR45","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1007\/978-3-030-01225-0_7","volume-title":"Computer Vision \u2013 ECCV 2018","author":"L Wei","year":"2018","unstructured":"Wei, L., Hu, L., Kim, V., Yumer, E., Li, H.: Real-time hair rendering using sequential adversarial networks. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11208, pp. 105\u2013122. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01225-0_7"},{"key":"5_CR46","doi-asserted-by":"crossref","unstructured":"Xiang, D., Prada, F., Wu, C., Hodgins, J.: MonoClothCap: towards temporally coherent clothing capture from monocular RGB video. In: 2020 International Conference on 3D Vision (3DV), pp. 322\u2013332. IEEE (2020)","DOI":"10.1109\/3DV50981.2020.00042"},{"issue":"6","key":"5_CR47","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3355089.3356511","volume":"38","author":"L Yang","year":"2019","unstructured":"Yang, L., Shi, Z., Zheng, Y., Zhou, K.: Dynamic hair modeling from monocular videos using deep neural networks. ACM Trans. Graph. (TOG) 38(6), 1\u201312 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"6","key":"5_CR48","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3355089.3356513","volume":"38","author":"W Yifan","year":"2019","unstructured":"Yifan, W., Serena, F., Wu, S., \u00d6ztireli, C., Sorkine-Hornung, O.: Differentiable surface splatting for point-based geometry processing. ACM Trans. Graph. (TOG) 38(6), 1\u201314 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"4","key":"5_CR49","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1145\/3072959.3073627","volume":"36","author":"M Zhang","year":"2017","unstructured":"Zhang, M., Chai, M., Wu, H., Yang, H., Zhou, K.: A datadriven approach to four-view image-based hair modeling. ACM Trans. Graph 36(4), 156 (2017)","journal-title":"ACM Trans. Graph"},{"issue":"2","key":"5_CR50","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1016\/j.visinf.2019.06.001","volume":"3","author":"M Zhang","year":"2019","unstructured":"Zhang, M., Zheng, Y.: Hair-GAN: recovering 3D hair structure from a single image using generative adversarial networks. Visual Inform. 3(2), 102\u2013112 (2019)","journal-title":"Visual Inform."},{"key":"5_CR51","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"5_CR52","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1007\/978-3-030-01252-6_15","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Zhou","year":"2018","unstructured":"Zhou, Y., et al.: HairNet: single-view hair reconstruction using convolutional neural networks. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11215, pp. 249\u2013265. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01252-6_15"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19827-4_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,7]],"date-time":"2024-10-07T00:59:24Z","timestamp":1728262764000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19827-4_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198267","9783031198274"],"references-count":52,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19827-4_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"2 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}