{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,27]],"date-time":"2025-08-27T15:49:39Z","timestamp":1756309779964,"version":"3.40.3"},"publisher-location":"Cham","reference-count":60,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031198236"},{"type":"electronic","value":"9783031198243"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19824-3_15","type":"book-chapter","created":{"date-parts":[[2022,11,10]],"date-time":"2022-11-10T21:14:32Z","timestamp":1668114872000},"page":"245-262","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Depth Field Networks For Generalizable Multi-view Scene Representation"],"prefix":"10.1007","author":[{"given":"Vitor","family":"Guizilini","sequence":"first","affiliation":[]},{"given":"Igor","family":"Vasiljevic","sequence":"additional","affiliation":[]},{"given":"Jiading","family":"Fang","sequence":"additional","affiliation":[]},{"given":"Rare","family":"Ambru","sequence":"additional","affiliation":[]},{"given":"Greg","family":"Shakhnarovich","sequence":"additional","affiliation":[]},{"given":"Matthew R.","family":"Walter","sequence":"additional","affiliation":[]},{"given":"Adrien","family":"Gaidon","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,11]]},"reference":[{"key":"15_CR1","doi-asserted-by":"crossref","unstructured":"Azinovi\u0107, D., Martin-Brualla, R., Goldman, D.B., Nie\u00dfner, M., Thies, J.: Neural RGB-D surface reconstruction. arXiv preprint arXiv:2104.04532 (2021)","DOI":"10.1109\/CVPR52688.2022.00619"},{"key":"15_CR2","doi-asserted-by":"crossref","unstructured":"Chang, J.R., Chen, Y.S.: Pyramid stereo matching network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5410\u20135418 (2018)","DOI":"10.1109\/CVPR.2018.00567"},{"key":"15_CR3","doi-asserted-by":"crossref","unstructured":"Dai, A., Chang, A.X., Savva, M., Halber, M., Funkhouser, T., Nie\u00dfner, M.: ScanNet: richly-annotated 3D reconstructions of indoor scenes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5828\u20135839 (2017)","DOI":"10.1109\/CVPR.2017.261"},{"key":"15_CR4","doi-asserted-by":"crossref","unstructured":"Deng, K., Liu, A., Zhu, J.Y., Ramanan, D.: Depth-supervised NeRF: fewer views and faster training for free. arXiv preprint arXiv:2107.02791 (2021)","DOI":"10.1109\/CVPR52688.2022.01254"},{"key":"15_CR5","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"15_CR6","doi-asserted-by":"crossref","unstructured":"Eigen, D., Fergus, R.: Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. In: Proceedings of the International Conference on Computer Vision (ICCV), pp. 2650\u20132658 (2015)","DOI":"10.1109\/ICCV.2015.304"},{"key":"15_CR7","unstructured":"Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. In: Advances in Neural Information Processing Systems (NeurIPS) (2014)"},{"key":"15_CR8","doi-asserted-by":"crossref","unstructured":"Esser, P., Rombach, R., Ommer, B.: Taming transformers for high-resolution image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12873\u201312883 (2021)","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"15_CR9","doi-asserted-by":"crossref","unstructured":"Fu, H., Gong, M., Wang, C., Batmanghelich, K., Tao, D.: Deep ordinal regression network for monocular depth estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2002\u20132011 (2018)","DOI":"10.1109\/CVPR.2018.00214"},{"key":"15_CR10","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac Aodha, O., Brostow, G.J.: Unsupervised monocular depth estimation with left-right consistency. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 270\u2013279 (2017)","DOI":"10.1109\/CVPR.2017.699"},{"key":"15_CR11","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac Aodha, O., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth prediction. In: Proceedings of the International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"15_CR12","doi-asserted-by":"crossref","unstructured":"Gordon, A., Li, H., Jonschkowski, R., Angelova, A.: Depth from videos in the wild: Unsupervised monocular depth learning from unknown cameras. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/ICCV.2019.00907"},{"key":"15_CR13","doi-asserted-by":"crossref","unstructured":"Guizilini, V., Ambrus, R., Pillai, S., Raventos, A., Gaidon, A.: 3D packing for self-supervised monocular depth estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2020)","DOI":"10.1109\/CVPR42600.2020.00256"},{"key":"15_CR14","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"15_CR15","unstructured":"Im, S., Jeon, H.G., Lin, S., Kweon, I.S.: DPSNet: End-to-end deep plane sweep stereo. In: Proceedings of the International Conference on Learning Representations (ICLR) (2019)"},{"key":"15_CR16","unstructured":"Jaegle, A., et al.: Perceiver IO: a general architecture for structured inputs & outputs. arXiv preprint arXiv:2107.14795 (2021)"},{"key":"15_CR17","unstructured":"Jaegle, A., Gimeno, F., Brock, A., Vinyals, O., Zisserman, A., Carreira, J.: Perceiver: General perception with iterative attention. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 4651\u20134664 (2021)"},{"key":"15_CR18","doi-asserted-by":"crossref","unstructured":"Kendall, A., et al.: End-to-end learning of geometry and context for deep stereo regression. In: Proceedings of the International Conference on Computer Vision (ICCV), pp. 66\u201375 (2017)","DOI":"10.1109\/ICCV.2017.17"},{"key":"15_CR19","doi-asserted-by":"crossref","unstructured":"Kopf, J., Rong, X., Huang, J.B.: Robust consistent video depth estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1611\u20131621 (2021)","DOI":"10.1109\/CVPR46437.2021.00166"},{"key":"15_CR20","doi-asserted-by":"crossref","unstructured":"Kusupati, U., Cheng, S., Chen, R., Su, H.: Normal assisted stereo depth estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2189\u20132199 (2020)","DOI":"10.1109\/CVPR42600.2020.00226"},{"key":"15_CR21","doi-asserted-by":"crossref","unstructured":"Laina, I., Rupprecht, C., Belagiannis, V., Tombari, F., Navab, N.: Deeper depth prediction with fully convolutional residual networks. In: Proceedings of the International Conference on 3D Vision (3DV), pp. 239\u2013248 (2016)","DOI":"10.1109\/3DV.2016.32"},{"key":"15_CR22","unstructured":"Lee, J.H., Han, M.K., Ko, D.W., Suh, I.H.: From big to small: multi-scale local planar guidance for monocular depth estimation. arXiv:1907.10326 (2019)"},{"key":"15_CR23","doi-asserted-by":"crossref","unstructured":"Li, Z., et al.: Revisiting stereo depth estimation from a sequence-to-sequence perspective with transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6197\u20136206 (2021)","DOI":"10.1109\/ICCV48922.2021.00614"},{"key":"15_CR24","doi-asserted-by":"crossref","unstructured":"Lipson, L., Teed, Z., Deng, J.: RAFT-Stereo: multilevel recurrent field transforms for stereo matching. In: Proceedings of the International Conference on 3D Vision (3DV), pp. 218\u2013227 (2021)","DOI":"10.1109\/3DV53792.2021.00032"},{"key":"15_CR25","doi-asserted-by":"crossref","unstructured":"Liu, C., Gu, J., Kim, K., Narasimhan, S.G., Kautz, J.: Neural $$\\rm RGB\\rightarrow \\rm D\\rm $$ sensing: Depth and uncertainty from a video camera. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10986\u201310995 (2019)","DOI":"10.1109\/CVPR.2019.01124"},{"key":"15_CR26","doi-asserted-by":"crossref","unstructured":"Long, X., Liu, L., Li, W., Theobalt, C., Wang, W.: Multi-view depth estimation using epipolar spatio-temporal networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8258\u20138267 (2021)","DOI":"10.1109\/CVPR46437.2021.00816"},{"key":"15_CR27","doi-asserted-by":"crossref","unstructured":"Long, X., Liu, L., Theobalt, C., Wang, W.: Occlusion-aware depth estimation with adaptive normal constraints. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 640\u2013657 (2020)","DOI":"10.1007\/978-3-030-58545-7_37"},{"key":"15_CR28","doi-asserted-by":"crossref","unstructured":"Luo, X., Huang, J.B., Szeliski, R., Matzen, K., Kopf, J.: Consistent video depth estimation. ACM Trans. Graphics (TOG) 39(4) (2020)","DOI":"10.1145\/3386569.3392377"},{"key":"15_CR29","doi-asserted-by":"crossref","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 405\u2013421 (2020)","DOI":"10.1007\/978-3-030-58452-8_24"},{"key":"15_CR30","doi-asserted-by":"crossref","unstructured":"Newcombe, R.A., et al.: KinectFusion: real-time dense surface mapping and tracking. In: Proceedings of the IEEE International Symposium on Mixed and Augmented Reality (ISMAR) (2011)","DOI":"10.1109\/ISMAR.2011.6092378"},{"key":"15_CR31","doi-asserted-by":"crossref","unstructured":"Novotny, D., Larlus, D., Vedaldi, A.: Learning 3D object categories by looking around them. In: Proceedings of the International Conference on Computer Vision (ICCV), pp. 5218\u20135227 (2017)","DOI":"10.1109\/ICCV.2017.558"},{"key":"15_CR32","doi-asserted-by":"crossref","unstructured":"Ranftl, R., Bochkovskiy, A., Koltun, V.: Vision transformers for dense prediction. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12179\u201312188 (2021)","DOI":"10.1109\/ICCV48922.2021.01196"},{"key":"15_CR33","doi-asserted-by":"crossref","unstructured":"Ranftl, R., Lasinger, K., Hafner, D., Schindler, K., Koltun, V.: Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer. IEEE Trans. Patt. Anal. Mach. Intell. 44, 1623\u20131637 (2020)","DOI":"10.1109\/TPAMI.2020.3019967"},{"key":"15_CR34","doi-asserted-by":"crossref","unstructured":"Rematas, K., et al.: Urban radiance fields. arXiv preprint arXiv:2111.14643 (2021)","DOI":"10.1109\/CVPR52688.2022.01259"},{"key":"15_CR35","doi-asserted-by":"crossref","unstructured":"Rombach, R., Esser, P., Ommer, B.: Geometry-free view synthesis: transformers and no 3D priors. In: Proceedings of the International Conference on Computer Vision (ICCV), pp. 14356\u201314366 (2021)","DOI":"10.1109\/ICCV48922.2021.01409"},{"key":"15_CR36","doi-asserted-by":"crossref","unstructured":"Sajjadi, M.S., et al..: Scene representation transformer: Geometry-free novel view synthesis through set-latent scene representations. arXiv preprint arXiv:2111.13152 (2021)","DOI":"10.1109\/CVPR52688.2022.00613"},{"key":"15_CR37","unstructured":"Saxena, A., Chung, S., Ng, A.: Learning depth from single monocular images. In: Advances in Neural Information Processing Systems (NeurIPS) (2005)"},{"key":"15_CR38","doi-asserted-by":"crossref","unstructured":"Schonberger, J.L., Frahm, J.M.: Structure-from-motion revisited. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4104\u20134113 (2016)","DOI":"10.1109\/CVPR.2016.445"},{"issue":"1","key":"15_CR39","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40537-019-0197-0","volume":"6","author":"C Shorten","year":"2019","unstructured":"Shorten, C., Khoshgoftaar, T.M.: A survey on image data augmentation for deep learning. J. Big Data 6(1), 1\u201348 (2019)","journal-title":"J. Big Data"},{"key":"15_CR40","doi-asserted-by":"crossref","unstructured":"Shotton, J., Glocker, B., Zach, C., Izadi, S., Criminisi, A., Fitzgibbon, A.: Scene coordinate regression forests for camera relocalization in RGB-D images. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2930\u20132937 (2013)","DOI":"10.1109\/CVPR.2013.377"},{"key":"15_CR41","doi-asserted-by":"crossref","unstructured":"Shu, C., Yu, K., Duan, Z., Yang, K.: Feature-metric loss for self-supervised learning of depth and egomotion. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 572\u2013588 (2020)","DOI":"10.1007\/978-3-030-58529-7_34"},{"key":"15_CR42","unstructured":"Sitzmann, V., Rezchikov, S., Freeman, B., Tenenbaum, J., Durand, F.: Light field networks: neural scene representations with single-evaluation rendering. In: Advances in Neural Information Processing Systems (NeurIPS) (2021)"},{"key":"15_CR43","doi-asserted-by":"crossref","unstructured":"Sun, J., Xie, Y., Chen, L., Zhou, X., Bao, H.: NeuralRecon: real-time coherent 3D reconstruction from monocular video. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 15598\u201315607 (2021)","DOI":"10.1109\/CVPR46437.2021.01534"},{"key":"15_CR44","unstructured":"Tang, C., Tan, P.: BA-Net: Dense bundle adjustment network. arXiv preprint arXiv:1806.04807 (2018)"},{"key":"15_CR45","unstructured":"Teed, Z., Deng, J.: DeepV2D: video to depth with differentiable structure from motion. In: Proceedings of the International Conference on Learning Representations (ICLR) (2020)"},{"key":"15_CR46","doi-asserted-by":"crossref","unstructured":"Teed, Z., Deng, J.: RAFT: recurrent all-pairs field transforms for optical flow. In: Proceedings of the European Conference on Computer Vision (ECCV) (2020)","DOI":"10.24963\/ijcai.2021\/662"},{"key":"15_CR47","unstructured":"Teed, Z., Deng, J.: DROID-SLAM: deep visual SLAM for monocular, stereo, and RGB-D cameras. In: Advances in Neural Information Processing Systems (NeurIPS) (2021)"},{"key":"15_CR48","doi-asserted-by":"crossref","unstructured":"Teed, Z., Deng, J.: Raft-3D: Scene flow using rigid-motion embeddings. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8375\u20138384 (2021)","DOI":"10.1109\/CVPR46437.2021.00827"},{"key":"15_CR49","doi-asserted-by":"crossref","unstructured":"Ummenhofer, B., et al.: DeMoN: depth and motion network for learning monocular stereo. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5038\u20135047 (2017)","DOI":"10.1109\/CVPR.2017.596"},{"key":"15_CR50","doi-asserted-by":"crossref","unstructured":"Vasiljevic, I., et al.: Neural ray surfaces for self-supervised learning of depth and ego-motion. In: Proceedings of the International Conference on 3D Vision (3DV) (2020)","DOI":"10.1109\/3DV50981.2020.00010"},{"key":"15_CR51","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems (NeurIPS) (2017)"},{"key":"15_CR52","doi-asserted-by":"crossref","unstructured":"Watson, J., Mac Aodha, O., Prisacariu, V., Brostow, G., Firman, M.: The temporal opportunist: Self-supervised multi-frame monocular depth. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1164\u20131174 (2021)","DOI":"10.1109\/CVPR46437.2021.00122"},{"key":"15_CR53","doi-asserted-by":"crossref","unstructured":"Wei, Y., Liu, S., Rao, Y., Zhao, W., Lu, J., Zhou, J.: NerfingMVS: guided optimization of neural radiance fields for indoor multi-view stereo. In: Proceedings of the International Conference on Computer Vision (ICCV) (2021)","DOI":"10.1109\/ICCV48922.2021.00556"},{"key":"15_CR54","unstructured":"Xie, Y., et al.: Neural fields in visual computing and beyond. arXiv preprint arXiv:2111.11426 (2021)"},{"key":"15_CR55","doi-asserted-by":"crossref","unstructured":"Yao, Y., Luo, Z., Li, S., Fang, T., Quan, L.: MVSNet: depth inference for unstructured multi-view stereo. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 767\u2013783 (2018)","DOI":"10.1007\/978-3-030-01237-3_47"},{"key":"15_CR56","doi-asserted-by":"crossref","unstructured":"Yifan, W., Doersch, C., Arandjelovi\u0107, R., Carreira, J., Zisserman, A.: Input-level inductive biases for 3D reconstruction. arXiv preprint arXiv:2112.03243 (2021)","DOI":"10.1109\/CVPR52688.2022.00608"},{"key":"15_CR57","doi-asserted-by":"crossref","unstructured":"Yu, A., Ye, V., Tancik, M., Kanazawa, A.: pixelNeRF: neural radiance fields from one or few images. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2021)","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"15_CR58","doi-asserted-by":"crossref","unstructured":"Zhou, H., Ummenhofer, B., Brox, T.: DeepTAM: deep tracking and mapping. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 822\u2013838 (2018)","DOI":"10.1007\/978-3-030-01270-0_50"},{"key":"15_CR59","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.G.: Unsupervised learning of depth and ego-motion from video. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.700"},{"key":"15_CR60","doi-asserted-by":"crossref","unstructured":"Zhu, Z., et al.: NICE-SLAM: neural implicit scalable encoding for SLAM. arXiv preprint arXiv:2112.12130 (2021)","DOI":"10.1109\/CVPR52688.2022.01245"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19824-3_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,12]],"date-time":"2022-11-12T00:10:22Z","timestamp":1668211822000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19824-3_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198236","9783031198243"],"references-count":60,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19824-3_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"11 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}