{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,27]],"date-time":"2025-07-27T07:45:37Z","timestamp":1753602337602,"version":"3.40.3"},"publisher-location":"Cham","reference-count":54,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031198113"},{"type":"electronic","value":"9783031198120"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19812-0_29","type":"book-chapter","created":{"date-parts":[[2022,10,29]],"date-time":"2022-10-29T14:03:42Z","timestamp":1667052222000},"page":"496-513","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Stereo Depth Estimation with\u00a0Echoes"],"prefix":"10.1007","author":[{"given":"Chenghao","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Kun","family":"Tian","sequence":"additional","affiliation":[]},{"given":"Bolin","family":"Ni","sequence":"additional","affiliation":[]},{"given":"Gaofeng","family":"Meng","sequence":"additional","affiliation":[]},{"given":"Bin","family":"Fan","sequence":"additional","affiliation":[]},{"given":"Zhaoxiang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Chunhong","family":"Pan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,30]]},"reference":[{"key":"29_CR1","unstructured":"Alwassel, H., Mahajan, D., Korbar, B., Torresani, L., Ghanem, B., Tran, D.: Self-supervised learning by cross-modal audio-video clustering. In: NeurIPS, pp. 9758\u20139770 (2020)"},{"key":"29_CR2","doi-asserted-by":"crossref","unstructured":"Arandjelovic, R., Zisserman, A.: Look, listen and learn. In: ICCV, pp. 609\u2013617 (2017)","DOI":"10.1109\/ICCV.2017.73"},{"key":"29_CR3","doi-asserted-by":"crossref","unstructured":"Chang, A., et al.: Matterport3D: learning from RGB-D data in indoor environments. In: 3DV, pp. 667\u2013676 (2017)","DOI":"10.1109\/3DV.2017.00081"},{"key":"29_CR4","doi-asserted-by":"crossref","unstructured":"Chang, J.R., Chen, Y.S.: Pyramid stereo matching network. In: CVPR, pp. 5410\u20135418 (2018)","DOI":"10.1109\/CVPR.2018.00567"},{"key":"29_CR5","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1007\/978-3-030-58539-6_2","volume-title":"Computer Vision \u2013 ECCV 2020","author":"C Chen","year":"2020","unstructured":"Chen, C., et al.: SoundSpaces: audio-visual navigation in\u00a03D\u00a0environments. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12351, pp. 17\u201336. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58539-6_2"},{"key":"29_CR6","doi-asserted-by":"crossref","unstructured":"Chen, Y., Xian, Y., Koepke, A., Shan, Y., Akata, Z.: Distilling audio-visual knowledge by compositional contrastive learning. In: CVPR, pp. 7016\u20137025 (2021)","DOI":"10.1109\/CVPR46437.2021.00694"},{"key":"29_CR7","unstructured":"Cheng, X., et al.: Hierarchical neural architecture search for deep stereo matching. In: NeurIPS, pp. 22158\u201322169 (2020)"},{"key":"29_CR8","doi-asserted-by":"crossref","unstructured":"Christensen, J.H., Hornauer, S., Stella, X.Y.: BatVision: learning to see 3D spatial layout with two ears. In: ICRA, pp. 1581\u20131587 (2020)","DOI":"10.1109\/ICRA40945.2020.9196934"},{"key":"29_CR9","doi-asserted-by":"crossref","unstructured":"Dwibedi, D., Misra, I., Hebert, M.: Cut, paste and learn: surprisingly easy synthesis for instance detection. In: ICCV, pp. 1301\u20131310 (2017)","DOI":"10.1109\/ICCV.2017.146"},{"key":"29_CR10","unstructured":"Gal, Y., Ghahramani, Z.: Dropout as a Bayesian approximation: representing model uncertainty in deep learning. In: ICML, pp. 1050\u20131059 (2016)"},{"key":"29_CR11","doi-asserted-by":"crossref","unstructured":"Gan, C., Huang, D., Zhao, H., Tenenbaum, J.B., Torralba, A.: Music gesture for visual sound separation. In: CVPR, pp. 10478\u201310487 (2020)","DOI":"10.1109\/CVPR42600.2020.01049"},{"key":"29_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"658","DOI":"10.1007\/978-3-030-58545-7_38","volume-title":"Computer Vision \u2013 ECCV 2020","author":"R Gao","year":"2020","unstructured":"Gao, R., Chen, C., Al-Halah, Z., Schissler, C., Grauman, K.: VisualEchoes: spatial image representation learning through echolocation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12354, pp. 658\u2013676. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58545-7_38"},{"key":"29_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1007\/978-3-030-01219-9_3","volume-title":"Computer Vision \u2013 ECCV 2018","author":"R Gao","year":"2018","unstructured":"Gao, R., Feris, R., Grauman, K.: Learning to separate object sounds by watching unlabeled video. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11207, pp. 36\u201354. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01219-9_3"},{"key":"29_CR14","doi-asserted-by":"crossref","unstructured":"Gao, R., Grauman, K.: 2.5 d visual sound. In: CVPR, pp. 324\u2013333 (2019)","DOI":"10.1109\/CVPR.2019.00041"},{"key":"29_CR15","doi-asserted-by":"crossref","unstructured":"Gao, R., Grauman, K.: Co-separating sounds of visual objects. In: ICCV, pp. 3879\u20133888 (2019)","DOI":"10.1109\/ICCV.2019.00398"},{"key":"29_CR16","doi-asserted-by":"crossref","unstructured":"Gao, R., Oh, T.H., Grauman, K., Torresani, L.: Listen to look: action recognition by previewing audio. In: CVPR, pp. 10457\u201310467 (2020)","DOI":"10.1109\/CVPR42600.2020.01047"},{"key":"29_CR17","doi-asserted-by":"crossref","unstructured":"Guo, X., Yang, K., Yang, W., Wang, X., Li, H.: Group-wise correlation stereo network. In: CVPR, pp. 3273\u20133282 (2019)","DOI":"10.1109\/CVPR.2019.00339"},{"key":"29_CR18","unstructured":"Kendall, A., Gal, Y.: What uncertainties do we need in Bayesian deep learning for computer vision? In: NeurIPS (2017)"},{"key":"29_CR19","doi-asserted-by":"crossref","unstructured":"Kendall, A., Martirosyan, H., Dasgupta, S., Henry, P.: End-to-end learning of geometry and context for deep stereo regression. In: ICCV, pp. 66\u201375 (2017)","DOI":"10.1109\/ICCV.2017.17"},{"issue":"2","key":"29_CR20","doi-asserted-by":"publisher","first-page":"373","DOI":"10.3758\/s13414-015-1015-1","volume":"78","author":"AJ Kolarik","year":"2016","unstructured":"Kolarik, A.J., Moore, B.C., Zahorik, P., Cirstea, S., Pardhan, S.: Auditory distance perception in humans: a review of cues, development, neuronal bases, and effects of sensory loss. Attention Percept. Psychophys. 78(2), 373\u2013395 (2016)","journal-title":"Attention Percept. Psychophys."},{"key":"29_CR21","unstructured":"Lakshminarayanan, B., Pritzel, A., Blundell, C.: Simple and scalable predictive uncertainty estimation using deep ensembles. In: NeurIPS (2017)"},{"key":"29_CR22","doi-asserted-by":"crossref","unstructured":"Liang, Z., et al.: Learning for disparity estimation through feature constancy. In: CVPR, pp. 2811\u20132820 (2018)","DOI":"10.1109\/CVPR.2018.00297"},{"key":"29_CR23","doi-asserted-by":"crossref","unstructured":"Morgado, P., Misra, I., Vasconcelos, N.: Robust audio-visual instance discrimination. In: CVPR, pp. 12934\u201312945 (2021)","DOI":"10.1109\/CVPR46437.2021.01274"},{"key":"29_CR24","unstructured":"Morgado, P., Nvasconcelos, N., Langlois, T., Wang, O.: Self-supervised generation of spatial audio for 360 video. In: NeurIPS (2018)"},{"key":"29_CR25","doi-asserted-by":"crossref","unstructured":"Morgado, P., Vasconcelos, N., Misra, I.: Audio-visual instance discrimination with cross-modal agreement. In: CVPR, pp. 12475\u201312486 (2021)","DOI":"10.1109\/CVPR46437.2021.01229"},{"key":"29_CR26","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"639","DOI":"10.1007\/978-3-030-01231-1_39","volume-title":"Computer Vision \u2013 ECCV 2018","author":"A Owens","year":"2018","unstructured":"Owens, A., Efros, A.A.: Audio-visual scene analysis with self-supervised multisensory features. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11210, pp. 639\u2013658. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01231-1_39"},{"key":"29_CR27","doi-asserted-by":"crossref","unstructured":"Pang, J., Sun, W., Ren, J., Yang, C., Yan, Q.: Cascade residual learning: a two-stage convolutional neural network for stereo matching. In: ICCV, pp. 878\u2013886 (2017)","DOI":"10.1109\/ICCVW.2017.108"},{"key":"29_CR28","doi-asserted-by":"crossref","unstructured":"Parida, K.K., Srivastava, S., Sharma, G.: Beyond image to depth: improving depth prediction using echoes. In: CVPR, pp. 8268\u20138277 (2021)","DOI":"10.1109\/CVPR46437.2021.00817"},{"key":"29_CR29","doi-asserted-by":"crossref","unstructured":"Poggi, M., Mattoccia, S.: Learning from scratch a confidence measure. In: BMVC, vol. 2, p. 4 (2016)","DOI":"10.5244\/C.30.46"},{"key":"29_CR30","doi-asserted-by":"crossref","unstructured":"Poggi, M., Pallotti, D., Tosi, F., Mattoccia, S.: Guided stereo matching. In: CVPR, pp. 979\u2013988 (2019)","DOI":"10.1109\/CVPR.2019.00107"},{"key":"29_CR31","doi-asserted-by":"crossref","unstructured":"Poggi, M., Tonioni, A., Tosi, F., Mattoccia, S., Di Stefano, L.: Continual adaptation for deep stereo. IEEE Trans. Pattern Anal. Mach. Intell. (2021)","DOI":"10.1109\/TPAMI.2021.3075815"},{"key":"29_CR32","doi-asserted-by":"crossref","unstructured":"Qiu, J., et al.: DeepLidar: deep surface normal guided depth prediction for outdoor scene from sparse lidar data and single color image. In: CVPR, pp. 3313\u20133322 (2019)","DOI":"10.1109\/CVPR.2019.00343"},{"key":"29_CR33","doi-asserted-by":"crossref","unstructured":"Savva, M., et al.: Habitat: a platform for embodied AI research. In: ICCV, pp. 9339\u20139347 (2019)","DOI":"10.1109\/ICCV.2019.00943"},{"key":"29_CR34","unstructured":"Schwarz, L.A.: Non-rigid registration using free-form deformations. Technische Universit\u00e4t M\u00fcnchen 6 (2007)"},{"key":"29_CR35","doi-asserted-by":"crossref","unstructured":"Shaked, A., Wolf, L.: Improved stereo matching with constant highway networks and reflective confidence learning. In: CVPR, pp. 4641\u20134650 (2017)","DOI":"10.1109\/CVPR.2017.730"},{"issue":"4","key":"29_CR36","doi-asserted-by":"publisher","first-page":"910","DOI":"10.1007\/s11263-019-01287-w","volume":"128","author":"X Song","year":"2020","unstructured":"Song, X., Zhao, X., Fang, L., Hu, H., Yu, Y.: EdgeStereo: an effective multi-task learning network for stereo matching and edge detection. Int. J. Comput. Vision 128(4), 910\u2013930 (2020)","journal-title":"Int. J. Comput. Vision"},{"key":"29_CR37","doi-asserted-by":"crossref","unstructured":"Srivastava, S., Jurie, F., Sharma, G.: Learning 2D to 3D lifting for object detection in 3D for autonomous vehicles. In: IROS, pp. 4504\u20134511 (2019)","DOI":"10.1109\/IROS40897.2019.8967624"},{"key":"29_CR38","unstructured":"Straub, J., et al.: The replica dataset: a digital replica of indoor spaces. arXiv preprint arXiv:1906.05797 (2019)"},{"key":"29_CR39","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1007\/978-3-030-01231-1_20","volume-title":"Computer Vision \u2013 ECCV 2018","author":"F Tosi","year":"2018","unstructured":"Tosi, F., Poggi, M., Benincasa, A., Mattoccia, S.: Beyond local reasoning for stereo confidence estimation with deep learning. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11210, pp. 323\u2013338. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01231-1_20"},{"key":"29_CR40","doi-asserted-by":"crossref","unstructured":"Tsiami, A., Koutras, P., Maragos, P.: Stavis: spatio-temporal audiovisual saliency network. In: CVPR, pp. 4766\u20134776 (2020)","DOI":"10.1109\/CVPR42600.2020.00482"},{"key":"29_CR41","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"638","DOI":"10.1007\/978-3-030-58548-8_37","volume-title":"Computer Vision \u2013 ECCV 2020","author":"AB Vasudevan","year":"2020","unstructured":"Vasudevan, A.B., Dai, D., Van Gool, L.: Semantic object prediction and spatial sound super-resolution with binaural sounds. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 638\u2013655. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_37"},{"key":"29_CR42","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NeurIPS (2017)"},{"key":"29_CR43","doi-asserted-by":"crossref","unstructured":"Veach, E., Guibas, L.: Bidirectional estimators for light transport. In: Photorealistic Rendering Techniques, pp. 145\u2013167 (1995)","DOI":"10.1007\/978-3-642-87825-1_11"},{"issue":"2","key":"29_CR44","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1111\/1467-9280.00319","volume":"12","author":"K Watanabe","year":"2001","unstructured":"Watanabe, K., Shimojo, S.: When sound affects vision: effects of auditory grouping on visual motion perception. Psychol. Sci. 12(2), 109\u2013116 (2001)","journal-title":"Psychol. Sci."},{"key":"29_CR45","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"722","DOI":"10.1007\/978-3-030-58452-8_42","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Watson","year":"2020","unstructured":"Watson, J., Aodha, O.M., Turmukhambetov, D., Brostow, G.J., Firman, M.: Learning stereo from single images. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 722\u2013740. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_42"},{"key":"29_CR46","doi-asserted-by":"crossref","unstructured":"Weng, X., Kitani, K.: Monocular 3D object detection with pseudo-lidar point cloud. In: ICCVW (2019)","DOI":"10.1109\/ICCVW.2019.00114"},{"key":"29_CR47","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"682","DOI":"10.1007\/978-3-030-58589-1_41","volume-title":"Computer Vision \u2013 ECCV 2020","author":"X Xiong","year":"2020","unstructured":"Xiong, X., Xiong, H., Xian, K., Zhao, C., Cao, Z., Li, X.: Sparse-to-dense depth completion revisited: sampling strategy and graph construction. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12366, pp. 682\u2013699. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58589-1_41"},{"key":"29_CR48","doi-asserted-by":"crossref","unstructured":"Xu, H., Zhang, J.: AANet: adaptive aggregation network for efficient stereo matching. In: CVPR, pp. 1959\u20131968 (2020)","DOI":"10.1109\/CVPR42600.2020.00203"},{"key":"29_CR49","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"660","DOI":"10.1007\/978-3-030-01234-2_39","volume-title":"Computer Vision \u2013 ECCV 2018","author":"G Yang","year":"2018","unstructured":"Yang, G., Zhao, H., Shi, J., Deng, Z., Jia, J.: SegStereo: exploiting semantic information for disparity estimation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 660\u2013676. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_39"},{"key":"29_CR50","doi-asserted-by":"crossref","unstructured":"Yang, K., Russell, B., Salamon, J.: Telling left from right: learning spatial correspondence of sight and sound. In: CVPR, pp. 9932\u20139941 (2020)","DOI":"10.1109\/CVPR42600.2020.00995"},{"key":"29_CR51","unstructured":"You, Y., et al.: Pseudo-lidar++: accurate depth for 3D object detection in autonomous driving. In: ICLR (2019)"},{"key":"29_CR52","doi-asserted-by":"crossref","unstructured":"Zhang, F., Prisacariu, V., Yang, R., Torr, P.H.: GA-Net: guided aggregation net for end-to-end stereo matching. In: CVPR, pp. 185\u2013194 (2019)","DOI":"10.1109\/CVPR.2019.00027"},{"issue":"9","key":"29_CR53","doi-asserted-by":"publisher","first-page":"1612","DOI":"10.1007\/s11431-020-1582-8","volume":"63","author":"CQ Zhao","year":"2020","unstructured":"Zhao, C.Q., Sun, Q.Y., Zhang, C.Z., Tang, Y., Qian, F.: Monocular depth estimation based on deep learning: an overview. SCIENCE CHINA Technol. Sci. 63(9), 1612\u20131627 (2020). https:\/\/doi.org\/10.1007\/s11431-020-1582-8","journal-title":"SCIENCE CHINA Technol. Sci."},{"key":"29_CR54","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"587","DOI":"10.1007\/978-3-030-01246-5_35","volume-title":"Computer Vision \u2013 ECCV 2018","author":"H Zhao","year":"2018","unstructured":"Zhao, H., Gan, C., Rouditchenko, A., Vondrick, C., McDermott, J., Torralba, A.: The sound of pixels. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11205, pp. 587\u2013604. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01246-5_35"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19812-0_29","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,27]],"date-time":"2023-04-27T13:08:21Z","timestamp":1682600901000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19812-0_29"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198113","9783031198120"],"references-count":54,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19812-0_29","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"30 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}