{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T17:01:05Z","timestamp":1775667665915,"version":"3.50.1"},"publisher-location":"Cham","reference-count":59,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197680","type":"print"},{"value":"9783031197697","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19769-7_41","type":"book-chapter","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T11:40:06Z","timestamp":1666438806000},"page":"709-726","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":27,"title":["Self-distilled Feature Aggregation for\u00a0Self-supervised Monocular Depth Estimation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6792-0739","authenticated-orcid":false,"given":"Zhengming","family":"Zhou","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4015-1615","authenticated-orcid":false,"given":"Qiulei","family":"Dong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,23]]},"reference":[{"key":"41_CR1","doi-asserted-by":"crossref","unstructured":"Cardace, A., Ramirez, P.Z., Salti, S., Di Stefano, L.: Shallow features guide unsupervised domain adaptation for semantic segmentation at class boundaries. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1160\u20131170 (2022)","DOI":"10.1109\/WACV51458.2022.00207"},{"key":"41_CR2","doi-asserted-by":"crossref","unstructured":"Chen, Y., Schmid, C., Sminchisescu, C.: Self-supervised learning with geometric constraints in monocular video: connecting flow, depth, and camera. In: ICCV, pp. 7063\u20137072 (2019)","DOI":"10.1109\/ICCV.2019.00716"},{"key":"41_CR3","doi-asserted-by":"crossref","unstructured":"Chen, Z., et al.: Revealing the reciprocal relations between self-supervised stereo and monocular depth estimation. In: ICCV, pp. 15529\u201315538 (2021)","DOI":"10.1109\/ICCV48922.2021.01524"},{"key":"41_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"52","DOI":"10.1007\/978-3-030-58577-8_4","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Cheng","year":"2020","unstructured":"Cheng, B., Saggu, I.S., Shah, R., Bansal, G., Bharadia, D.: $$S^3$$Net: semantic-aware self-supervised depth estimation with monocular videos and synthetic data. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12375, pp. 52\u201369. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58577-8_4"},{"issue":"23","key":"41_CR5","doi-asserted-by":"publisher","first-page":"26912","DOI":"10.1109\/JSEN.2021.3120753","volume":"21","author":"Z Cheng","year":"2021","unstructured":"Cheng, Z., Zhang, Y., Tang, C.: Swin-depth: using transformers and multi-scale fusion for monocular-based depth estimation. IEEE Sens. J. 21(23), 26912\u201326920 (2021)","journal-title":"IEEE Sens. J."},{"key":"41_CR6","doi-asserted-by":"crossref","unstructured":"Choi, H., et al.: Adaptive confidence thresholding for monocular depth estimation. In: ICCV, pp. 12808\u201312818 (2021)","DOI":"10.1109\/ICCV48922.2021.01257"},{"key":"41_CR7","unstructured":"Clevert, D.A., Unterthiner, T., Hochreiter, S.: Fast and accurate deep network learning by exponential linear units (ELUs). arXiv preprint arXiv:1511.07289 (2015)"},{"key":"41_CR8","doi-asserted-by":"crossref","unstructured":"Cordts, M., et al.: The cityscapes dataset for semantic urban scene understanding. In: CVPR, pp. 3213\u20133223 (2016)","DOI":"10.1109\/CVPR.2016.350"},{"key":"41_CR9","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: CVPR, pp. 248\u2013255 (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"41_CR10","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: ICLR (2021)"},{"key":"41_CR11","unstructured":"Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. In: Advances in Neural Information Processing Systems, pp. 2366\u20132374 (2014)"},{"key":"41_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-46484-8_45","volume-title":"Computer Vision \u2013 ECCV 2016","author":"R Garg","year":"2016","unstructured":"Garg, R., B.G., V.K., Carneiro, G., Reid, I.: Unsupervised CNN for single view depth estimation: geometry to the rescue. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9912, pp. 740\u2013756. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46484-8_45"},{"key":"41_CR13","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The KITTI vision benchmark suite. In: CVPR, pp. 3354\u20133361 (2012)","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"41_CR14","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac Aodha, O., Brostow, G.J.: Unsupervised monocular depth estimation with left-right consistency. In: CVPR, pp. 270\u2013279 (2017)","DOI":"10.1109\/CVPR.2017.699"},{"key":"41_CR15","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac Aodha, O., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth estimation. In: ICCV, pp. 3828\u20133838 (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"41_CR16","first-page":"12626","volume":"33","author":"JL GonzalezBello","year":"2020","unstructured":"GonzalezBello, J.L., Kim, M.: Forget about the lidar: self-supervised depth estimators with med probability volumes. Adv. Neural. Inf. Process. Syst. 33, 12626\u201312637 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"41_CR17","doi-asserted-by":"crossref","unstructured":"GonzalezBello, J.L., Kim, M.: PLADE-Net: towards pixel-level accuracy for self-supervised single-view depth estimation with neural positional encoding and distilled matting loss. In: CVPR, pp. 6851\u20136860 (2021)","DOI":"10.1109\/CVPR46437.2021.00678"},{"key":"41_CR18","unstructured":"GonzalezBello, J.L., Kim, M.: Self-supervised deep monocular depth estimation with ambiguity boosting. IEEE TPAMI (2021)"},{"issue":"6","key":"41_CR19","doi-asserted-by":"publisher","first-page":"1789","DOI":"10.1007\/s11263-021-01453-z","volume":"129","author":"J Gou","year":"2021","unstructured":"Gou, J., Yu, B., Maybank, S.J., Tao, D.: Knowledge distillation: a survey. IJCV 129(6), 1789\u20131819 (2021)","journal-title":"IJCV"},{"key":"41_CR20","doi-asserted-by":"crossref","unstructured":"Guizilini, V., Ambrus, R., Pillai, S., Raventos, A., Gaidon, A.: 3d packing for self-supervised monocular depth estimation. In: CVPR, pp. 2485\u20132494 (2020)","DOI":"10.1109\/CVPR42600.2020.00256"},{"key":"41_CR21","doi-asserted-by":"crossref","unstructured":"Guizilini, V., Hou, R., Li, J., Ambrus, R., Gaidon, A.: Semantically-guided representation learning for self-supervised monocular depth. In: International Conference on Learning Representations (ICLR) (2020)","DOI":"10.1109\/CVPR42600.2020.00256"},{"key":"41_CR22","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"506","DOI":"10.1007\/978-3-030-01252-6_30","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Guo","year":"2018","unstructured":"Guo, X., Li, H., Yi, S., Ren, J., Wang, X.: Learning monocular depth by distilling cross-domain stereo networks. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11215, pp. 506\u2013523. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01252-6_30"},{"key":"41_CR23","doi-asserted-by":"crossref","unstructured":"Hirschmuller, H.: Accurate and efficient stereo processing by semi-global matching and mutual information. In: CVPR, vol. 2, pp. 807\u2013814. IEEE (2005)","DOI":"10.1109\/CVPR.2005.56"},{"key":"41_CR24","doi-asserted-by":"crossref","unstructured":"Huang, S., Lu, Z., Cheng, R., He, C.: FaPN: feature-aligned pyramid network for dense image prediction. In: ICCV, pp. 864\u2013873 (2021)","DOI":"10.1109\/ICCV48922.2021.00090"},{"issue":"1","key":"41_CR25","first-page":"550","volume":"44","author":"Z Huang","year":"2021","unstructured":"Huang, Z., Wei, Y., Wang, X., Liu, W., Huang, T.S., Shi, H.: AlignSeg: feature-aligned segmentation networks. IEEE TPAMI 44(1), 550\u2013557 (2021)","journal-title":"IEEE TPAMI"},{"key":"41_CR26","unstructured":"Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International Conference on Machine Learning, pp. 448\u2013456 (2015)"},{"key":"41_CR27","doi-asserted-by":"crossref","unstructured":"Ji, P., Li, R., Bhanu, B., Xu, Y.: MonoIndoor: towards good practice of self-supervised monocular depth estimation for indoor environments. In: ICCV, pp. 12787\u201312796 (2021)","DOI":"10.1109\/ICCV48922.2021.01255"},{"key":"41_CR28","doi-asserted-by":"crossref","unstructured":"Jiao, Y., Tran, T.D., Shi, G.: EffiScene: efficient per-pixel rigidity inference for unsupervised joint learning of optical flow, depth, camera pose and motion segmentation. In: CVPR, pp. 5538\u20135547 (2021)","DOI":"10.1109\/CVPR46437.2021.00549"},{"key":"41_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694\u2013711. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43"},{"key":"41_CR30","doi-asserted-by":"crossref","unstructured":"Johnston, A., Carneiro, G.: Self-supervised monocular trained depth estimation using self-attention and discrete disparity volume. In: CVPR, pp. 4756\u20134765 (2020)","DOI":"10.1109\/CVPR42600.2020.00481"},{"key":"41_CR31","doi-asserted-by":"crossref","unstructured":"Jung, H., Park, E., Yoo, S.: Fine-grained semantics-aware representation enhancement for self-supervised monocular depth estimation. In: ICCV, pp. 12642\u201312652 (2021)","DOI":"10.1109\/ICCV48922.2021.01241"},{"key":"41_CR32","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"41_CR33","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"582","DOI":"10.1007\/978-3-030-58565-5_35","volume-title":"Computer Vision \u2013 ECCV 2020","author":"M Klingner","year":"2020","unstructured":"Klingner, M., Term\u00f6hlen, J.-A., Mikolajczyk, J., Fingscheidt, T.: Self-supervised monocular depth estimation: solving the dynamic object problem by semantic guidance. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12365, pp. 582\u2013600. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58565-5_35"},{"key":"41_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"435","DOI":"10.1007\/978-3-030-58520-4_26","volume-title":"Computer Vision \u2013 ECCV 2020","author":"X Li","year":"2020","unstructured":"Li, X., et al.: Improving semantic segmentation via decoupled body and edge supervision. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12362, pp. 435\u2013452. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58520-4_26"},{"key":"41_CR35","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"775","DOI":"10.1007\/978-3-030-58452-8_45","volume-title":"Computer Vision \u2013 ECCV 2020","author":"X Li","year":"2020","unstructured":"Li, X., et al.: Semantic flow for fast and accurate scene parsing. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 775\u2013793. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_45"},{"key":"41_CR36","doi-asserted-by":"crossref","unstructured":"Liu, L., Song, X., Wang, M., Liu, Y., Zhang, L.: Self-supervised monocular depth estimation for all day images using domain separation. In: ICCV, pp. 12737\u201312746 (2021)","DOI":"10.1109\/ICCV48922.2021.01250"},{"key":"41_CR37","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: ICCV, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"41_CR38","doi-asserted-by":"crossref","unstructured":"Mahjourian, R., Wicke, M., Angelova, A.: Unsupervised learning of depth and ego-motion from monocular video using 3d geometric constraints. In: CVPR, pp. 5667\u20135675 (2018)","DOI":"10.1109\/CVPR.2018.00594"},{"key":"41_CR39","first-page":"8026","volume":"32","author":"A Paszke","year":"2019","unstructured":"Paszke, A., et al.: PyTorch: an imperative style, high-performance deep learning library. Adv. Neural. Inf. Process. Syst. 32, 8026\u20138037 (2019)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"41_CR40","doi-asserted-by":"crossref","unstructured":"Peng, R., Wang, R., Lai, Y., Tang, L., Cai, Y.: Excavating the potential capacity of self-supervised monocular depth estimation. In: ICCV, pp. 15560\u201315569 (2021)","DOI":"10.1109\/ICCV48922.2021.01527"},{"key":"41_CR41","doi-asserted-by":"crossref","unstructured":"Pilzer, A., Lathuiliere, S., Sebe, N., Ricci, E.: Refine and distill: exploiting cycle-inconsistency and knowledge distillation for unsupervised monocular depth estimation. In: CVPR, pp. 9768\u20139777 (2019)","DOI":"10.1109\/CVPR.2019.01000"},{"key":"41_CR42","doi-asserted-by":"crossref","unstructured":"Poggi, M., Aleotti, F., Tosi, F., Mattoccia, S.: On the uncertainty of self-supervised monocular depth estimation. In: CVPR, pp. 3227\u20133237 (2020)","DOI":"10.1109\/CVPR42600.2020.00329"},{"key":"41_CR43","doi-asserted-by":"crossref","unstructured":"Ramamonjisoa, M., Firman, M., Watson, J., Lepetit, V., Turmukhambetov, D.: Single image depth prediction with wavelet decomposition. In: CVPR, pp. 11089\u201311098 (2021)","DOI":"10.1109\/CVPR46437.2021.01094"},{"key":"41_CR44","doi-asserted-by":"crossref","unstructured":"Ranftl, R., Bochkovskiy, A., Koltun, V.: Vision transformers for dense prediction. In: ICCV, pp. 12179\u201312188 (2021)","DOI":"10.1109\/ICCV48922.2021.01196"},{"key":"41_CR45","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"572","DOI":"10.1007\/978-3-030-58529-7_34","volume-title":"Computer Vision \u2013 ECCV 2020","author":"C Shu","year":"2020","unstructured":"Shu, C., Yu, K., Duan, Z., Yang, K.: Feature-metric loss for self-supervised learning of depth and egomotion. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12364, pp. 572\u2013588. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58529-7_34"},{"key":"41_CR46","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"41_CR47","doi-asserted-by":"crossref","unstructured":"Tosi, F., Aleotti, F., Poggi, M., Mattoccia, S.: Learning monocular depth estimation infusing traditional stereo knowledge. In: CVPR, pp. 9799\u20139809 (2019)","DOI":"10.1109\/CVPR.2019.01003"},{"key":"41_CR48","doi-asserted-by":"crossref","unstructured":"Uhrig, J., Schneider, N., Schneider, L., Franke, U., Brox, T., Geiger, A.: Sparsity invariant CNNs. In: 2017 International Conference on 3D Vision (3DV), pp. 11\u201320 (2017)","DOI":"10.1109\/3DV.2017.00012"},{"key":"41_CR49","doi-asserted-by":"crossref","unstructured":"Wang, K., et al.: Regularizing nighttime weirdness: efficient self-supervised monocular depth estimation in the dark. In: ICCV, pp. 16055\u201316064 (2021)","DOI":"10.1109\/ICCV48922.2021.01575"},{"key":"41_CR50","doi-asserted-by":"crossref","unstructured":"Wang, L., Wang, Y., Wang, L., Zhan, Y., Wang, Y., Lu, H.: Can scale-consistent monocular depth be learned in a self-supervised scale-invariant manner? In: ICCV, pp. 12727\u201312736 (2021)","DOI":"10.1109\/ICCV48922.2021.01249"},{"key":"41_CR51","doi-asserted-by":"crossref","unstructured":"Wang, W., et al.: Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. In: ICCV, pp. 568\u2013578 (2021)","DOI":"10.1109\/ICCV48922.2021.00061"},{"issue":"4","key":"41_CR52","first-page":"600","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE TIP 13(4), 600\u2013612 (2004)","journal-title":"IEEE TIP"},{"key":"41_CR53","doi-asserted-by":"crossref","unstructured":"Watson, J., Firman, M., Brostow, G.J., Turmukhambetov, D.: Self-supervised monocular depth hints. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00225"},{"key":"41_CR54","doi-asserted-by":"crossref","unstructured":"Yang, G., Tang, H., Ding, M., Sebe, N., Ricci, E.: Transformer-based attention networks for continuous pixel-wise prediction. In: ICCV, pp. 16269\u201316279 (2021)","DOI":"10.1109\/ICCV48922.2021.01596"},{"key":"41_CR55","doi-asserted-by":"crossref","unstructured":"Yin, Z., Shi, J.: GeoNet: unsupervised learning of dense depth, optical flow and camera pose. In: CVPR, pp. 1983\u20131992 (2018)","DOI":"10.1109\/CVPR.2018.00212"},{"key":"41_CR56","doi-asserted-by":"crossref","unstructured":"Zhou, J., Wang, Y., Qin, K., Zeng, W.: Moving indoor: unsupervised video depth learning in challenging environments. In: ICCV, pp. 8618\u20138627 (2019)","DOI":"10.1109\/ICCV.2019.00871"},{"key":"41_CR57","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.G.: Unsupervised learning of depth and ego-motion from video. In: CVPR, pp. 1851\u20131858 (2017)","DOI":"10.1109\/CVPR.2017.700"},{"key":"41_CR58","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Fan, X., Shi, P., Xin, Y.: R-MSFM: recurrent multi-scale feature modulation for monocular depth estimating. In: ICCV, pp. 12777\u201312786 (2021)","DOI":"10.1109\/ICCV48922.2021.01254"},{"key":"41_CR59","doi-asserted-by":"crossref","unstructured":"Zhu, S., Brazil, G., Liu, X.: The edge of depth: explicit constraints between segmentation and depth. In: CVPR, pp. 13116\u201313125 (2020)","DOI":"10.1109\/CVPR42600.2020.01313"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19769-7_41","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T10:09:17Z","timestamp":1728209357000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19769-7_41"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197680","9783031197697"],"references-count":59,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19769-7_41","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"23 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}