{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T17:10:36Z","timestamp":1777655436406,"version":"3.51.4"},"publisher-location":"Cham","reference-count":26,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031730207","type":"print"},{"value":"9783031730214","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T00:00:00Z","timestamp":1732147200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T00:00:00Z","timestamp":1732147200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73021-4_4","type":"book-chapter","created":{"date-parts":[[2024,11,20]],"date-time":"2024-11-20T09:15:51Z","timestamp":1732094151000},"page":"54-69","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["GroCo: Ground Constraint for\u00a0Metric Self-supervised Monocular Depth"],"prefix":"10.1007","author":[{"given":"Aur\u00e9lien","family":"Cecille","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0374-3814","authenticated-orcid":false,"given":"Stefan","family":"Duffner","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8587-6997","authenticated-orcid":false,"given":"Franck","family":"Davoine","sequence":"additional","affiliation":[]},{"given":"Thibault","family":"Neveu","sequence":"additional","affiliation":[]},{"given":"R\u00e9mi","family":"Agier","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,21]]},"reference":[{"key":"4_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40064-016-3573-7","volume":"5","author":"MO Aqel","year":"2016","unstructured":"Aqel, M.O., Marhaban, M.H., Saripan, M.I., Ismail, N.B.: Review of visual odometry: types, approaches, challenges, and applications. Springerplus 5, 1\u201326 (2016)","journal-title":"Springerplus"},{"key":"4_CR2","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"4_CR3","unstructured":"Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. In: Advances in Neural Information Processing Systems, vol. 27 (2014)"},{"issue":"11","key":"4_CR4","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: the KITTI dataset. Int. J. Robot. Res. 32(11), 1231\u20131237 (2013)","journal-title":"Int. J. Robot. Res."},{"key":"4_CR5","doi-asserted-by":"publisher","unstructured":"Godard, C., Aodha, O.M., Firman, M., Brostow, G.: Digging into self-supervised monocular depth estimation. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 3827\u20133837. IEEE (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00393, https:\/\/ieeexplore.ieee.org\/document\/9009796\/","DOI":"10.1109\/ICCV.2019.00393"},{"key":"4_CR6","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac\u00a0Aodha, O., Brostow, G.J.: Unsupervised monocular depth estimation with left-right consistency. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 270\u2013279 (2017)","DOI":"10.1109\/CVPR.2017.699"},{"key":"4_CR7","doi-asserted-by":"crossref","unstructured":"Guizilini, V., Ambrus, R., Burgard, W., Gaidon, A.: Sparse auxiliary networks for unified monocular depth prediction and completion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11078\u201311088 (2021)","DOI":"10.1109\/CVPR46437.2021.01093"},{"key":"4_CR8","doi-asserted-by":"publisher","unstructured":"Guizilini, V., Ambrus, R., Pillai, S., Raventos, A., Gaidon, A.: 3D packing for self-supervised monocular depth estimation. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 2482\u20132491. IEEE (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00256, https:\/\/ieeexplore.ieee.org\/document\/9156708\/","DOI":"10.1109\/CVPR42600.2020.00256"},{"key":"4_CR9","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"4_CR10","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"4_CR11","doi-asserted-by":"publisher","unstructured":"Kinoshita, G., Nishino, K.: Camera height doesn\u2019t change: unsupervised monocular scale-aware road-scene depth estimation. https:\/\/doi.org\/10.48550\/arXiv.2312.04530, http:\/\/arxiv.org\/abs\/2312.04530","DOI":"10.48550\/arXiv.2312.04530"},{"key":"4_CR12","unstructured":"Koledi\u0107, K., Petrovi\u0107, L., Petrovi\u0107, I., Markovi\u0107, I.: GenDepth: generalizing monocular depth estimation for arbitrary camera parameters via ground plane embedding, http:\/\/arxiv.org\/abs\/2312.06021"},{"key":"4_CR13","unstructured":"Lee, J.H., Han, M.K., Ko, D.W., Suh, I.H.: From big to small: multi-scale local planar guidance for monocular depth estimation. arXiv preprint arXiv:1907.10326 (2019)"},{"key":"4_CR14","doi-asserted-by":"crossref","unstructured":"Li, Z., Chen, Z., Liu, X., Jiang, J.: DepthFormer: exploiting long-range correlation and local information for accurate monocular depth estimation. arXiv preprint arXiv:2203.14211 (2022)","DOI":"10.1007\/s11633-023-1458-0"},{"key":"4_CR15","doi-asserted-by":"publisher","unstructured":"Lyu, X., et al.: HR-depth: high resolution self-supervised monocular depth estimation. 35(3), 2294\u20132301 (2021). https:\/\/doi.org\/10.1609\/aaai.v35i3.16329, https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/16329","DOI":"10.1609\/aaai.v35i3.16329"},{"key":"4_CR16","doi-asserted-by":"crossref","unstructured":"Uhrig, J., Schneider, N., Schneider, L., Franke, U., Brox, T., Geiger, A.: Sparsity invariant CNNs. In: International Conference on 3D Vision (3DV) (2017)","DOI":"10.1109\/3DV.2017.00012"},{"key":"4_CR17","doi-asserted-by":"publisher","unstructured":"Van\u00a0Dijk, T., De\u00a0Croon, G.: How do neural networks see depth in single images? In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 2183\u20132191. IEEE (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00227, https:\/\/ieeexplore.ieee.org\/document\/9009532\/","DOI":"10.1109\/ICCV.2019.00227"},{"key":"4_CR18","doi-asserted-by":"publisher","unstructured":"Vijayanarasimhan, S., Ricco, S., Schmid, C., Sukthankar, R., Fragkiadaki, K.: SfM-net: learning of structure and motion from video (2017). https:\/\/doi.org\/10.48550\/arXiv.1704.07804, http:\/\/arxiv.org\/abs\/1704.07804","DOI":"10.48550\/arXiv.1704.07804"},{"key":"4_CR19","doi-asserted-by":"publisher","unstructured":"Wagstaff, B., Kelly, J.: Self-supervised scale recovery for monocular depth and egomotion estimation. In: 2021 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2620\u20132627 (2021). https:\/\/doi.org\/10.1109\/IROS51168.2021.9635938, https:\/\/ieeexplore.ieee.org\/document\/9635938. ISSN 2153-0866","DOI":"10.1109\/IROS51168.2021.9635938"},{"key":"4_CR20","doi-asserted-by":"publisher","unstructured":"Xiang, J., Wang, Y., An, L., Liu, H., Wang, Z., Liu, J.: Visual attention-based self-supervised absolute depth estimation using geometric priors in autonomous driving. IEEE Robot. Autom. Lett. 7(4), 11998\u201312005. https:\/\/doi.org\/10.1109\/LRA.2022.3210298, https:\/\/ieeexplore.ieee.org\/abstract\/document\/9904826","DOI":"10.1109\/LRA.2022.3210298"},{"key":"4_CR21","doi-asserted-by":"crossref","unstructured":"Xue, F., Zhuo, G., Huang, Z., Fu, W., Wu, Z., Ang, M.H.: Toward hierarchical self-supervised monocular absolute depth estimation for autonomous driving applications. In: 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2330\u20132337. IEEE (2020)","DOI":"10.1109\/IROS45743.2020.9340802"},{"key":"4_CR22","doi-asserted-by":"publisher","unstructured":"Yang, X., Ma, Z., Ji, Z., Ren, Z.: GEDepth: ground embedding for monocular depth estimation. In: 2023 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 12673\u201312681. IEEE. https:\/\/doi.org\/10.1109\/ICCV51070.2023.01168, https:\/\/ieeexplore.ieee.org\/document\/10378086\/","DOI":"10.1109\/ICCV51070.2023.01168"},{"key":"4_CR23","doi-asserted-by":"crossref","unstructured":"Zhang, N., Nex, F., Vosselman, G., Kerle, N.: Lite-mono: a lightweight CNN and transformer architecture for self-supervised monocular depth estimation, pp. 18537\u201318546 (2023). https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/html\/Zhang_Lite-Mono_A_Lightweight_CNN_and_Transformer_Architecture_for_Self-Supervised_Monocular_CVPR_2023_paper.html","DOI":"10.1109\/CVPR52729.2023.01778"},{"key":"4_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1007\/978-3-031-19839-7_9","volume-title":"Computer Vision - ECCV 2022","author":"S Zhang","year":"2022","unstructured":"Zhang, S., Zhang, J., Tao, D.: Towards scale-aware, robust, and generalizable unsupervised monocular depth estimation by integrating IMU motion dynamics. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13698, pp. 143\u2013160. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19839-7_9"},{"key":"4_CR25","doi-asserted-by":"publisher","unstructured":"Zhao, C., et al.: MonoViT: self-supervised monocular depth estimation with a vision transformer. In: 2022 International Conference on 3D Vision (3DV), pp. 668\u2013678 (2022). https:\/\/doi.org\/10.1109\/3DV57658.2022.00077, https:\/\/ieeexplore.ieee.org\/abstract\/document\/10044409. ISSN 2475-7888","DOI":"10.1109\/3DV57658.2022.00077"},{"key":"4_CR26","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.G.: Unsupervised learning of depth and ego-motion from video, pp. 1851\u20131858 (2017). https:\/\/openaccess.thecvf.com\/content_cvpr_2017\/html\/Zhou_Unsupervised_Learning_of_CVPR_2017_paper.html","DOI":"10.1109\/CVPR.2017.700"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73021-4_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,20]],"date-time":"2024-11-20T09:40:53Z","timestamp":1732095653000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73021-4_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,21]]},"ISBN":["9783031730207","9783031730214"],"references-count":26,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73021-4_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,21]]},"assertion":[{"value":"21 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}