{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T14:42:51Z","timestamp":1775227371295,"version":"3.50.1"},"publisher-location":"Cham","reference-count":42,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031923869","type":"print"},{"value":"9783031923876","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-92387-6_5","type":"book-chapter","created":{"date-parts":[[2025,5,25]],"date-time":"2025-05-25T18:42:18Z","timestamp":1748198538000},"page":"71-86","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["Depth Any Canopy: Leveraging Depth Foundation Models for\u00a0Canopy Height Estimation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5067-2118","authenticated-orcid":false,"given":"Daniele","family":"Rege Cambrin","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9273-7303","authenticated-orcid":false,"given":"Isaac","family":"Corley","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1263-7522","authenticated-orcid":false,"given":"Paolo","family":"Garza","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,12]]},"reference":[{"key":"5_CR1","unstructured":"Andersen, H.E., Reutebuch, S.E., Schreuder, G.F.: Automated individual tree measurement through morphological analysis of a lidar-based canopy surface model. In: Proceedings of of the 1st International Precision Forestry Symposium, pp. 11\u201321 (2001)"},{"key":"5_CR2","doi-asserted-by":"publisher","first-page":"269","DOI":"10.1016\/j.isprsjprs.2022.11.011","volume":"195","author":"A Becker","year":"2023","unstructured":"Becker, A., Russo, S., Puliti, S., Lang, N., Schindler, K., Wegner, J.D.: Country-wide retrieval of forest structure from optical and sar satellite imagery with deep ensembles. ISPRS J. Photogramm. Remote. Sens. 195, 269\u2013286 (2023)","journal-title":"ISPRS J. Photogramm. Remote. Sens."},{"key":"5_CR3","unstructured":"Birkl, R., Wofk, D., M\u00fcller, M.: Midas v3. 1\u2013a model zoo for robust monocular relative depth estimation. arXiv preprint arXiv:2307.14460 (2023)"},{"key":"5_CR4","doi-asserted-by":"publisher","unstructured":"Cha, K., Seo, J., Lee, T.: A billion-scale foundation model for remote sensing images. IEEE J. Selected Topics Appl. Earth Observat. Remote Sens., 1\u201317 (2024). https:\/\/doi.org\/10.1109\/JSTARS.2024.3401772","DOI":"10.1109\/JSTARS.2024.3401772"},{"key":"5_CR5","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale (2021). https:\/\/arxiv.org\/abs\/2010.11929"},{"issue":"9","key":"5_CR6","doi-asserted-by":"publisher","DOI":"10.1088\/1748-9326\/ac8694","volume":"17","author":"R Dubayah","year":"2022","unstructured":"Dubayah, R., et al.: Gedi launches a new era of biomass inference from space. Environ. Res. Lett. 17(9), 095001 (2022)","journal-title":"Environ. Res. Lett."},{"key":"5_CR7","unstructured":"Fogel, F., et\u00a0al.: Open-canopy: a country-scale benchmark for canopy height estimation at very high resolution. arXiv preprint arXiv:2407.09392 (2024)"},{"key":"5_CR8","unstructured":"Fogel, F., et al.: Open-canopy: a country-scale benchmark for canopy height estimation at very high resolution (2024). https:\/\/arxiv.org\/abs\/2407.09392"},{"issue":"5","key":"5_CR9","doi-asserted-by":"publisher","first-page":"650","DOI":"10.5589\/m03-023","volume":"29","author":"DL Gaveau","year":"2003","unstructured":"Gaveau, D.L., Hill, R.A.: Quantifying canopy height underestimation by laser pulse penetration in small-footprint airborne laser scanning data. Can. J. Remote. Sens. 29(5), 650\u2013657 (2003)","journal-title":"Can. J. Remote. Sens."},{"issue":"1","key":"5_CR10","doi-asserted-by":"publisher","DOI":"10.1117\/1.3361375","volume":"4","author":"TU Kampe","year":"2010","unstructured":"Kampe, T.U., Johnson, B.R., Kuester, M.A., Keller, M.: Neon: the first continental-scale ecological observatory with airborne remote sensing of vegetation canopy biochemistry and structure. J. Appl. Remote Sens. 4(1), 043510 (2010)","journal-title":"J. Appl. Remote Sens."},{"key":"5_CR11","doi-asserted-by":"crossref","unstructured":"Ke, B., Obukhov, A., Huang, S., Metzger, N., Daudt, R.C., Schindler, K.: Repurposing diffusion-based image generators for monocular depth estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9492\u20139502 (2024)","DOI":"10.1109\/CVPR52733.2024.00907"},{"key":"5_CR12","unstructured":"Khanna, S., et al.: Diffusionsat: a generative foundation model for satellite imagery. In: The Twelfth International Conference on Learning Representations (2023)"},{"key":"5_CR13","doi-asserted-by":"crossref","unstructured":"Kirillov, A., et\u00a0al.: Segment anything. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4015\u20134026 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"5_CR14","doi-asserted-by":"crossref","unstructured":"Kuckreja, K., Danish, M.S., Naseer, M., Das, A., Khan, S., Khan, F.S.: Geochat: grounded large vision-language model for remote sensing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 27831\u201327840 (2024)","DOI":"10.1109\/CVPR52733.2024.02629"},{"key":"5_CR15","unstructured":"Lacoste, A., Luccioni, A., Schmidt, V., Dandres, T.: Quantifying the carbon emissions of machine learning. arXiv preprint arXiv:1910.09700 (2019)"},{"issue":"11","key":"5_CR16","doi-asserted-by":"publisher","first-page":"1778","DOI":"10.1038\/s41559-023-02206-6","volume":"7","author":"N Lang","year":"2023","unstructured":"Lang, N., Jetz, W., Schindler, K., Wegner, J.D.: A high-resolution canopy height model of the earth. Nat. Ecol. Evol. 7(11), 1778\u20131789 (2023)","journal-title":"Nat. Ecol. Evol."},{"key":"5_CR17","unstructured":"Lasinger, K., Ranftl, R., Schindler, K., Koltun, V.: Towards robust monocular depth estimation: mixing datasets for zero-shot cross-dataset transfer. arXiv preprint arXiv:1907.01341 (2019)"},{"key":"5_CR18","doi-asserted-by":"crossref","unstructured":"Li, Z., Snavely, N.: Megadepth: learning single-view depth prediction from internet photos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2041\u20132050 (2018)","DOI":"10.1109\/CVPR.2018.00218"},{"key":"5_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2023.3331717","volume":"61","author":"J Lin","year":"2023","unstructured":"Lin, J., Gao, F., Shi, X., Dong, J., Du, Q.: Ss-mae: spatial-spectral masked autoencoder for multisource remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 61, 1\u201314 (2023). https:\/\/doi.org\/10.1109\/TGRS.2023.3331717","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"5_CR20","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"5_CR21","doi-asserted-by":"publisher","unstructured":"Minderer, M., et\u00a0al.: Simple open-vocabulary object detection. In: European Conference on Computer Vision, pp. 728\u2013755. Springer, Heidelberg (2022). https:\/\/doi.org\/10.1007\/978-3-031-20080-9_42","DOI":"10.1007\/978-3-031-20080-9_42"},{"key":"5_CR22","unstructured":"Oquab, M., et al.: Dinov2: learning robust visual features without supervision (2024). https:\/\/arxiv.org\/abs\/2304.07193"},{"key":"5_CR23","unstructured":"Pauls, J., et al.: Estimating canopy height at scale. arXiv preprint arXiv:2406.01076 (2024)"},{"key":"5_CR24","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"5_CR25","doi-asserted-by":"crossref","unstructured":"Ranftl, R., Bochkovskiy, A., Koltun, V.: Vision transformers for dense prediction. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12179\u201312188 (2021)","DOI":"10.1109\/ICCV48922.2021.01196"},{"issue":"5","key":"5_CR26","doi-asserted-by":"publisher","first-page":"1475","DOI":"10.1080\/01431160701736380","volume":"29","author":"J Rosette","year":"2008","unstructured":"Rosette, J., North, P., Suarez, J.: Vegetation height estimates for a mixed temperate forest using satellite laser altimetry. Int. J. Remote Sens. 29(5), 1475\u20131493 (2008)","journal-title":"Int. J. Remote Sens."},{"key":"5_CR27","unstructured":"Satellogic: Earthview. https:\/\/huggingface.co\/datasets\/satellogic\/EarthView. Accessed 06 July 2024"},{"key":"5_CR28","doi-asserted-by":"crossref","unstructured":"Schops, T., et al.: A multi-view stereo benchmark with high-resolution images and multi-camera videos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3260\u20133269 (2017)","DOI":"10.1109\/CVPR.2017.272"},{"issue":"3","key":"5_CR29","doi-asserted-by":"publisher","first-page":"1136","DOI":"10.1016\/j.foreco.2008.11.022","volume":"257","author":"JO Sexton","year":"2009","unstructured":"Sexton, J.O., Bax, T., Siqueira, P., Swenson, J.J., Hensley, S.: A comparison of lidar, radar, and field measurements of canopy height in pine and hardwood forests of southeastern north america. For. Ecol. Manag. 257(3), 1136\u20131147 (2009)","journal-title":"For. Ecol. Manag."},{"key":"5_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"746","DOI":"10.1007\/978-3-642-33715-4_54","volume-title":"Computer Vision \u2013 ECCV 2012","author":"N Silberman","year":"2012","unstructured":"Silberman, N., Hoiem, D., Kohli, P., Fergus, R.: Indoor segmentation and support inference from RGBD images. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7576, pp. 746\u2013760. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33715-4_54"},{"key":"5_CR31","doi-asserted-by":"publisher","unstructured":"Simard, M., Pinto, N., Fisher, J.B., Baccini, A.: Mapping forest canopy height globally with spaceborne lidar. J. Geophys. Res. 116(G4) (2011). https:\/\/doi.org\/10.1029\/2011jg001708","DOI":"10.1029\/2011jg001708"},{"key":"5_CR32","doi-asserted-by":"publisher","DOI":"10.1016\/j.rse.2023.113888","volume":"300","author":"J Tolan","year":"2024","unstructured":"Tolan, J., et al.: Very high resolution canopy height maps from rgb imagery using self-supervised vision transformer and convolutional decoder trained on aerial lidar. Remote Sens. Environ. 300, 113888 (2024)","journal-title":"Remote Sens. Environ."},{"key":"5_CR33","doi-asserted-by":"crossref","unstructured":"Tomppo, E., et\u00a0al.: National forest inventories. Pathways Common Report. Eur. Sci. Found. 1, 541\u2013553 (2010)","DOI":"10.1007\/978-90-481-3233-1"},{"key":"5_CR34","unstructured":"Vasiljevic, I., et\u00a0al.: Diode: a dense indoor and outdoor depth dataset. arXiv preprint arXiv:1908.00463 (2019)"},{"key":"5_CR35","doi-asserted-by":"publisher","unstructured":"Wang, S., Liu, C., Li, W., Jia, S., Yue, H.: Hybrid model for estimating forest canopy heights using fused multimodal spaceborne lidar data and optical imagery. Int. J. Appl. Earth Obs. Geoinf. 122, 103431 (2023). https:\/\/doi.org\/10.1016\/j.jag.2023.103431","DOI":"10.1016\/j.jag.2023.103431"},{"key":"5_CR36","unstructured":"Watch, G.F.: Global forest watch. World Resources Institute, Washington, DC (2002).https:\/\/globalforestwatch.org\/. Accessed Mar 2002"},{"key":"5_CR37","unstructured":"Wu, H., et al.: Q-align: teaching lmms for visual scoring via discrete text-defined levels (2023). https:\/\/arxiv.org\/abs\/2312.17090"},{"key":"5_CR38","doi-asserted-by":"crossref","unstructured":"Yang, L., Kang, B., Huang, Z., Xu, X., Feng, J., Zhao, H.: Depth anything: unleashing the power of large-scale unlabeled data. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10371\u201310381 (2024)","DOI":"10.1109\/CVPR52733.2024.00987"},{"key":"5_CR39","unstructured":"Yang, L., et al.: Depth anything v2 (2024). https:\/\/arxiv.org\/abs\/2406.09414"},{"key":"5_CR40","doi-asserted-by":"publisher","unstructured":"Zang, J., et al.: Field-measured canopy height may not be as accurate and heritable as believed: evidence from advanced 3d sensing. Plant Methods 19(1) (2023). https:\/\/doi.org\/10.1186\/s13007-023-01012-2","DOI":"10.1186\/s13007-023-01012-2"},{"key":"5_CR41","unstructured":"Zheng, Z., Zhong, Y., Zhang, L., Ermon, S.: Segment any change. arXiv preprint arXiv:2402.01188 (2024)"},{"key":"5_CR42","doi-asserted-by":"crossref","unstructured":"Zhou, C., et al.: A comprehensive survey on pretrained foundation models: a history from bert to chatgpt (2023)","DOI":"10.1007\/s13042-024-02443-6"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-92387-6_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,25]],"date-time":"2025-05-25T18:42:27Z","timestamp":1748198547000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-92387-6_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031923869","9783031923876"],"references-count":42,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-92387-6_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"12 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}