{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,24]],"date-time":"2025-06-24T04:02:40Z","timestamp":1750737760490,"version":"3.41.0"},"publisher-location":"Singapore","reference-count":20,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819665785","type":"print"},{"value":"9789819665792","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-6579-2_6","type":"book-chapter","created":{"date-parts":[[2025,6,23]],"date-time":"2025-06-23T12:06:35Z","timestamp":1750680395000},"page":"74-88","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["USAM-Net: A U-Net-Based Network for\u00a0Improved Stereo Correspondence and\u00a0Scene Depth Estimation Using Features from\u00a0a\u00a0Pre-trained Image Segmentation Network"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-6943-2003","authenticated-orcid":false,"given":"Joseph Emmanuel DL","family":"Dayo","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7140-1707","authenticated-orcid":false,"given":"Prospero C.","family":"Naval","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,24]]},"reference":[{"key":"6_CR1","doi-asserted-by":"crossref","unstructured":"Cantrell, K.J., Miller, C.D., Morato, C.W.: Practical depth estimation with image segmentation and serial u-nets. In: 6th International Conference on Vehicle Technology and Intelligent Transport Systems, pp. 406\u2013414 (2020)","DOI":"10.5220\/0009781804060414"},{"key":"6_CR2","unstructured":"Dayo, J.E.: GitHub - jedld\/OpenStereo: Fork of OpenStereo: A Comprehensive Benchmark for Stereo Matching and Strong Baseline \u2014 github.com. https:\/\/github.com\/jedld\/OpenStereo. Accessed 29 Sept 2024"},{"key":"6_CR3","unstructured":"Guo, X., et al.: Openstereo: A comprehensive benchmark for stereo matching and strong baseline. arXiv preprint arXiv:2312.00343 (2023)"},{"key":"6_CR4","doi-asserted-by":"crossref","unstructured":"Huang, B., Zheng, J.Q., Giannarou, S., Elson, D.: H-net: Unsupervised attention-based stereo depth estimation leveraging epipolar geometry (2022)","DOI":"10.1109\/CVPRW56347.2022.00492"},{"key":"6_CR5","doi-asserted-by":"publisher","unstructured":"Jan, A., Seo, S.: Monocular depth estimation using res-UNet with an attention model. Appl. Sci. 13(10), 6319 (2023). https:\/\/doi.org\/10.3390\/app13106319","DOI":"10.3390\/app13106319"},{"key":"6_CR6","doi-asserted-by":"crossref","unstructured":"Kirillov, A., et al.: Segment anything. arXiv:2304.02643 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"6_CR7","doi-asserted-by":"publisher","unstructured":"Liang, Z., et al.: Learning for disparity estimation through feature constancy. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2811\u20132820 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00297","DOI":"10.1109\/CVPR.2018.00297"},{"key":"6_CR8","doi-asserted-by":"crossref","unstructured":"Mayer, N., et al.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: IEEE International Conference on Computer Vision and Pattern Recognition (CVPR) (2016). http:\/\/lmb.informatik.uni-freiburg.de\/Publications\/2016\/MIFDB16, arXiv:1512.02134","DOI":"10.1109\/CVPR.2016.438"},{"key":"6_CR9","doi-asserted-by":"crossref","unstructured":"Menze, M., Geiger, A.: Object scene flow for autonomous vehicles. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2015)","DOI":"10.1109\/CVPR.2015.7298925"},{"key":"6_CR10","doi-asserted-by":"crossref","unstructured":"Menze, M., Heipke, C., Geiger, A.: Joint 3D estimation of vehicles and scene flow. In: ISPRS Workshop on Image Sequence Analysis (ISA) (2015)","DOI":"10.5194\/isprsannals-II-3-W5-427-2015"},{"key":"6_CR11","doi-asserted-by":"crossref","unstructured":"Silberman, N., Hoiem, D., Kohli, P., Fergus, R.: Indoor segmentation and support inference from RGBD images. In: ECCV (2012)","DOI":"10.1007\/978-3-642-33715-4_54"},{"key":"6_CR12","doi-asserted-by":"publisher","unstructured":"Oktay, O., et al.: Attention U-Net: Learning where to look for the pancreas (04 2018). https:\/\/doi.org\/10.48550\/arXiv.1804.03999","DOI":"10.48550\/arXiv.1804.03999"},{"key":"6_CR13","doi-asserted-by":"publisher","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: Convolutional networks for biomedical image segmentation, vol.\u00a09351, pp. 234\u2013241 (10 2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"6_CR14","doi-asserted-by":"publisher","unstructured":"Scharstein, D., et al.: High-resolution stereo datasets with subpixel-accurate ground truth. vol.\u00a08753, pp. 31\u201342 (09 2014). https:\/\/doi.org\/10.1007\/978-3-319-11752-2_3","DOI":"10.1007\/978-3-319-11752-2_3"},{"key":"6_CR15","doi-asserted-by":"crossref","unstructured":"Shen, Z., Dai, Y., Rao, Z.: CFNet: Cascade and fused cost volume for robust stereo matching (04 2021)","DOI":"10.1109\/CVPR46437.2021.01369"},{"issue":"4","key":"6_CR16","doi-asserted-by":"publisher","first-page":"910","DOI":"10.1007\/s11263-019-01287-w","volume":"128","author":"X Song","year":"2020","unstructured":"Song, X., Zhao, X., Fang, L., Hu, H., Yu, Y.: EdgeStereo: an effective multi-task learning network for stereo matching and edge detection. Int. J. Comput. Vision 128(4), 910\u2013930 (2020). https:\/\/doi.org\/10.1007\/s11263-019-01287-w","journal-title":"Int. J. Comput. Vision"},{"key":"6_CR17","unstructured":"Victor, V.S., Neigel, P.: Survey on semantic stereo matching\/semantic depth estimation. arXiv preprint arXiv:2109.10123 (2021)"},{"key":"6_CR18","doi-asserted-by":"crossref","unstructured":"Xu, G., Wang, X., Ding, X., Yang, X.: Iterative geometry encoding volume for stereo matching (2023). https:\/\/arxiv.org\/abs\/2303.06615","DOI":"10.1109\/CVPR52729.2023.02099"},{"key":"6_CR19","doi-asserted-by":"crossref","unstructured":"Yang, G., Song, X., Huang, C., Deng, Z., Shi, J., Zhou, B.: Drivingstereo: a large-scale dataset for stereo matching in autonomous driving scenarios. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.00099"},{"key":"6_CR20","doi-asserted-by":"publisher","first-page":"660","DOI":"10.1007\/978-3-030-01234-2_39","volume-title":"Computer Vision \u2013 ECCV 2018: 15th European Conference, Munich, Germany, September 8\u201314, 2018, Proceedings, Part VII","author":"G Yang","year":"2018","unstructured":"Yang, G., Zhao, H., Shi, J., Deng, Z., Jia, J.: SegStereo: exploiting semantic information for disparity estimation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision \u2013 ECCV 2018: 15th European Conference, Munich, Germany, September 8\u201314, 2018, Proceedings, Part VII, pp. 660\u2013676. Springer International Publishing, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_39"}],"container-title":["Lecture Notes in Computer Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-6579-2_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,23]],"date-time":"2025-06-23T12:06:39Z","timestamp":1750680399000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-6579-2_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819665785","9789819665792"],"references-count":20,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-6579-2_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"24 June 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Auckland","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Zealand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/iconip2024.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}