{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,4]],"date-time":"2025-07-04T05:39:09Z","timestamp":1751607549308,"version":"3.40.3"},"publisher-location":"Cham","reference-count":43,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031250712"},{"type":"electronic","value":"9783031250729"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25072-9_11","type":"book-chapter","created":{"date-parts":[[2023,2,17]],"date-time":"2023-02-17T08:40:04Z","timestamp":1676623204000},"page":"161-174","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Learning an\u00a0Efficient Multimodal Depth Completion Model"],"prefix":"10.1007","author":[{"given":"Dewang","family":"Hou","sequence":"first","affiliation":[]},{"given":"Yuanyuan","family":"Du","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,18]]},"reference":[{"unstructured":"Bachlechner, T., Majumder, B.P., Mao, H., Cottrell, G., McAuley, J.: Rezero is all you need: fast convergence at large depth. In: Uncertainty in Artificial Intelligence, pp. 1352\u20131361. PMLR (2021)","key":"11_CR1"},{"doi-asserted-by":"crossref","unstructured":"Chen, Y., Yang, B., Liang, M., Urtasun, R.: Learning joint 2D\u20133D representations for depth completion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 10023\u201310032 (2019)","key":"11_CR2","DOI":"10.1109\/ICCV.2019.01012"},{"doi-asserted-by":"crossref","unstructured":"Cheng, X., Wang, P., Guan, C., Yang, R.: CSPN++: learning context and resource aware convolutional spatial propagation networks for depth completion. In: Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), vol. 34, pp. 10615\u201310622 (2020)","key":"11_CR3","DOI":"10.1609\/aaai.v34i07.6635"},{"key":"11_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1007\/978-3-030-01270-0_7","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Cheng","year":"2018","unstructured":"Cheng, X., Wang, P., Yang, R.: Depth estimation via affinity learned with convolutional spatial propagation network. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11220, pp. 108\u2013125. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01270-0_7"},{"issue":"10","key":"11_CR5","doi-asserted-by":"publisher","first-page":"2361","DOI":"10.1109\/TPAMI.2019.2947374","volume":"42","author":"X Cheng","year":"2019","unstructured":"Cheng, X., Wang, P., Yang, R.: Learning depth with convolutional spatial propagation network. IEEE Trans. Pattern Anal. Mach. Intell. 42(10), 2361\u20132379 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"11_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"450","DOI":"10.1007\/978-3-030-01449-0_38","volume-title":"Advanced Concepts for Intelligent Vision Systems","author":"M Dimitrievski","year":"2018","unstructured":"Dimitrievski, M., Veelaert, P., Philips, W.: Learning morphological operators for depth completion. In: Blanc-Talon, J., Helbert, D., Philips, W., Popescu, D., Scheunders, P. (eds.) ACIVS 2018. LNCS, vol. 11182, pp. 450\u2013461. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01449-0_38"},{"doi-asserted-by":"crossref","unstructured":"Eldesokey, A., Felsberg, M., Holmquist, K., Persson, M.: Uncertainty-aware CNNs for depth completion: uncertainty from beginning to end. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12014\u201312023 (2020)","key":"11_CR7","DOI":"10.1109\/CVPR42600.2020.01203"},{"unstructured":"Eldesokey, A., Felsberg, M., Khan, F.S.: Propagating confidences through CNNs for sparse data regression. arXiv preprint arXiv:1805.11913 (2018)","key":"11_CR8"},{"doi-asserted-by":"crossref","unstructured":"Fu, C., Dong, C., Mertz, C., Dolan, J.M.: Depth completion via inductive fusion of planar LiDAR and monocular camera. In: IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 10843\u201310848 (2020)","key":"11_CR9","DOI":"10.1109\/IROS45743.2020.9341385"},{"unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: Proceedings on the International Conference on Neural Information Processing Systems (NIPS), vol. 27 (2014)","key":"11_CR10"},{"issue":"2","key":"11_CR11","doi-asserted-by":"publisher","first-page":"1808","DOI":"10.1109\/LRA.2021.3060396","volume":"6","author":"J Gu","year":"2021","unstructured":"Gu, J., Xiang, Z., Ye, Y., Wang, L.: DenseLiDAR: a real-time pseudo dense depth guided depth completion network. IEEE Robot. Autom. Lett. 6(2), 1808\u20131815 (2021)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"11_CR12","doi-asserted-by":"publisher","first-page":"806","DOI":"10.1109\/TCI.2020.2981761","volume":"6","author":"P Hambarde","year":"2020","unstructured":"Hambarde, P., Murala, S.: S2DNet: depth estimation from single image and sparse samples. IEEE Trans. Comput. Imaging 6, 806\u2013817 (2020)","journal-title":"IEEE Trans. Comput. Imaging"},{"unstructured":"Hu, J., et al.: Deep depth completion: a survey. arXiv preprint arXiv:2205.05335 (2022)","key":"11_CR13"},{"doi-asserted-by":"crossref","unstructured":"Hu, M., Wang, S., Li, B., Ning, S., Fan, L., Gong, X.: PENet: towards precise and efficient image guided depth completion. In: IEEE International Conference on Robotics and Automation (ICRA), pp. 13656\u201313662 (2021)","key":"11_CR14","DOI":"10.1109\/ICRA48506.2021.9561035"},{"doi-asserted-by":"crossref","unstructured":"Imran, S., Long, Y., Liu, X., Morris, D.: Depth coefficients for depth completion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12438\u201312447 (2019)","key":"11_CR15","DOI":"10.1109\/CVPR.2019.01273"},{"doi-asserted-by":"crossref","unstructured":"Jaritz, M., De Charette, R., Wirbel, E., Perrotton, X., Nashashibi, F.: Sparse and dense data with CNNs: depth completion and semantic segmentation. In: IEEE International Conference on 3D Vision (3DV), pp. 52\u201360 (2018)","key":"11_CR16","DOI":"10.1109\/3DV.2018.00017"},{"doi-asserted-by":"crossref","unstructured":"Kwon, H., Tai, Y.W., Lin, S.: Data-driven depth map refinement via multi-scale sparse representation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 159\u2013167 (2015)","key":"11_CR17","DOI":"10.1109\/CVPR.2015.7298611"},{"key":"11_CR18","doi-asserted-by":"publisher","first-page":"79801","DOI":"10.1109\/ACCESS.2020.2990212","volume":"8","author":"S Lee","year":"2020","unstructured":"Lee, S., Lee, J., Kim, D., Kim, J.: Deep architecture with cross guidance between single image and sparse LiDAR data for depth completion. IEEE Access 8, 79801\u201379810 (2020)","journal-title":"IEEE Access"},{"doi-asserted-by":"crossref","unstructured":"Liao, Y., Huang, L., Wang, Y., Kodagoda, S., Yu, Y., Liu, Y.: Parse geometry from a line: monocular depth estimation with partial laser observation. In: IEEE International Conference on Robotics and Automation (ICRA), pp. 5059\u20135066 (2017)","key":"11_CR19","DOI":"10.1109\/ICRA.2017.7989590"},{"doi-asserted-by":"crossref","unstructured":"Lin, Y., Cheng, T., Zhong, Q., Zhou, W., Yang, H.: Dynamic spatial propagation network for depth completion. arXiv preprint arXiv:2202.09769 (2022)","key":"11_CR20","DOI":"10.1609\/aaai.v36i2.20055"},{"unstructured":"Liu, S., De Mello, S., Gu, J., Zhong, G., Yang, M.H., Kautz, J.: Learning affinity via spatial propagation networks. In: Proceedings on the International Conference on Neural Information Processing Systems (NIPS), vol. 30 (2017)","key":"11_CR21"},{"unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)","key":"11_CR22"},{"doi-asserted-by":"crossref","unstructured":"Lu, K., Barnes, N., Anwar, S., Zheng, L.: From depth what can you see? Depth completion via auxiliary image reconstruction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11306\u201311315 (2020)","key":"11_CR23","DOI":"10.1109\/CVPR42600.2020.01132"},{"doi-asserted-by":"crossref","unstructured":"Lu, K., Barnes, N., Anwar, S., Zheng, L.: Depth completion auto-encoder. In: IEEE\/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW), pp. 63\u201373 (2022)","key":"11_CR24","DOI":"10.1109\/WACVW54805.2022.00012"},{"doi-asserted-by":"crossref","unstructured":"Ma, F., Karaman, S.: Sparse-to-Dense: depth prediction from sparse depth samples and a single image. In: IEEE International Conference on Robotics and Automation (ICRA), pp. 4796\u20134803 (2018)","key":"11_CR25","DOI":"10.1109\/ICRA.2018.8460184"},{"unstructured":"Mathieu, M., Couprie, C., LeCun, Y.: Deep multi-scale video prediction beyond mean square error. arXiv preprint arXiv:1511.05440 (2015)","key":"11_CR26"},{"key":"11_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"120","DOI":"10.1007\/978-3-030-58601-0_8","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Park","year":"2020","unstructured":"Park, J., Joo, K., Hu, Z., Liu, C.-K., So Kweon, I.: Non-local spatial propagation network for depth completion. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12358, pp. 120\u2013136. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58601-0_8"},{"doi-asserted-by":"crossref","unstructured":"Qiu, J., et al.: DeepLiDAR: deep surface normal guided depth prediction for outdoor scene from sparse LiDAR data and single color image. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3313\u20133322 (2019)","key":"11_CR28","DOI":"10.1109\/CVPR.2019.00343"},{"doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.C.: MobileNetV2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4510\u20134520 (2018)","key":"11_CR29","DOI":"10.1109\/CVPR.2018.00474"},{"doi-asserted-by":"crossref","unstructured":"Schuster, R., Wasenmuller, O., Unger, C., Stricker, D.: SSGP: sparse spatial guided propagation for robust and generic interpolation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 197\u2013206 (2021)","key":"11_CR30","DOI":"10.1109\/WACV48630.2021.00024"},{"doi-asserted-by":"crossref","unstructured":"Senushkin, D., Romanov, M., Belikov, I., Patakin, N., Konushin, A.: Decoder modulation for indoor depth completion. In: IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2181\u20132188 (2021)","key":"11_CR31","DOI":"10.1109\/IROS51168.2021.9636870"},{"doi-asserted-by":"crossref","unstructured":"Shi, W., et al.: Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1874\u20131883 (2016)","key":"11_CR32","DOI":"10.1109\/CVPR.2016.207"},{"key":"11_CR33","doi-asserted-by":"publisher","first-page":"1116","DOI":"10.1109\/TIP.2020.3040528","volume":"30","author":"J Tang","year":"2020","unstructured":"Tang, J., Tian, F.P., Feng, W., Li, J., Tan, P.: Learning guided convolutional network for depth completion. IEEE Trans. Image Process. 30, 1116\u20131129 (2020)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Uhrig, J., Schneider, N., Schneider, L., Franke, U., Brox, T., Geiger, A.: Sparsity invariant CNNs. In: IEEE International Conference on 3D Vision (3DV), pp. 11\u201320 (2017)","key":"11_CR34","DOI":"10.1109\/3DV.2017.00012"},{"doi-asserted-by":"crossref","unstructured":"Van Gansbeke, W., Neven, D., De Brabandere, B., Van Gool, L.: Sparse and noisy LiDAR completion with RGB guidance and uncertainty. In: IEEE International Conference on Machine Vision Applications (MVA), pp. 1\u20136 (2019)","key":"11_CR35","DOI":"10.23919\/MVA.2019.8757939"},{"issue":"1","key":"11_CR36","doi-asserted-by":"publisher","first-page":"98","DOI":"10.1109\/MSP.2008.930649","volume":"26","author":"Z Wang","year":"2009","unstructured":"Wang, Z., Bovik, A.C.: Mean squared error: love it or leave it? A new look at signal fidelity measures. IEEE Sig. Process. Mag. 26(1), 98\u2013117 (2009)","journal-title":"IEEE Sig. Process. Mag."},{"doi-asserted-by":"crossref","unstructured":"Xu, Y., Zhu, X., Shi, J., Zhang, G., Bao, H., Li, H.: Depth completion from sparse LiDAR data with depth-normal constraints. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 2811\u20132820 (2019)","key":"11_CR37","DOI":"10.1109\/ICCV.2019.00290"},{"doi-asserted-by":"crossref","unstructured":"Xu, Z., Yin, H., Yao, J.: Deformable spatial propagation networks for depth completion. In: IEEE International Conference on Image Processing (ICIP), pp. 913\u2013917 (2020)","key":"11_CR38","DOI":"10.1109\/ICIP40778.2020.9191138"},{"doi-asserted-by":"crossref","unstructured":"Yan, Z., et al.: RigNet: repetitive image guided network for depth completion. arXiv preprint arXiv:2107.13802 (2021)","key":"11_CR39","DOI":"10.1007\/978-3-031-19812-0_13"},{"issue":"8","key":"11_CR40","doi-asserted-by":"publisher","first-page":"3443","DOI":"10.1109\/TIP.2014.2329776","volume":"23","author":"J Yang","year":"2014","unstructured":"Yang, J., Ye, X., Li, K., Hou, C., Wang, Y.: Color-guided depth recovery from RGB-D data using an adaptive autoregressive model. IEEE Trans. Image Process. 23(8), 3443\u20133458 (2014)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Yang, Y., Wong, A., Soatto, S.: Dense depth posterior (DDP) from single image and sparse range. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3353\u20133362 (2019)","key":"11_CR41","DOI":"10.1109\/CVPR.2019.00347"},{"key":"11_CR42","doi-asserted-by":"publisher","first-page":"5264","DOI":"10.1109\/TIP.2021.3079821","volume":"30","author":"S Zhao","year":"2021","unstructured":"Zhao, S., Gong, M., Fu, H., Tao, D.: Adaptive context-aware multi-modal network for depth completion. IEEE Trans. Image Process. 30, 5264\u20135276 (2021)","journal-title":"IEEE Trans. Image Process."},{"unstructured":"Zhong, Y., Wu, C.Y., You, S., Neumann, U.: Deep RGB-D canonical correlation analysis for sparse depth completion. In: Proceedings on the International Conference on Neural Information Processing Systems (NIPS), vol. 32 (2019)","key":"11_CR43"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25072-9_11","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T13:19:58Z","timestamp":1728911998000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25072-9_11"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031250712","9783031250729"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25072-9_11","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"18 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}