{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T01:08:12Z","timestamp":1743124092406,"version":"3.40.3"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031250712"},{"type":"electronic","value":"9783031250729"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25072-9_13","type":"book-chapter","created":{"date-parts":[[2023,2,17]],"date-time":"2023-02-17T08:40:04Z","timestamp":1676623204000},"page":"192-207","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Depth Completion Using Laplacian Pyramid-Based Depth Residuals"],"prefix":"10.1007","author":[{"given":"Haosong","family":"Yue","sequence":"first","affiliation":[]},{"given":"Qiang","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Zhong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jing","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xingming","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,18]]},"reference":[{"key":"13_CR1","unstructured":"Qiang, L., Haosong, Y., Zhanggang, L., Wei, W., Zhong, L., Weihai, C.: SEHLNet: separate estimation of high- and low-frequency components for depth completion. In: ICRA (2022)"},{"key":"13_CR2","doi-asserted-by":"crossref","unstructured":"Hu, M., Wang, S., Li, B., Ning, S., Fan, L., Gong, X.: PENet: towards precise and efficient image guided depth completion. In: ICRA (2021)","DOI":"10.1109\/ICRA48506.2021.9561035"},{"issue":"11","key":"13_CR3","doi-asserted-by":"publisher","first-page":"4381","DOI":"10.1109\/TCSVT.2021.3049869","volume":"31","author":"M Song","year":"2021","unstructured":"Song, M., Lim, S., Kim, W.: Monocular depth estimation using Laplacian pyramid-based depth residuals. IEEE Trans. Circ. Syst. Video Technol. 31(11), 4381\u20134393 (2021)","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"issue":"11","key":"13_CR4","doi-asserted-by":"publisher","first-page":"5034","DOI":"10.1109\/TNNLS.2020.3026669","volume":"32","author":"X Chen","year":"2021","unstructured":"Chen, X., Chen, X., Zhang, Y., Fu, X., Zha, Z.J.: Laplacian pyramid neural network for dense continuous-value regression for complex scenes. IEEE Trans. Neural Netw. Learn. Syst. 32(11), 5034\u20135046 (2021)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"13_CR5","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"438","DOI":"10.1007\/978-3-030-01270-0_26","volume-title":"Computer Vision \u2013 ECCV 2018","author":"J Jeon","year":"2018","unstructured":"Jeon, J., Lee, S.: Reconstruction-based pairwise depth dataset for depth image enhancement using CNN. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11220, pp. 438\u2013454. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01270-0_26"},{"key":"13_CR6","doi-asserted-by":"crossref","unstructured":"Li, D., et al.: Involution: inverting the inherence of convolution for visual recognition (2021)","DOI":"10.1109\/CVPR46437.2021.01214"},{"key":"13_CR7","doi-asserted-by":"crossref","unstructured":"Burt, P.J., Adelson, E.H.: The Laplacian pyramid as a compact image code. In: Readings in Computer Vision, vol. 31, no. 4, pp. 671\u2013679 (1987)","DOI":"10.1016\/B978-0-08-051581-6.50065-9"},{"issue":"11","key":"13_CR8","doi-asserted-by":"publisher","first-page":"2599","DOI":"10.1109\/TPAMI.2018.2865304","volume":"41","author":"WS Lai","year":"2018","unstructured":"Lai, W.S., Huang, J.B., Ahuja, N., Yang, M.H.: Fast and accurate image super-resolution with deep Laplacian pyramid networks. IEEE Trans. Pattern Anal. Mach. Intell. 41(11), 2599\u20132613 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"13_CR9","doi-asserted-by":"crossref","unstructured":"Cheng, X., Wang, P., Yang, R.: Learning depth with convolutional spatial propagation network. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01270-0_7"},{"key":"13_CR10","unstructured":"Liu, S., Mello, S.D., Gu, J., Zhong, G., Yang, M.H., Kautz, J.: Learning affinity via spatial propagation networks (2017)"},{"key":"13_CR11","doi-asserted-by":"crossref","unstructured":"Cheng, X., Wang, P., Guan, C., Yang, R.: CSPN++: learning context and resource aware convolutional spatial propagation networks for depth completion (2019)","DOI":"10.1609\/aaai.v34i07.6635"},{"key":"13_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"120","DOI":"10.1007\/978-3-030-58601-0_8","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Park","year":"2020","unstructured":"Park, J., Joo, K., Hu, Z., Liu, C.-K., So Kweon, I.: Non-local spatial propagation network for depth completion. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12358, pp. 120\u2013136. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58601-0_8"},{"key":"13_CR13","doi-asserted-by":"crossref","unstructured":"Xu, Z., Wang, Y., Yao, J.: Deformable spatial propagation network for depth completion (2020)","DOI":"10.1109\/ICIP40778.2020.9191138"},{"key":"13_CR14","doi-asserted-by":"crossref","unstructured":"Song, X., et al.: ApolloCar3D: a large 3D car instance understanding benchmark for autonomous driving. IEEE (2018)","DOI":"10.1109\/CVPR.2019.00560"},{"key":"13_CR15","doi-asserted-by":"crossref","unstructured":"Liao, Y., Huang, L., Yue, W., Kodagoda, S., Yong, L.: Parse geometry from a line: monocular depth estimation with partial laser observation. IEEE (2017)","DOI":"10.1109\/ICRA.2017.7989590"},{"issue":"1","key":"13_CR16","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1089\/cpb.2007.9935","volume":"11","author":"C Armbruester","year":"2008","unstructured":"Armbruester, C., Wolter, M., Kuhlen, T., Spijkers, W., Fimm, B.: Depth perception in virtual reality: distance estimations in peri-and extrapersonal space. CyberPsychology 11(1), 9\u201315 (2008)","journal-title":"CyberPsychology"},{"key":"13_CR17","doi-asserted-by":"crossref","unstructured":"Uhrig, J., Schneider, N., Schneider, L., Franke, U., Brox, T., Geiger, A.: Sparsity invariant CNNs, pp. 11\u201320. IEEE Computer Society (2017)","DOI":"10.1109\/3DV.2017.00012"},{"key":"13_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"499","DOI":"10.1007\/978-3-030-20887-5_31","volume-title":"Computer Vision \u2013 ACCV 2018","author":"N Chodosh","year":"2019","unstructured":"Chodosh, N., Wang, C., Lucey, S.: Deep convolutional compressed sensing for LiDAR depth completion. In: Jawahar, C.V., Li, H., Mori, G., Schindler, K. (eds.) ACCV 2018. LNCS, vol. 11361, pp. 499\u2013513. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-20887-5_31"},{"key":"13_CR19","doi-asserted-by":"crossref","unstructured":"Ma, F., Cavalheiro, G.V., Karaman, S.: Self-supervised sparse-to-dense: self-supervised depth completion from LiDAR and monocular camera (2018)","DOI":"10.1109\/ICRA.2019.8793637"},{"key":"13_CR20","doi-asserted-by":"crossref","unstructured":"Chen, Y., Yang, B., Liang, M., Urtasun, R.: Learning joint 2D\u20133D representations for depth completion. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV) (2020)","DOI":"10.1109\/ICCV.2019.01012"},{"key":"13_CR21","doi-asserted-by":"crossref","unstructured":"Qiu, J., Cui, Z., Zhang, Y., Zhang, X., Pollefeys, M.: DeepLiDAR: deep surface normal guided depth prediction for outdoor scene from sparse lidar data and single color image. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.00343"},{"key":"13_CR22","doi-asserted-by":"publisher","first-page":"5264","DOI":"10.1109\/TIP.2021.3079821","volume":"30","author":"S Zhao","year":"2021","unstructured":"Zhao, S., Gong, M., Fu, H., Tao, D.: Adaptive context-aware multi-modal network for depth completion. IEEE Trans. Image Process. 30, 5264\u20135276 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"13_CR23","doi-asserted-by":"crossref","unstructured":"Liu, L., Song, X., Lyu, X., Diao, J., Zhang, L.: FCFR-Net: feature fusion based coarse-to-fine residual learning for monocular depth completion (2020)","DOI":"10.1609\/aaai.v35i3.16311"},{"key":"13_CR24","doi-asserted-by":"publisher","first-page":"1116","DOI":"10.1109\/TIP.2020.3040528","volume":"30","author":"J Tang","year":"2021","unstructured":"Tang, J., Tian, F.P., Feng, W., Li, J., Tan, P.: Learning guided convolutional network for depth completion. IEEE Trans. Image Process. 30, 1116\u20131129 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"13_CR25","doi-asserted-by":"crossref","unstructured":"Lin, Y., Cheng, T., Zhong, Q., Zhou, W., Yang, H.: Dynamic spatial propagation network for depth completion (2022)","DOI":"10.1609\/aaai.v36i2.20055"},{"key":"13_CR26","doi-asserted-by":"crossref","unstructured":"Van Gansbeke, W., Neven, D., Brabandere, B.D., Van Gool, L.: Sparse and noisy LiDAR completion with RGB guidance and uncertainty. In: 2019 16th International Conference on Machine Vision Applications (MVA) (2019)","DOI":"10.23919\/MVA.2019.8757939"},{"key":"13_CR27","unstructured":"Paszke, A., Gross, S., Massa, F., Lerer, A., Chintala, S.: PyTorch: an imperative style, high-performance deep learning library (2019)"},{"key":"13_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"214","DOI":"10.1007\/978-3-031-19812-0_13","volume-title":"Computer Vision - ECCV 2022","author":"Z Yan","year":"2021","unstructured":"Yan, Z., et al.: RigNet: repetitive image guided network for depth completion. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13687, pp. 214\u2013230. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-031-19812-0_13"},{"key":"13_CR29","doi-asserted-by":"crossref","unstructured":"Nazir, D., Liwicki, M., Stricker, D., Afzal, M.Z.: SemAttNet: towards attention-based semantic aware guided depth completion. arXiv e-prints (2022)","DOI":"10.1109\/ACCESS.2022.3214316"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25072-9_13","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T15:38:07Z","timestamp":1710257887000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25072-9_13"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031250712","9783031250729"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25072-9_13","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"18 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}