{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,27]],"date-time":"2025-08-27T15:52:03Z","timestamp":1756309923977,"version":"3.40.3"},"publisher-location":"Cham","reference-count":34,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031250712"},{"type":"electronic","value":"9783031250729"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25072-9_1","type":"book-chapter","created":{"date-parts":[[2023,2,17]],"date-time":"2023-02-17T08:40:04Z","timestamp":1676623204000},"page":"3-20","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["MIPI 2022 Challenge on\u00a0RGB+ToF Depth Completion: Dataset and\u00a0Report"],"prefix":"10.1007","author":[{"given":"Wenxiu","family":"Sun","sequence":"first","affiliation":[]},{"given":"Qingpeng","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Chongyi","family":"Li","sequence":"additional","affiliation":[]},{"given":"Ruicheng","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Shangchen","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Jun","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Qingyu","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Chen Change","family":"Loy","sequence":"additional","affiliation":[]},{"given":"Jinwei","family":"Gu","sequence":"additional","affiliation":[]},{"given":"Dewang","family":"Hou","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Liying","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Yu","family":"Li","sequence":"additional","affiliation":[]},{"given":"Huaijia","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Ruizheng","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Jiangbo","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Jiaya","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Qiang","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Haosong","family":"Yue","sequence":"additional","affiliation":[]},{"given":"Danyang","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Lehang","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Jiaxuan","family":"Quan","sequence":"additional","affiliation":[]},{"given":"Jixiang","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Yufei","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yuchao","family":"Dai","sequence":"additional","affiliation":[]},{"given":"Peng","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Hu","family":"Yan","sequence":"additional","affiliation":[]},{"given":"Houbiao","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Siyuan","family":"Su","sequence":"additional","affiliation":[]},{"given":"Xuanhe","family":"Li","sequence":"additional","affiliation":[]},{"given":"Rui","family":"Ren","sequence":"additional","affiliation":[]},{"given":"Yunlong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yufan","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Dong","family":"Lao","sequence":"additional","affiliation":[]},{"given":"Alex","family":"Wong","sequence":"additional","affiliation":[]},{"given":"Katie","family":"Chang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,18]]},"reference":[{"unstructured":"Baruch, G., et al.: Arkitscenes-a diverse real-world dataset for 3D indoor scene understanding using mobile RGB-D data. arXiv preprint arXiv:2111.08897 (2021)","key":"1_CR1"},{"issue":"11","key":"1_CR2","doi-asserted-by":"publisher","first-page":"5034","DOI":"10.1109\/TNNLS.2020.3026669","volume":"32","author":"X Chen","year":"2020","unstructured":"Chen, X., Chen, X., Zhang, Y., Fu, X., Zha, Z.J.: Laplacian pyramid neural network for dense continuous-value regression for complex scenes. IEEE Trans. Neural Netw. Learn. Syst. 32(11), 5034\u20135046 (2020)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"doi-asserted-by":"crossref","unstructured":"Chen, Z., Badrinarayanan, V., Drozdov, G., Rabinovich, A.: Estimating depth from RGB and sparse sensing. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 167\u2013182 (2018)","key":"1_CR3","DOI":"10.1007\/978-3-030-01225-0_11"},{"doi-asserted-by":"crossref","unstructured":"Cheng, X., Wang, P., Guan, C., Yang, R.: CSPN++: learning context and resource aware convolutional spatial propagation networks for depth completion. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 10615\u201310622 (2020)","key":"1_CR4","DOI":"10.1609\/aaai.v34i07.6635"},{"doi-asserted-by":"crossref","unstructured":"Cheng, X., Wang, P., Yang, R.: Depth estimation via affinity learned with convolutional spatial propagation network. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 103\u2013119 (2018)","key":"1_CR5","DOI":"10.1007\/978-3-030-01270-0_7"},{"issue":"10","key":"1_CR6","doi-asserted-by":"publisher","first-page":"2361","DOI":"10.1109\/TPAMI.2019.2947374","volume":"42","author":"X Cheng","year":"2019","unstructured":"Cheng, X., Wang, P., Yang, R.: Learning depth with convolutional spatial propagation network. IEEE Trans. Pattern Anal. Mach. Intell. 42(10), 2361\u20132379 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1_CR7","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1007\/BFb0086566","volume-title":"Constructive Theory of Functions of Several Variables","author":"J Duchon","year":"1977","unstructured":"Duchon, J.: Splines minimizing rotation-invariant semi-norms in sobolev spaces. In: Schempp, W., Zeller, K. (eds.) Constructive Theory of Functions of Several Variables, pp. 85\u2013100. Springer, Heidelberg (1977). https:\/\/doi.org\/10.1007\/BFb0086566"},{"doi-asserted-by":"crossref","unstructured":"Eldesokey, A., Felsberg, M., Holmquist, K., Persson, M.: Uncertainty-aware CNNs for depth completion: uncertainty from beginning to end. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12014\u201312023 (2020)","key":"1_CR8","DOI":"10.1109\/CVPR42600.2020.01203"},{"issue":"10","key":"1_CR9","doi-asserted-by":"publisher","first-page":"2423","DOI":"10.1109\/TPAMI.2019.2929170","volume":"42","author":"A Eldesokey","year":"2019","unstructured":"Eldesokey, A., Felsberg, M., Khan, F.S.: Confidence propagation through CNNs for guided sparse depth regression. IEEE Trans. Pattern Anal. Mach. Intell. 42(10), 2423\u20132436 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"unstructured":"Howard, A.G., et al.: MobileNets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)","key":"1_CR10"},{"unstructured":"Hu, J., et al.: Deep depth completion: a survey. arXiv preprint arXiv:2205.05335 (2022)","key":"1_CR11"},{"doi-asserted-by":"crossref","unstructured":"Hu, M., Wang, S., Li, B., Ning, S., Fan, L., Gong, X.: PENet: towards precise and efficient image guided depth completion. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 13656\u201313662. IEEE (2021)","key":"1_CR12","DOI":"10.1109\/ICRA48506.2021.9561035"},{"doi-asserted-by":"crossref","unstructured":"Imran, S., Long, Y., Liu, X., Morris, D.: Depth coefficients for depth completion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12438\u201312447. IEEE (2019)","key":"1_CR13","DOI":"10.1109\/CVPR.2019.01273"},{"doi-asserted-by":"crossref","unstructured":"Jeon, J., Lee, S.: Reconstruction-based pairwise depth dataset for depth image enhancement using CNN. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 422\u2013438 (2018)","key":"1_CR14","DOI":"10.1007\/978-3-030-01270-0_26"},{"doi-asserted-by":"crossref","unstructured":"Lee, B.U., Lee, K., Kweon, I.S.: Depth completion using plane-residual representation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13916\u201313925 (2021)","key":"1_CR15","DOI":"10.1109\/CVPR46437.2021.01370"},{"doi-asserted-by":"crossref","unstructured":"Li, A., Yuan, Z., Ling, Y., Chi, W., Zhang, C., et al.: A multi-scale guided cascade hourglass network for depth completion. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 32\u201340 (2020)","key":"1_CR16","DOI":"10.1109\/WACV45572.2020.9093407"},{"doi-asserted-by":"crossref","unstructured":"Lin, Y., Cheng, T., Zhong, Q., Zhou, W., Yang, H.: Dynamic spatial propagation network for depth completion. arXiv preprint arXiv:2202.09769 (2022)","key":"1_CR17","DOI":"10.1609\/aaai.v36i2.20055"},{"doi-asserted-by":"crossref","unstructured":"Liu, Q., Yue, H., Lyu, Z., Wang, W., Liu, Z., Chen, W.: SEHLNet: separate estimation of high-and low-frequency components for depth completion. In: 2022 International Conference on Robotics and Automation (ICRA), pp. 668\u2013674. IEEE (2022)","key":"1_CR18","DOI":"10.1109\/ICRA46639.2022.9811840"},{"doi-asserted-by":"crossref","unstructured":"L\u00f3pez-Randulfe, J., Veiga, C., Rodr\u00edguez-Andina, J.J., Farina, J.: A quantitative method for selecting denoising filters, based on a new edge-sensitive metric. In: 2017 IEEE International Conference on Industrial Technology (ICIT), pp. 974\u2013979. IEEE (2017)","key":"1_CR19","DOI":"10.1109\/ICIT.2017.7915493"},{"doi-asserted-by":"crossref","unstructured":"Lopez-Rodriguez, A., Busam, B., Mikolajczyk, K.: Project to adapt: domain adaptation for depth completion from noisy and sparse sensor data. In: Proceedings of the Asian Conference on Computer Vision (2020)","key":"1_CR20","DOI":"10.1007\/978-3-030-69525-5_20"},{"issue":"1","key":"1_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41598-021-01763-9","volume":"11","author":"G Luetzenburg","year":"2021","unstructured":"Luetzenburg, G., Kroon, A., Bj\u00f8rk, A.A.: Evaluation of the apple iPhone 12 pro lidar for an application in geosciences. Sci. Rep. 11(1), 1\u20139 (2021)","journal-title":"Sci. Rep."},{"doi-asserted-by":"crossref","unstructured":"Ma, F., Cavalheiro, G.V., Karaman, S.: Self-supervised sparse-to-dense: self-supervised depth completion from lidar and monocular camera. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 3288\u20133295. IEEE (2019)","key":"1_CR22","DOI":"10.1109\/ICRA.2019.8793637"},{"doi-asserted-by":"crossref","unstructured":"McCormac, J., Handa, A., Leutenegger, S., Davison, A.J.: Scenenet RGB-D: can 5M synthetic images beat generic imagenet pre-training on indoor segmentation? In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2678\u20132687 (2017)","key":"1_CR23","DOI":"10.1109\/ICCV.2017.292"},{"key":"1_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"120","DOI":"10.1007\/978-3-030-58601-0_8","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Park","year":"2020","unstructured":"Park, J., Joo, K., Hu, Z., Liu, C.-K., So Kweon, I.: Non-local spatial propagation network for depth completion. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12358, pp. 120\u2013136. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58601-0_8"},{"doi-asserted-by":"crossref","unstructured":"Qu, C., Nguyen, T., Taylor, C.: Depth completion via deep basis fitting. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 71\u201380 (2020)","key":"1_CR25","DOI":"10.1109\/WACV45572.2020.9093349"},{"issue":"3","key":"1_CR26","doi-asserted-by":"publisher","first-page":"1623","DOI":"10.1109\/TPAMI.2020.3019967","volume":"44","author":"R Ranftl","year":"2020","unstructured":"Ranftl, R., Lasinger, K., Hafner, D., Schindler, K., Koltun, V.: Towards robust monocular depth estimation: mixing datasets for zero-shot cross-dataset transfer. IEEE Trans. Pattern Anal. Mach. Intell. 44(3), 1623\u20131637 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"doi-asserted-by":"crossref","unstructured":"Shi, W., et al.: Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1874\u20131883 (2016)","key":"1_CR27","DOI":"10.1109\/CVPR.2016.207"},{"key":"1_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"746","DOI":"10.1007\/978-3-642-33715-4_54","volume-title":"Computer Vision \u2013 ECCV 2012","author":"N Silberman","year":"2012","unstructured":"Silberman, N., Hoiem, D., Kohli, P., Fergus, R.: Indoor segmentation and support inference from RGBD images. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7576, pp. 746\u2013760. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33715-4_54"},{"issue":"11","key":"1_CR29","doi-asserted-by":"publisher","first-page":"4381","DOI":"10.1109\/TCSVT.2021.3049869","volume":"31","author":"M Song","year":"2021","unstructured":"Song, M., Lim, S., Kim, W.: Monocular depth estimation using laplacian pyramid-based depth residuals. IEEE Trans. Circuits Syst. Video Technol. 31(11), 4381\u20134393 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"doi-asserted-by":"crossref","unstructured":"Sun, P., et al.: Scalability in perception for autonomous driving: Waymo open dataset. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2446\u20132454 (2020)","key":"1_CR30","DOI":"10.1109\/CVPR42600.2020.00252"},{"key":"1_CR31","doi-asserted-by":"publisher","first-page":"1116","DOI":"10.1109\/TIP.2020.3040528","volume":"30","author":"J Tang","year":"2020","unstructured":"Tang, J., Tian, F.P., Feng, W., Li, J., Tan, P.: Learning guided convolutional network for depth completion. IEEE Trans. Image Process. 30, 1116\u20131129 (2020)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Uhrig, J., Schneider, N., Schneider, L., Franke, U., Brox, T., Geiger, A.: Sparsity invariant CNNs. In: International Conference on 3D Vision (3DV) (2017)","key":"1_CR32","DOI":"10.1109\/3DV.2017.00012"},{"doi-asserted-by":"crossref","unstructured":"Van Gansbeke, W., Neven, D., De Brabandere, B., Van Gool, L.: Sparse and noisy lidar completion with RGB guidance and uncertainty. In: 2019 16th International Conference on Machine Vision Applications (MVA), pp. 1\u20136. IEEE (2019)","key":"1_CR33","DOI":"10.23919\/MVA.2019.8757939"},{"issue":"2","key":"1_CR34","doi-asserted-by":"publisher","first-page":"1495","DOI":"10.1109\/LRA.2021.3058072","volume":"6","author":"A Wong","year":"2021","unstructured":"Wong, A., Cicek, S., Soatto, S.: Learning topology from synthetic data for unsupervised depth completion. IEEE Robot. Autom. Lett. 6(2), 1495\u20131502 (2021)","journal-title":"IEEE Robot. Autom. Lett."}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25072-9_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T13:18:41Z","timestamp":1728911921000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25072-9_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031250712","9783031250729"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25072-9_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"18 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}