{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T00:49:00Z","timestamp":1774226940350,"version":"3.50.1"},"publisher-location":"Cham","reference-count":61,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198236","type":"print"},{"value":"9783031198243","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19824-3_6","type":"book-chapter","created":{"date-parts":[[2022,11,10]],"date-time":"2022-11-10T21:14:32Z","timestamp":1668114872000},"page":"87-105","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Learning Online Multi-sensor Depth Fusion"],"prefix":"10.1007","author":[{"given":"Erik","family":"Sandstr\u00f6m","sequence":"first","affiliation":[]},{"given":"Martin R.","family":"Oswald","sequence":"additional","affiliation":[]},{"given":"Suryansh","family":"Kumar","sequence":"additional","affiliation":[]},{"given":"Silvan","family":"Weder","sequence":"additional","affiliation":[]},{"given":"Fisher","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Cristian","family":"Sminchisescu","sequence":"additional","affiliation":[]},{"given":"Luc","family":"Van Gool","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,11]]},"reference":[{"key":"6_CR1","doi-asserted-by":"crossref","unstructured":"Agresti, G., Minto, L., Marin, G., Zanuttigh, P.: Deep learning for confidence information in stereo and ToF data fusion. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 697\u2013705 (2017)","DOI":"10.1109\/ICCVW.2017.88"},{"key":"6_CR2","doi-asserted-by":"publisher","first-page":"161","DOI":"10.1016\/j.inffus.2018.11.006","volume":"49","author":"G Agresti","year":"2019","unstructured":"Agresti, G., Minto, L., Marin, G., Zanuttigh, P.: Stereo and ToF data fusion by learning from synthetic data. Inf. Fusion 49, 161\u2013173 (2019)","journal-title":"Inf. Fusion"},{"key":"6_CR3","doi-asserted-by":"publisher","first-page":"136471","DOI":"10.1109\/ACCESS.2019.2942375","volume":"7","author":"MK Ali","year":"2019","unstructured":"Ali, M.K., Rajput, A., Shahzad, M., Khan, F., Akhtar, F., B\u00f6rner, A.: Multi-sensor depth fusion framework for real-time 3d reconstruction. IEEE Access 7, 136471\u2013136480 (2019)","journal-title":"IEEE Access"},{"key":"6_CR4","unstructured":"Bo\u017ei\u010d, A., Palafox, P., Thies, J., Dai, A., Nie\u00dfner, M.: Transformerfusion: monocular RGB scene reconstruction using transformers. arXiv preprint arXiv:2107.02191 (2021)"},{"key":"6_CR5","doi-asserted-by":"crossref","unstructured":"Bylow, E., Olsson, C., Kahl, F.: Robust online 3d reconstruction combining a depth sensor and sparse feature points. In: 2016 23rd International Conference on Pattern Recognition (ICPR), pp. 3709\u20133714 (2016)","DOI":"10.1109\/ICPR.2016.7900211"},{"key":"6_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1007\/978-3-030-20205-7_22","volume-title":"Image Analysis","author":"E Bylow","year":"2019","unstructured":"Bylow, E., Maier, R., Kahl, F., Olsson, C.: Combining depth fusion and photometric stereo for fine-detailed 3d models. In: Felsberg, M., Forss\u00e9n, P.-E., Sintorn, I.-M., Unger, J. (eds.) SCIA 2019. LNCS, vol. 11482, pp. 261\u2013274. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-20205-7_22"},{"issue":"5","key":"6_CR7","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3182157","volume":"37","author":"YP Cao","year":"2018","unstructured":"Cao, Y.P., Kobbelt, L., Hu, S.M.: Real-time high-accuracy three-dimensional reconstruction with consumer RGB-D cameras. ACM Trans. Graph. (TOG) 37(5), 1\u201316 (2018)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"6_CR8","doi-asserted-by":"crossref","unstructured":"Chang, J.R., Chen, Y.S.: Pyramid stereo matching network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5410\u20135418 (2018)","DOI":"10.1109\/CVPR.2018.00567"},{"key":"6_CR9","doi-asserted-by":"crossref","unstructured":"Choe, J., Im, S., Rameau, F., Kang, M., Kweon, I.S.: VolumeFusion: deep depth fusion for 3d scene reconstruction. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 16086\u201316095, October 2021","DOI":"10.1109\/ICCV48922.2021.01578"},{"key":"6_CR10","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"640","DOI":"10.1007\/978-3-642-37447-0_49","volume-title":"Computer Vision","author":"O Choi","year":"2013","unstructured":"Choi, O., Lee, S.: Fusion of time-of-flight and stereo for disambiguation of depth measurements. In: Lee, K.M., Matsushita, Y., Rehg, J.M., Hu, Z. (eds.) ACCV 2012. LNCS, vol. 7727, pp. 640\u2013653. Springer, Heidelberg (2013). https:\/\/doi.org\/10.1007\/978-3-642-37447-0_49"},{"key":"6_CR11","doi-asserted-by":"crossref","unstructured":"Curless, B., Levoy, M.: A volumetric method for building complex models from range images. In: Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, pp. 303\u2013312 (1996)","DOI":"10.1145\/237170.237269"},{"issue":"4","key":"6_CR12","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3054739","volume":"36","author":"A Dai","year":"2017","unstructured":"Dai, A., Nie\u00dfner, M., Zollh\u00f6fer, M., Izadi, S., Theobalt, C.: BundleFusion: real-time globally consistent 3d reconstruction using on-the-fly surface reintegration. ACM Trans. Graph. (ToG) 36(4), 1 (2017)","journal-title":"ACM Trans. Graph. (ToG)"},{"issue":"11","key":"6_CR13","doi-asserted-by":"publisher","first-page":"2260","DOI":"10.1109\/TPAMI.2015.2408361","volume":"37","author":"C Dal Mutto","year":"2015","unstructured":"Dal Mutto, C., Zanuttigh, P., Cortelazzo, G.M.: Probabilistic TOF and stereo data fusion based on mixed pixels measurement models. IEEE Trans. Pattern Anal. Mach. Intell. 37(11), 2260\u20132272 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"6_CR14","doi-asserted-by":"publisher","first-page":"2739","DOI":"10.1109\/TMM.2021.3087017","volume":"24","author":"Y Deng","year":"2021","unstructured":"Deng, Y., Xiao, J., Zhou, S.Z.: TOF and stereo data fusion using dynamic search range stereo matching. IEEE Trans. Multimedia 24, 2739\u20132751 (2021)","journal-title":"IEEE Trans. Multimedia"},{"key":"6_CR15","doi-asserted-by":"crossref","unstructured":"Dong, W., Wang, Q., Wang, X., Zha, H.: PSDF fusion: probabilistic signed distance function for on-the-fly 3d data fusion and scene reconstruction. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 701\u2013717 (2018)","DOI":"10.1007\/978-3-030-01240-3_43"},{"key":"6_CR16","doi-asserted-by":"crossref","unstructured":"Duan, Y., Pei, M., Wang, Y.: Probabilistic depth map fusion of kinect and stereo in real-time. In: 2012 IEEE International Conference on Robotics and Biomimetics (ROBIO), pp. 2317\u20132322. IEEE (2012)","DOI":"10.1109\/ROBIO.2012.6491315"},{"issue":"4","key":"6_CR17","first-page":"1309","volume":"31","author":"Y Duan","year":"2015","unstructured":"Duan, Y., Pei, M., Wang, Y., Yang, M., Qin, I., Jia, Y.: A unified probabilistic framework for real-time depth map fusion. J. Inf. Sci. Eng. 31(4), 1309\u20131327 (2015)","journal-title":"J. Inf. Sci. Eng."},{"issue":"11","key":"6_CR18","doi-asserted-by":"publisher","first-page":"2178","DOI":"10.1109\/TPAMI.2015.2400465","volume":"37","author":"GD Evangelidis","year":"2015","unstructured":"Evangelidis, G.D., Hansard, M., Horaud, R.: Fusion of range and stereo data for high-resolution scene-modeling. IEEE Trans. Pattern Anal. Mach. Intell. 37(11), 2178\u20132192 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"11","key":"6_CR19","doi-asserted-by":"publisher","first-page":"2895","DOI":"10.1109\/TVCG.2018.2868533","volume":"24","author":"S Golodetz","year":"2018","unstructured":"Golodetz, S., Cavallari, T., Lord, N.A., Prisacariu, V.A., Murray, D.W., Torr, P.H.: Collaborative large-scale dense 3d reconstruction with online inter-agent pose optimisation. IEEE Trans. Visual. Comput. Graph. 24(11), 2895\u20132905 (2018)","journal-title":"IEEE Trans. Visual. Comput. Graph."},{"key":"6_CR20","doi-asserted-by":"crossref","unstructured":"Gu, P., et al.: A 3d reconstruction method using multisensor fusion in large-scale indoor scenes. Complexity 2020 (2020)","DOI":"10.1155\/2020\/6973790"},{"key":"6_CR21","doi-asserted-by":"crossref","unstructured":"Handa, A., Whelan, T., McDonald, J., Davison, A.J.: A benchmark for RGB-D visual odometry, 3d reconstruction and slam. In: 2014 IEEE International Conference on Robotics and Automation (ICRA), pp. 1524\u20131531. IEEE (2014)","DOI":"10.1109\/ICRA.2014.6907054"},{"issue":"2","key":"6_CR22","doi-asserted-by":"publisher","first-page":"328","DOI":"10.1109\/TPAMI.2007.1166","volume":"30","author":"H Hirschmuller","year":"2007","unstructured":"Hirschmuller, H.: Stereo processing by semiglobal matching and mutual information. IEEE Trans. Pattern Anal. Mach. Intell. 30(2), 328\u2013341 (2007)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"6_CR23","doi-asserted-by":"crossref","unstructured":"Huang, J., Huang, S.S., Song, H., Hu, S.M.: Di-fusion: online implicit 3d reconstruction with deep priors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8932\u20138941 (2021)","DOI":"10.1109\/CVPR46437.2021.00882"},{"key":"6_CR24","doi-asserted-by":"crossref","unstructured":"Izadi, S., et al.: KinectFusion: real-time 3d reconstruction and interaction using a moving depth camera. In: Proceedings of the 24th annual ACM Symposium on User Interface Software and Technology, pp. 559\u2013568. ACM (2011)","DOI":"10.1145\/2047196.2047270"},{"issue":"11","key":"6_CR25","doi-asserted-by":"publisher","first-page":"1241","DOI":"10.1109\/TVCG.2015.2459891","volume":"21","author":"O K\u00e4hler","year":"2015","unstructured":"K\u00e4hler, O., Prisacariu, V.A., Ren, C.Y., Sun, X., Torr, P.H.S., Murray, D.W.: Very high frame rate volumetric integration of depth images on mobile devices. IEEE Trans. Vis. Comput. Graph. 21(11), 1241\u20131250 (2015)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"issue":"3","key":"6_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2487228.2487237","volume":"32","author":"M Kazhdan","year":"2013","unstructured":"Kazhdan, M., Hoppe, H.: Screened poisson surface reconstruction. ACM Trans. Graph. (ToG) 32(3), 1\u201313 (2013)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"6_CR27","doi-asserted-by":"crossref","unstructured":"Kim, Y.M., Theobalt, C., Diebel, J., Kosecka, J., Miscusik, B., Thrun, S.: Multi-view image and TOF sensor fusion for dense 3d reconstruction. In: 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV workshops, pp. 1542\u20131549. IEEE (2009)","DOI":"10.1109\/ICCVW.2009.5457430"},{"key":"6_CR28","unstructured":"Lefloch, D., Weyrich, T., Kolb, A.: Anisotropic point-based fusion. In: 2015 18th International Conference on Information Fusion (Fusion), pp. 2121\u20132128. IEEE (2015)"},{"issue":"4","key":"6_CR29","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1145\/37402.37422","volume":"21","author":"WE Lorensen","year":"1987","unstructured":"Lorensen, W.E., Cline, H.E.: Marching cubes: a high resolution 3d surface construction algorithm. ACM siggraph Comput. Graph. 21(4), 163\u2013169 (1987)","journal-title":"ACM siggraph Comput. Graph."},{"issue":"11","key":"6_CR30","first-page":"2579","volume":"9","author":"L Van der Maaten","year":"2008","unstructured":"Van der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11), 2579\u20132605 (2008)","journal-title":"J. Mach. Learn. Res."},{"key":"6_CR31","doi-asserted-by":"crossref","unstructured":"Maddern, W., Newman, P.: Real-time probabilistic fusion of sparse 3d lidar and dense stereo. In: 2016 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2181\u20132188. IEEE (2016)","DOI":"10.1109\/IROS.2016.7759342"},{"key":"6_CR32","doi-asserted-by":"crossref","unstructured":"Savva, M., et al.: Habitat: a platform for embodied AI Research. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00943"},{"key":"6_CR33","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"386","DOI":"10.1007\/978-3-319-46478-7_24","volume-title":"Computer Vision","author":"G Marin","year":"2016","unstructured":"Marin, G., Zanuttigh, P., Mattoccia, S.: Reliable fusion of ToF and stereo depth driven by confidence measures. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9911, pp. 386\u2013401. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46478-7_24"},{"key":"6_CR34","doi-asserted-by":"crossref","unstructured":"Martins, D., Van Hecke, K., De Croon, G.: Fusion of stereo and still monocular depth estimates in a self-supervised learning context. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 849\u2013856. IEEE (2018)","DOI":"10.1109\/ICRA.2018.8461116"},{"key":"6_CR35","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"414","DOI":"10.1007\/978-3-030-58571-6_25","volume-title":"Computer Vision","author":"Z Murez","year":"2020","unstructured":"Murez, Z., van As, T., Bartolozzi, J., Sinha, A., Badrinarayanan, V., Rabinovich, A.: Atlas: end-to-end 3D scene reconstruction from posed images. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12352, pp. 414\u2013431. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58571-6_25"},{"key":"6_CR36","doi-asserted-by":"crossref","unstructured":"Newcombe, R.A., et al.: KinectFusion: real-time dense surface mapping and tracking. In: ISMAR, vol. 11, pp. 127\u2013136 (2011)","DOI":"10.1109\/ISMAR.2011.6092378"},{"key":"6_CR37","doi-asserted-by":"crossref","unstructured":"Newcombe, R.A., Lovegrove, S.J., Davison, A.J.: DTAM: dense tracking and mapping in real-time. In: ICCV (2011)","DOI":"10.1109\/ICCV.2011.6126513"},{"key":"6_CR38","doi-asserted-by":"publisher","unstructured":"Nie\u00dfner, M., Zollh\u00f6fer, M., Izadi, S., Stamminger, M.: Real-time 3d reconstruction at scale using voxel hashing. ACM Trans. Graph. (TOG) 32 (2013). https:\/\/doi.org\/10.1145\/2508363.2508374","DOI":"10.1145\/2508363.2508374"},{"key":"6_CR39","doi-asserted-by":"publisher","unstructured":"Oleynikova, H., Taylor, Z., Fehr, M., Siegwart, R., Nieto, J.I.: Voxblox: incremental 3d euclidean signed distance fields for on-board MAV planning. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems, IROS 2017, Vancouver, BC, Canada, 24\u201328 September 2017, pp. 1366\u20131373. IEEE (2017). https:\/\/doi.org\/10.1109\/IROS.2017.8202315","DOI":"10.1109\/IROS.2017.8202315"},{"key":"6_CR40","doi-asserted-by":"crossref","unstructured":"Park, K., Kim, S., Sohn, K.: High-precision depth estimation with the 3d lidar and stereo fusion. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 2156\u20132163. IEEE (2018)","DOI":"10.1109\/ICRA.2018.8461048"},{"key":"6_CR41","doi-asserted-by":"crossref","unstructured":"Patil, V., Van Gansbeke, W., Dai, D., Van Gool, L.: Don\u2019t forget the past: Recurrent depth estimation from monocular video. IEEE Robot. Autom. Lett. 5(4), 6813\u20136820 (2020)","DOI":"10.1109\/LRA.2020.3017478"},{"key":"6_CR42","doi-asserted-by":"crossref","unstructured":"Poggi, M., Mattoccia, S.: Deep stereo fusion: combining multiple disparity hypotheses with deep-learning. In: 2016 Fourth International Conference on 3D Vision (3DV), pp. 138\u2013147. IEEE (2016)","DOI":"10.1109\/3DV.2016.22"},{"issue":"5","key":"6_CR43","doi-asserted-by":"publisher","first-page":"487","DOI":"10.3390\/rs11050487","volume":"11","author":"C Pu","year":"2019","unstructured":"Pu, C., Song, R., Tylecek, R., Li, N., Fisher, R.B.: SDF-MAN: semi-supervised disparity fusion with multi-scale adversarial networks. Remote Sens. 11(5), 487 (2019)","journal-title":"Remote Sens."},{"key":"6_CR44","doi-asserted-by":"publisher","unstructured":"Qiu, J., et al.: Deeplidar: deep surface normal guided depth prediction for outdoor scene from sparse lidar data and single color image. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16\u201320 June 2019, pp. 3313\u20133322. Computer Vision Foundation\/IEEE (2019). https:\/\/doi.org\/10.1109\/CVPR.2019.00343, https:\/\/openaccess.thecvf.com\/content_CVPR_2019\/html\/Qiu_DeepLiDAR_Deep_Surface_Normal_Guided_Depth_Prediction_for_Outdoor_Scene_CVPR_2019_paper.html","DOI":"10.1109\/CVPR.2019.00343"},{"key":"6_CR45","doi-asserted-by":"crossref","unstructured":"Rozumnyi, D., Cherabier, I., Pollefeys, M., Oswald, M.R.: Learned semantic multi-sensor depth map fusion. In: International Conference on Computer Vision Workshop (ICCVW), Workshop on 3D Reconstruction in the Wild, 2019. Seoul, South Korea (2019)","DOI":"10.1109\/ICCVW.2019.00264"},{"key":"6_CR46","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"501","DOI":"10.1007\/978-3-319-46487-9_31","volume-title":"Computer Vision","author":"JL Sch\u00f6nberger","year":"2016","unstructured":"Sch\u00f6nberger, J.L., Zheng, E., Frahm, J.-M., Pollefeys, M.: Pixelwise view selection for unstructured multi-view stereo. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 501\u2013518. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46487-9_31"},{"key":"6_CR47","doi-asserted-by":"crossref","unstructured":"Schops, T., Sattler, T., Pollefeys, M.: BAD SLAM: bundle adjusted direct RGB-D SLAM. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00022"},{"key":"6_CR48","doi-asserted-by":"crossref","unstructured":"Steinbrucker, F., Kerl, C., Cremers, D., Sturm, J.: Large-scale multi-resolution surface reconstruction from RGB-D sequences. In: 2013 IEEE International Conference on Computer Vision, pp. 3264\u20133271 (2013)","DOI":"10.1109\/ICCV.2013.405"},{"key":"6_CR49","unstructured":"Straub, J., et al.: The replica dataset: a digital replica of indoor spaces. arXiv preprint arXiv:1906.05797 (2019)"},{"key":"6_CR50","doi-asserted-by":"crossref","unstructured":"Sucar, E., Liu, S., Ortiz, J., Davison, A.: iMAP: implicit mapping and positioning in real-time. In: Proceedings of the IEEE International Conference on Computer Vision (2021)","DOI":"10.1109\/ICCV48922.2021.00617"},{"key":"6_CR51","doi-asserted-by":"crossref","unstructured":"Sun, J., Xie, Y., Chen, L., Zhou, X., Bao, H.: NeuralRecon: real-time coherent 3d reconstruction from monocular video. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 15598\u201315607 (2021)","DOI":"10.1109\/CVPR46437.2021.01534"},{"key":"6_CR52","doi-asserted-by":"crossref","unstructured":"Van Baar, J., Beardsley, P., Pollefeys, M., Gross, M.: Sensor fusion for depth estimation, including TOF and thermal sensors. In: 2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission, pp. 472\u2013478. IEEE (2012)","DOI":"10.1109\/3DIMPVT.2012.69"},{"key":"6_CR53","doi-asserted-by":"crossref","unstructured":"Wasenm\u00fcller, O., Meyer, M., Stricker, D.: Corbs: comprehensive RGB-D benchmark for slam using kinect v2. In: 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1\u20137. IEEE (2016)","DOI":"10.1109\/WACV.2016.7477636"},{"key":"6_CR54","doi-asserted-by":"crossref","unstructured":"Weder, S., Sch\u00f6nberger, J.L., Pollefeys, M., Oswald, M.R.: RoutedFusion: learning real-time depth map fusion. ArXiv abs\/2001.04388 (2020)","DOI":"10.1109\/CVPR42600.2020.00494"},{"key":"6_CR55","doi-asserted-by":"crossref","unstructured":"Weder, S., Schonberger, J.L., Pollefeys, M., Oswald, M.R.: NeuralFusion: online depth fusion in latent space. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3162\u20133172 (2021)","DOI":"10.1109\/CVPR46437.2021.00318"},{"key":"6_CR56","doi-asserted-by":"crossref","unstructured":"Yan, Z., Tian, Y., Shi, X., Guo, P., Wang, P., Zha, H.: Continual neural mapping: learning an implicit scene representation from sequential observations. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 15782\u201315792, October 2021","DOI":"10.1109\/ICCV48922.2021.01549"},{"issue":"5","key":"6_CR57","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3389412","volume":"39","author":"S Yang","year":"2020","unstructured":"Yang, S., et al.: Noise-resilient reconstruction of panoramas and 3d scenes using robot-mounted unsynchronized commodity RGB-D cameras. ACM Trans. Graph. (TOG) 39(5), 1\u201315 (2020)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"11","key":"6_CR58","doi-asserted-by":"publisher","first-page":"3217","DOI":"10.1109\/TVCG.2019.2919619","volume":"26","author":"S Yang","year":"2019","unstructured":"Yang, S., Li, B., Liu, M., Lai, Y.K., Kobbelt, L., Hu, S.M.: HeteroFusion: dense scene reconstruction integrating multi-sensors. IEEE Trans. Visual. Comput. Graph. 26(11), 3217\u20133230 (2019)","journal-title":"IEEE Trans. Visual. Comput. Graph."},{"issue":"4","key":"6_CR59","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2461912.2461967","volume":"32","author":"QY Zhou","year":"2013","unstructured":"Zhou, Q.Y., Koltun, V.: Dense scene reconstruction with points of interest. ACM Trans. Graph. (ToG) 32(4), 1\u20138 (2013)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"6_CR60","doi-asserted-by":"crossref","unstructured":"Zhu, Z., et al.: Nice-slam: neural implicit scalable encoding for slam. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12786\u201312796 (2022)","DOI":"10.1109\/CVPR52688.2022.01245"},{"key":"6_CR61","doi-asserted-by":"crossref","unstructured":"Zollh\u00f6fer, M., et al.: State of the art on 3d reconstruction with RGB-D cameras. In: Computer Graphics Forum, vol. 37, pp. 625\u2013652. Wiley Online Library (2018)","DOI":"10.1111\/cgf.13386"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19824-3_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,12]],"date-time":"2022-11-12T00:05:12Z","timestamp":1668211512000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19824-3_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198236","9783031198243"],"references-count":61,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19824-3_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"11 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}