{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T22:35:32Z","timestamp":1772836532173,"version":"3.50.1"},"publisher-location":"Cham","reference-count":73,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031726835","type":"print"},{"value":"9783031726842","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72684-2_27","type":"book-chapter","created":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T19:04:22Z","timestamp":1730574262000},"page":"475-493","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":23,"title":["DVLO: Deep Visual-LiDAR Odometry with\u00a0Local-to-Global Feature Fusion and\u00a0Bi-directional Structure Alignment"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-8047-3814","authenticated-orcid":false,"given":"Jiuming","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-6747-7481","authenticated-orcid":false,"given":"Dong","family":"Zhuo","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8014-6410","authenticated-orcid":false,"given":"Zhiheng","family":"Feng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0009-9521-2682","authenticated-orcid":false,"given":"Siting","family":"Zhu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9213-5970","authenticated-orcid":false,"given":"Chensheng","family":"Peng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6753-0303","authenticated-orcid":false,"given":"Zhe","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9959-1634","authenticated-orcid":false,"given":"Hesheng","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,3]]},"reference":[{"key":"27_CR1","doi-asserted-by":"crossref","unstructured":"Ali, S.A., Aouada, D., Reis, G., Stricker, D.: Delo: deep evidential lidar odometry using partial optimal transport. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4517\u20134526 (2023)","DOI":"10.1109\/ICCVW60793.2023.00486"},{"key":"27_CR2","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1016\/j.neunet.2022.03.005","volume":"150","author":"Y Almalioglu","year":"2022","unstructured":"Almalioglu, Y., Turan, M., Saputra, M.R.U., de Gusm\u00e3o, P.P., Markham, A., Trigoni, N.: SelfVIO: self-supervised deep monocular visual-inertial odometry and depth estimation. Neural Netw. 150, 119\u2013136 (2022)","journal-title":"Neural Netw."},{"issue":"4","key":"27_CR3","doi-asserted-by":"publisher","first-page":"1496","DOI":"10.1007\/s12559-022-10010-w","volume":"14","author":"Y An","year":"2022","unstructured":"An, Y., Shi, J., Gu, D., Liu, Q.: Visual-lidar slam based on unsupervised multi-channel deep neural networks. Cogn. Comput. 14(4), 1496\u20131508 (2022)","journal-title":"Cogn. Comput."},{"key":"27_CR4","doi-asserted-by":"crossref","unstructured":"Aydemir, E., Fetic, N., Unel, M.: H-vlo: hybrid lidar-camera fusion for self-supervised odometry. In: 2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3302\u20133307. IEEE (2022)","DOI":"10.1109\/IROS47612.2022.9981111"},{"key":"27_CR5","doi-asserted-by":"crossref","unstructured":"Chen, S., Xu, H., Li, R., Liu, G., Fu, C.W., Liu, S.: SIRA-PCR: sim-to-real adaptation for 3d point cloud registration. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14394\u201314405 (2023)","DOI":"10.1109\/ICCV51070.2023.01324"},{"key":"27_CR6","doi-asserted-by":"publisher","first-page":"11642","DOI":"10.1109\/ACCESS.2023.3241961","volume":"11","author":"HM Cho","year":"2023","unstructured":"Cho, H.M., Kim, E.: Dynamic object-aware visual odometry (VO) estimation based on optical flow matching. IEEE Access 11, 11642\u201311651 (2023)","journal-title":"IEEE Access"},{"key":"27_CR7","doi-asserted-by":"crossref","unstructured":"Deng, J., et al.: Nerf-loam: neural implicit representation for large-scale incremental lidar odometry and mapping. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8218\u20138227 (2023)","DOI":"10.1109\/ICCV51070.2023.00755"},{"key":"27_CR8","unstructured":"Deng, T., et al.: Compact 3d gaussian splatting for dense visual slam. arXiv preprint arXiv:2403.11247 (2024)"},{"key":"27_CR9","doi-asserted-by":"crossref","unstructured":"Deng, T., et al.: PLGSLAM: progressive neural scene represenation with local to global bundle adjustment. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 19657\u201319666 (2024)","DOI":"10.1109\/CVPR52733.2024.01859"},{"issue":"1","key":"27_CR10","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1109\/MRA.2022.3228492","volume":"30","author":"T Deng","year":"2023","unstructured":"Deng, T., Xie, H., Wang, J., Chen, W.: Long-term visual simultaneous localization and mapping: using a bayesian persistence filter-based global map prediction. IEEE Robot. Autom. Mag. 30(1), 36\u201349 (2023)","journal-title":"IEEE Robot. Autom. Mag."},{"key":"27_CR11","doi-asserted-by":"crossref","unstructured":"Ding, F., Palffy, A., Gavrila, D.M., Lu, C.X.: Hidden gems: 4d radar scene flow learning using cross-modal supervision. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9340\u20139349 (2023)","DOI":"10.1109\/CVPR52729.2023.00901"},{"issue":"3","key":"27_CR12","doi-asserted-by":"publisher","first-page":"8233","DOI":"10.1109\/LRA.2022.3187248","volume":"7","author":"F Ding","year":"2022","unstructured":"Ding, F., Pan, Z., Deng, Y., Deng, J., Lu, C.X.: Self-supervised scene flow estimation with 4-d automotive radar. IEEE Robot. Autom. Lett. 7(3), 8233\u20138240 (2022)","journal-title":"IEEE Robot. Autom. Lett."},{"issue":"11","key":"27_CR13","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: the kitti dataset. The Int. J. Robot. Res. 32(11), 1231\u20131237 (2013)","journal-title":"The Int. J. Robot. Res."},{"key":"27_CR14","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The kitti vision benchmark suite. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3354\u20133361 (2012)","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"27_CR15","doi-asserted-by":"crossref","unstructured":"Graeter, J., Wilczynski, A., Lauer, M.: Limo: lidar-monocular visual odometry. In: 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 7872\u20137879. IEEE (2018)","DOI":"10.1109\/IROS.2018.8594394"},{"key":"27_CR16","doi-asserted-by":"crossref","unstructured":"Huang, S.S., Ma, Z.Y., Mu, T.J., Fu, H., Hu, S.M.: Lidar-monocular visual odometry using point and line features. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 1091\u20131097. IEEE (2020)","DOI":"10.1109\/ICRA40945.2020.9196613"},{"key":"27_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1007\/978-3-030-58555-6_3","volume-title":"Computer Vision \u2013 ECCV 2020","author":"T Huang","year":"2020","unstructured":"Huang, T., Liu, Z., Chen, X., Bai, X.: EPNet: enhancing point features with image semantics for 3d object detection. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020, Part XV. LNCS, vol. 12360, pp. 35\u201352. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58555-6_3"},{"key":"27_CR18","doi-asserted-by":"crossref","unstructured":"Ilg, E., Mayer, N., Saikia, T., Keuper, M., Dosovitskiy, A., Brox, T.: Flownet 2.0: evolution of optical flow estimation with deep networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2462\u20132470 (2017)","DOI":"10.1109\/CVPR.2017.179"},{"key":"27_CR19","unstructured":"Jiang, C., et al.: NeuroGauss4D-PCI: 4d neural fields and gaussian deformation fields for point cloud interpolation. arXiv preprint arXiv:2405.14241 (2024)"},{"key":"27_CR20","doi-asserted-by":"crossref","unstructured":"Jiang, C., et al.: 3dsflabelling: boosting 3d scene flow estimation by pseudo auto-labelling. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 15173\u201315183 (2024)","DOI":"10.1109\/CVPR52733.2024.01437"},{"key":"27_CR21","doi-asserted-by":"crossref","unstructured":"Kendall, A., Grimes, M., Cipolla, R.: Posenet: a convolutional network for real-time 6-dof camera relocalization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2938\u20132946 (2015)","DOI":"10.1109\/ICCV.2015.336"},{"issue":"2015","key":"27_CR22","first-page":"486","volume":"1","author":"KR Konda","year":"2015","unstructured":"Konda, K.R., Memisevic, R.: Learning visual odometry with a convolutional network. VISAPP 1(2015), 486\u2013490 (2015)","journal-title":"VISAPP"},{"key":"27_CR23","doi-asserted-by":"crossref","unstructured":"Leutenegger, S., Furgale, P., Rabaud, V., Chli, M., Konolige, K., Siegwart, R.: Keyframe-based visual-inertial slam using nonlinear optimization. In: Proceedings of Robotis Science and Systems (RSS), vol. 2013 (2013)","DOI":"10.15607\/RSS.2013.IX.037"},{"key":"27_CR24","doi-asserted-by":"crossref","unstructured":"Li, B., Hu, M., Wang, S., Wang, L., Gong, X.: Self-supervised visual-lidar odometry with flip consistency. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 3844\u20133852 (2021)","DOI":"10.1109\/WACV48630.2021.00389"},{"issue":"3","key":"27_CR25","doi-asserted-by":"publisher","first-page":"6171","DOI":"10.1109\/LRA.2022.3166457","volume":"7","author":"B Li","year":"2022","unstructured":"Li, B., Wang, S., Ye, H., Gong, X., Xiang, Z.: Cross-modal knowledge distillation for depth privileged monocular visual odometry. IEEE Robot. Autom. Lett. 7(3), 6171\u20136178 (2022)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"27_CR26","doi-asserted-by":"crossref","unstructured":"Li, Q., et al.: Lo-net: deep real-time lidar odometry. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8473\u20138482 (2019)","DOI":"10.1109\/CVPR.2019.00867"},{"key":"27_CR27","doi-asserted-by":"crossref","unstructured":"Li, X., et\u00a0al.: Logonet: towards accurate 3d object detection with local-to-global cross-modal fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17524\u201317534 (2023)","DOI":"10.1109\/CVPR52729.2023.01681"},{"key":"27_CR28","doi-asserted-by":"crossref","unstructured":"Liu, H., Lu, T., Xu, Y., Liu, J., Li, W., Chen, L.: Camliflow: bidirectional camera-lidar fusion for joint optical flow and scene flow estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5791\u20135801 (2022)","DOI":"10.1109\/CVPR52688.2022.00570"},{"key":"27_CR29","doi-asserted-by":"publisher","first-page":"2378","DOI":"10.1109\/TPAMI.2023.3330866","volume":"46","author":"H Liu","year":"2023","unstructured":"Liu, H., Lu, T., Xu, Y., Liu, J., Wang, L.: Learning optical flow and scene flow with bidirectional camera-lidar fusion. IEEE Trans. Pattern Anal. Mach. Intell. 46, 2378\u20132395 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR30","doi-asserted-by":"crossref","unstructured":"Liu, J., Wang, G., Jiang, C., Liu, Z., Wang, H.: Translo: a window-based masked point transformer framework for large-scale lidar odometry. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a037, pp. 1683\u20131691 (2023)","DOI":"10.1609\/aaai.v37i2.25256"},{"key":"27_CR31","doi-asserted-by":"crossref","unstructured":"Liu, J., Wang, G., Liu, Z., Jiang, C., Pollefeys, M., Wang, H.: Regformer: an efficient projection-aware transformer network for large-scale point cloud registration. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 8451\u20138460, October 2023","DOI":"10.1109\/ICCV51070.2023.00776"},{"key":"27_CR32","doi-asserted-by":"crossref","unstructured":"Liu, J., et al.: Difflow3d: toward robust uncertainty-aware scene flow estimation with iterative diffusion-based refinement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 15109\u201315119 (2024)","DOI":"10.1109\/CVPR52733.2024.01431"},{"issue":"6","key":"27_CR33","doi-asserted-by":"publisher","first-page":"8907","DOI":"10.1109\/TII.2024.3378829","volume":"20","author":"Q Liu","year":"2024","unstructured":"Liu, Q., Chen, N., Liu, Z., Wang, H.: Toward learning-based visuomotor navigation with neural radiance fields. IEEE Trans. Industr. Inf. 20(6), 8907\u20138916 (2024). https:\/\/doi.org\/10.1109\/TII.2024.3378829","journal-title":"IEEE Trans. Industr. Inf."},{"key":"27_CR34","doi-asserted-by":"publisher","unstructured":"Liu, Q., Cui, X., Liu, Z., Wang, H.: Boosting explore-exploit behavior for navigating vehicle by maintaining informative topological frontier. IEEE Trans. Intell. Veh. 1\u201311 (2024). https:\/\/doi.org\/10.1109\/TIV.2024.3415734","DOI":"10.1109\/TIV.2024.3415734"},{"key":"27_CR35","doi-asserted-by":"publisher","unstructured":"Liu, Q., Wang, G., Liu, Z., Wang, H.: Visuomotor navigation for embodied robots with spatial memory and semantic reasoning cognition. IEEE Trans. Neural Netw. Learn. Syst. 1\u201312 (2024). https:\/\/doi.org\/10.1109\/TNNLS.2024.3418857","DOI":"10.1109\/TNNLS.2024.3418857"},{"key":"27_CR36","doi-asserted-by":"crossref","unstructured":"Liu, X., Qi, C.R., Guibas, L.J.: Flownet3d: learning scene flow in 3d point clouds. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 529\u2013537 (2019)","DOI":"10.1109\/CVPR.2019.00062"},{"key":"27_CR37","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"27_CR38","unstructured":"Ma, X., et al.: Image as set of points. In: International Conference on Learning Representations (ICLR) (2023)"},{"key":"27_CR39","doi-asserted-by":"crossref","unstructured":"Mayer, N., et al.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4040\u20134048 (2016)","DOI":"10.1109\/CVPR.2016.438"},{"key":"27_CR40","doi-asserted-by":"crossref","unstructured":"Naumann, J., Xu, B., Leutenegger, S., Zuo, X.: Nerf-VO: real-time sparse visual odometry with neural radiance fields. arXiv preprint arXiv:2312.13471 (2023)","DOI":"10.1109\/LRA.2024.3421192"},{"key":"27_CR41","unstructured":"Nicolai, A., Skeele, R., Eriksen, C., Hollinger, G.A.: Deep learning for laser based odometry estimation. In: RSS workshop Limits and Potentials of Deep Learning in Robotics, vol.\u00a0184, p.\u00a01 (2016)"},{"key":"27_CR42","doi-asserted-by":"crossref","unstructured":"Ouyang, B., Raviv, D.: Occlusion guided scene flow estimation on 3d point clouds. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2805\u20132814 (2021)","DOI":"10.1109\/CVPRW53098.2021.00315"},{"key":"27_CR43","doi-asserted-by":"crossref","unstructured":"Peng, C., et al.: Delflow: dense efficient learning of scene flow for large-scale point clouds. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 16901\u201316910 (2023)","DOI":"10.1109\/ICCV51070.2023.01550"},{"issue":"10","key":"27_CR44","doi-asserted-by":"publisher","first-page":"2478","DOI":"10.1109\/TPAMI.2019.2909895","volume":"42","author":"EJ Shamwell","year":"2019","unstructured":"Shamwell, E.J., Lindgren, K., Leung, S., Nothwang, W.D.: Unsupervised deep visual-inertial odometry with online error correction for RGB-d imagery. IEEE Trans. Pattern Anal. Mach. Intell. 42(10), 2478\u20132493 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR45","doi-asserted-by":"publisher","first-page":"1030","DOI":"10.1109\/TMRB.2024.3407369","volume":"6","author":"J Shan","year":"2024","unstructured":"Shan, J., Li, Y., Xie, T., Wang, H.: ENeRF-SLAM: a dense endoscopic slam with neural implicit representation. IEEE Trans. Med. Robot. Bionics 6, 1030\u20131041 (2024)","journal-title":"IEEE Trans. Med. Robot. Bionics"},{"issue":"2","key":"27_CR46","doi-asserted-by":"publisher","first-page":"115","DOI":"10.1007\/s10514-019-09881-0","volume":"44","author":"YS Shin","year":"2020","unstructured":"Shin, Y.S., Park, Y.S., Kim, A.: Dvl-slam: sparse depth enhanced direct visual-lidar slam. Auton. Robot. 44(2), 115\u2013130 (2020)","journal-title":"Auton. Robot."},{"key":"27_CR47","doi-asserted-by":"publisher","first-page":"3384","DOI":"10.1109\/TIV.2022.3215141","volume":"8","author":"C Shu","year":"2022","unstructured":"Shu, C., Luo, Y.: Multi-modal feature constraint based tightly coupled monocular visual-lidar odometry and mapping. IEEE Trans. Intell. Veh. 8, 3384\u20133393 (2022)","journal-title":"IEEE Trans. Intell. Veh."},{"key":"27_CR48","unstructured":"Shubodh, S., Omama, M., Zaidi, H., Parihar, U.S., Krishna, M.: LIP-Loc: LiDAR image pretraining for cross-modal localization. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 948\u2013957 (2024)"},{"issue":"4","key":"27_CR49","doi-asserted-by":"publisher","first-page":"730","DOI":"10.1109\/TPAMI.2015.2469274","volume":"38","author":"S Song","year":"2015","unstructured":"Song, S., Chandraker, M., Guest, C.C.: High accuracy monocular SFM and scale correction for autonomous driving. IEEE Trans. Pattern Anal. Mach. Intell. 38(4), 730\u2013743 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR50","doi-asserted-by":"crossref","unstructured":"Sun, D., Yang, X., Liu, M.Y., Kautz, J.: PWC-Net: CNNs for optical flow using pyramid, warping, and cost volume. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8934\u20138943 (2018)","DOI":"10.1109\/CVPR.2018.00931"},{"key":"27_CR51","doi-asserted-by":"publisher","first-page":"22064","DOI":"10.1109\/JSEN.2023.3302401","volume":"23","author":"L Sun","year":"2023","unstructured":"Sun, L., Ding, G., Qiu, Y., Yoshiyasu, Y., Kanehiro, F.: TransFusionOdom: transformer-based lidar-inertial fusion odometry estimation. IEEE Sens. J. 23, 22064\u201322079 (2023)","journal-title":"IEEE Sens. J."},{"key":"27_CR52","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"402","DOI":"10.1007\/978-3-030-58536-5_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Z Teed","year":"2020","unstructured":"Teed, Z., Deng, J.: RAFT: recurrent all-pairs field transforms for optical flow. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020, Part II. LNCS, vol. 12347, pp. 402\u2013419. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_24"},{"key":"27_CR53","doi-asserted-by":"crossref","unstructured":"Valente, M., Joly, C., de\u00a0La\u00a0Fortelle, A.: Deep sensor fusion for real-time odometry estimation. In: 2019 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 6679\u20136685. IEEE (2019)","DOI":"10.1109\/IROS40897.2019.8967803"},{"issue":"5","key":"27_CR54","first-page":"5749","volume":"45","author":"G Wang","year":"2022","unstructured":"Wang, G., Wu, X., Jiang, S., Liu, Z., Wang, H.: Efficient 3d deep lidar odometry. IEEE Trans. Pattern Anal. Mach. Intell. 45(5), 5749\u20135765 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR55","doi-asserted-by":"crossref","unstructured":"Wang, G., Wu, X., Liu, Z., Wang, H.: PWCLO-Net: deep LiDAR odometry in 3d point clouds using hierarchical embedding mask optimization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 15910\u201315919 (2021)","DOI":"10.1109\/CVPR46437.2021.01565"},{"key":"27_CR56","doi-asserted-by":"crossref","unstructured":"Wang, S., Clark, R., Wen, H., Trigoni, N.: DeepVO: towards end-to-end visual odometry with deep recurrent convolutional neural networks. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 2043\u20132050. IEEE (2017)","DOI":"10.1109\/ICRA.2017.7989236"},{"issue":"16","key":"27_CR57","doi-asserted-by":"publisher","first-page":"3340","DOI":"10.3390\/rs13163340","volume":"13","author":"W Wang","year":"2021","unstructured":"Wang, W., Liu, J., Wang, C., Luo, B., Zhang, C.: DV-LOAM: direct visual lidar odometry and mapping. Remote Sens. 13(16), 3340 (2021)","journal-title":"Remote Sens."},{"key":"27_CR58","doi-asserted-by":"crossref","unstructured":"Wang, W., et al.: DeepPCO: end-to-end point cloud odometry through deep parallel neural network. In: 2019 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3248\u20133254. IEEE (2019)","DOI":"10.1109\/IROS40897.2019.8967756"},{"key":"27_CR59","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: Animatabledreamer: text-guided non-rigid 3d model generation and reconstruction with canonical score distillation. arXiv preprint arXiv:2312.03795 (2023)","DOI":"10.1007\/978-3-031-72698-9_19"},{"key":"27_CR60","unstructured":"Wang, Y., Wang, X., Chen, Z., Wang, Z., Sun, F., Zhu, J.: Vidu4d: single generated video to high-fidelity 4d reconstruction with dynamic gaussian surfels. arXiv preprint arXiv:2405.16822 (2024)"},{"key":"27_CR61","unstructured":"Wu, W., et al.: DVN-SLAM: dynamic visual neural slam based on local-global encoding. arXiv preprint arXiv:2403.11776 (2024)"},{"key":"27_CR62","doi-asserted-by":"crossref","unstructured":"Wu, W., et al.: EMIE-MAP: large-scale road surface reconstruction based on explicit mesh and implicit encoding. arXiv preprint arXiv:2403.11789 (2024)","DOI":"10.1007\/978-3-031-73021-4_22"},{"key":"27_CR63","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"88","DOI":"10.1007\/978-3-030-58558-7_6","volume-title":"Computer Vision \u2013 ECCV 2020","author":"W Wu","year":"2020","unstructured":"Wu, W., Wang, Z.Y., Li, Z., Liu, W., Fuxin, L.: PointPWC-net: cost volume on point clouds for (Self-)supervised scene flow estimation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020, Part V. LNCS, vol. 12350, pp. 88\u2013107. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58558-7_6"},{"key":"27_CR64","doi-asserted-by":"publisher","first-page":"11203","DOI":"10.1109\/TPAMI.2023.3262817","volume":"45","author":"Z Yuan","year":"2023","unstructured":"Yuan, Z., Wang, Q., Cheng, K., Hao, T., Yang, X.: SDV-LOAM: semi-direct visual-lidar odometry and mapping. IEEE Trans. Pattern Anal. Mach. Intell. 45, 11203\u201311220 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR65","unstructured":"Zhan, H., Weerasekera, C.S., Bian, J.W., Garg, R., Reid, I.: DF-VO: what should be learnt for visual odometry? arXiv preprint arXiv:2103.00933 (2021)"},{"key":"27_CR66","doi-asserted-by":"crossref","unstructured":"Zhang, J., Singh, S.: Loam: Lidar odometry and mapping in real-time. In: Robotics: Science and Systems, vol.\u00a02, pp.\u00a01\u20139. Berkeley, CA (2014)","DOI":"10.15607\/RSS.2014.X.007"},{"key":"27_CR67","doi-asserted-by":"crossref","unstructured":"Zhang, J., Singh, S.: Visual-lidar odometry and mapping: low-drift, robust, and fast. In: 2015 IEEE International Conference on Robotics and Automation (ICRA), pp. 2174\u20132181. IEEE (2015)","DOI":"10.1109\/ICRA.2015.7139486"},{"key":"27_CR68","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Yang, Y., Fang, H., Geng, R., Jensfelt, P.: Deflow: decoder of scene flow network in autonomous driving. arXiv preprint arXiv:2401.16122 (2024)","DOI":"10.1109\/ICRA57147.2024.10610278"},{"key":"27_CR69","doi-asserted-by":"crossref","unstructured":"Zhang, Y.X., Gui, J., Cong, X., Gong, X., Tao, W.: A comprehensive survey and taxonomy on point cloud registration based on deep learning. arXiv preprint arXiv:2404.13830 (2024)","DOI":"10.24963\/ijcai.2024\/922"},{"key":"27_CR70","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.G.: Unsupervised learning of depth and ego-motion from video. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1851\u20131858 (2017)","DOI":"10.1109\/CVPR.2017.700"},{"key":"27_CR71","unstructured":"Zhu, S., Qin, R., Wang, G., Liu, J., Wang, H.: Semgauss-slam: dense semantic gaussian splatting slam. arXiv preprint arXiv:2403.07494 (2024)"},{"key":"27_CR72","doi-asserted-by":"crossref","unstructured":"Zhu, S., Wang, G., Blum, H., Liu, J., Song, L., Pollefeys, M., Wang, H.: SNI-SLAM: semantic neural implicit slam. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 21167\u201321177 (2024)","DOI":"10.1109\/CVPR52733.2024.02000"},{"key":"27_CR73","first-page":"5065","volume":"9","author":"G Zhuoins","year":"2023","unstructured":"Zhuoins, G., Lu, S., Xiong, L., Zhouins, H., Zheng, L., Zhou, M.: 4DRVO-net: deep 4d radar-visual odometry using multi-modal and multi-scale adaptive fusion. IEEE Trans. Intell. Veh. 9, 5065\u20135079 (2023)","journal-title":"IEEE Trans. Intell. Veh."}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72684-2_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T18:18:18Z","timestamp":1732990698000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72684-2_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,3]]},"ISBN":["9783031726835","9783031726842"],"references-count":73,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72684-2_27","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,3]]},"assertion":[{"value":"3 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}