{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:20:36Z","timestamp":1740108036945,"version":"3.37.3"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"21","license":[{"start":{"date-parts":[[2022,6,24]],"date-time":"2022-06-24T00:00:00Z","timestamp":1656028800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,6,24]],"date-time":"2022-06-24T00:00:00Z","timestamp":1656028800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100016152","name":"YUTP","doi-asserted-by":"crossref","award":["015LC0- 243"],"award-info":[{"award-number":["015LC0- 243"]}],"id":[{"id":"10.13039\/501100016152","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2022,11]]},"DOI":"10.1007\/s00521-022-07484-y","type":"journal-article","created":{"date-parts":[[2022,6,24]],"date-time":"2022-06-24T08:42:24Z","timestamp":1656060144000},"page":"18823-18836","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Lightweight spatial attentive network for vehicular visual odometry estimation in urban environments"],"prefix":"10.1007","volume":"34","author":[{"given":"Nivesh","family":"Gadipudi","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4721-9400","authenticated-orcid":false,"given":"Irraivan","family":"Elamvazuthi","sequence":"additional","affiliation":[]},{"given":"Cheng-Kai","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Sivajothi","family":"Paramasivam","sequence":"additional","affiliation":[]},{"given":"Steven","family":"Su","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,6,24]]},"reference":[{"key":"7484_CR1","doi-asserted-by":"publisher","first-page":"513","DOI":"10.1177\/0278364917734298","volume":"37","author":"S Wang","year":"2018","unstructured":"Wang S, Clark R, Wen H, Trigoni A (2018) End-to-end, sequence-to-sequence probabilistic visual odometry through deep neural networks. Int J Rob Res 37:513\u2013542. https:\/\/doi.org\/10.1177\/0278364917734298","journal-title":"Int J Rob Res"},{"issue":"4","key":"7484_CR2","doi-asserted-by":"publisher","first-page":"289","DOI":"10.1007\/s40903-015-0032-7","volume":"1","author":"K Yousif","year":"2015","unstructured":"Yousif K, Bab-Hadiashar A, Hoseinnezhad R (2015) An overview to visual odometry and visual slam: applications to mobile robotics. Intell Indus Syst 1(4):289\u2013311. https:\/\/doi.org\/10.1007\/s40903-015-0032-7","journal-title":"Intell Indus Syst"},{"issue":"8","key":"7484_CR3","doi-asserted-by":"publisher","first-page":"3209","DOI":"10.1007\/s00521-020-05192-z","volume":"33","author":"M Zhai","year":"2021","unstructured":"Zhai M, Xiang X (2021) Geometry understanding from autonomous driving scenarios based on feature refinement. Neural Comput Appl 33(8):3209\u20133220. https:\/\/doi.org\/10.1007\/s00521-020-05192-z","journal-title":"Neural Comput Appl"},{"key":"7484_CR4","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1016\/j.isprsjprs.2020.06.010","volume":"166","author":"K Liu","year":"2020","unstructured":"Liu K, Li Q, Qiu G (2020) Posegan: a pose-to-image translation framework for camera localization. ISPRS J Photogramm Remote Sens 166:308\u2013315. https:\/\/doi.org\/10.1016\/j.isprsjprs.2020.06.010","journal-title":"ISPRS J Photogramm Remote Sens"},{"doi-asserted-by":"crossref","unstructured":"Klein, G., Murray, D.: Parallel tracking and mapping for small ar workspaces. In: Proceedings of the IEEE and ACM International symposium on mixed and augmented reality, pp. 225\u2013234 (2007)","key":"7484_CR5","DOI":"10.1109\/ISMAR.2007.4538852"},{"issue":"6","key":"7484_CR6","doi-asserted-by":"publisher","first-page":"1052","DOI":"10.1109\/TPAMI.2007.1049","volume":"29","author":"AJ Davison","year":"2007","unstructured":"Davison AJ, Reid ID, Molton ND, Stasse O (2007) Monoslam: real-time single camera slam. IEEE Trans Pattern Anal Mach Intell 29(6):1052\u20131067. https:\/\/doi.org\/10.1109\/TPAMI.2007.1049","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"5","key":"7484_CR7","doi-asserted-by":"publisher","first-page":"1147","DOI":"10.1109\/TRO.2015.2463671","volume":"31","author":"R Mur-Artal","year":"2015","unstructured":"Mur-Artal R, Montiel JMM, Tardos JD (2015) ORB-SLAM: a versatile and accurate monocular slam system. IEEE Trans Robot 31(5):1147\u20131163. https:\/\/doi.org\/10.1109\/TRO.2015.2463671","journal-title":"IEEE Trans Robot"},{"issue":"5","key":"7484_CR8","doi-asserted-by":"publisher","first-page":"1383","DOI":"10.1007\/s00521-017-3032-6","volume":"29","author":"MW Cao","year":"2018","unstructured":"Cao MW, Jia W, Zhao Y, Li SJ, Liu XP (2018) Fast and robust absolute camera pose estimation with known focal length. Neural Comput Appl 29(5):1383\u20131398. https:\/\/doi.org\/10.1007\/s00521-017-3032-6","journal-title":"Neural Comput Appl"},{"doi-asserted-by":"crossref","unstructured":"Newcombe, R.A., Lovegrove, S.J., Davison, A.J.: Dtam: Dense tracking and mapping in real-time. In: Proceedings of the IEEE International conference on computer vision (ICCV), pp. 2320\u20132327 (2011)","key":"7484_CR9","DOI":"10.1109\/ICCV.2011.6126513"},{"doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? the kitti vision benchmark suite. In: Proceedings of the IEEE Conference on computer vision and pattern recognition (CVPR), pp. 3354\u20133361 (2012)","key":"7484_CR10","DOI":"10.1109\/CVPR.2012.6248074"},{"issue":"3","key":"7484_CR11","doi-asserted-by":"publisher","first-page":"346","DOI":"10.1016\/j.cviu.2007.09.014","volume":"110","author":"H Bay","year":"2008","unstructured":"Bay H, Ess A, Tuytelaars T, Van Gool L (2008) Speeded-up robust features (surf). Comput Vis Image Underst 110(3):346\u2013359. https:\/\/doi.org\/10.1016\/j.cviu.2007.09.014","journal-title":"Comput Vis Image Underst"},{"doi-asserted-by":"crossref","unstructured":"Muja, M., Lowe, D.G.: Fast matching of binary features. In: Proceedings of the IEEE Conference on Computer and Robot Vision, pp. 404\u2013410 (2012)","key":"7484_CR12","DOI":"10.1109\/CRV.2012.60"},{"doi-asserted-by":"crossref","unstructured":"Rublee, E., Rabaud, V., Konolige, K., Bradski, G.: Orb: An efficient alternative to sift or surf. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2564\u20132571 (2011)","key":"7484_CR13","DOI":"10.1109\/ICCV.2011.6126544"},{"doi-asserted-by":"crossref","unstructured":"Pumarola, A., Vakhitov, A., Agudo, A., Sanfeliu, A., Moreno-Noguer, F.: Pl-slam: Real-time monocular visual slam with points and lines. In: Proceedings of the IEEE International Conference on robotics and automation (ICRA), pp. 4503\u20134508 (2017)","key":"7484_CR14","DOI":"10.1109\/ICRA.2017.7989522"},{"doi-asserted-by":"crossref","unstructured":"McCormac, J., Clark, R., Bloesch, M., Davison, A., Leutenegger, S.: Fusion++: Volumetric object-level slam. In: Proceedings of the IEEE International Conference on 3D Vision (3DV), pp. 32\u201341 (2018)","key":"7484_CR15","DOI":"10.1109\/3DV.2018.00015"},{"doi-asserted-by":"crossref","unstructured":"Herrera, D.C., Kim, K., Kannala, J., Pulli, K., Heikkil\u00e4, J.: Dt-slam: Deferred triangulation for robust slam. In: Proceedings of the IEEE International Conference on 3D Vision (3DV), vol. 1, pp. 609\u2013616 (2014)","key":"7484_CR16","DOI":"10.1109\/3DV.2014.49"},{"doi-asserted-by":"crossref","unstructured":"Engel, J., Sch\u00f6ps, T., Cremers, D.: Lsd-slam: Large-scale direct monocular slam. In: Proceedings of the European Conference on computer vision (ECCV), pp. 834\u2013849 (2014)","key":"7484_CR17","DOI":"10.1007\/978-3-319-10605-2_54"},{"doi-asserted-by":"crossref","unstructured":"Forster, C., Pizzoli, M., Scaramuzza, D.: Svo: Fast semi-direct monocular visual odometry. In: Proceedings of the IEEE International Conference on robotics and automation (ICRA), pp. 15\u201322 (2014)","key":"7484_CR18","DOI":"10.1109\/ICRA.2014.6906584"},{"issue":"3","key":"7484_CR19","doi-asserted-by":"publisher","first-page":"611","DOI":"10.1109\/TPAMI.2017.2658577","volume":"40","author":"J Engel","year":"2017","unstructured":"Engel J, Koltun V, Cremers D (2017) Direct sparse odometry. IEEE Trans Pattern Anal Mach Intell 40(3):611\u2013625. https:\/\/doi.org\/10.1109\/TPAMI.2017.2658577","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"4","key":"7484_CR20","doi-asserted-by":"publisher","first-page":"1363","DOI":"10.1109\/TRO.2020.2991614","volume":"36","author":"J Zubizarreta","year":"2020","unstructured":"Zubizarreta J, Aguinaga I, Montiel JMM (2020) Direct sparse mapping. IEEE Trans Robot 36(4):1363\u20131370. https:\/\/doi.org\/10.1109\/TRO.2020.2991614","journal-title":"IEEE Trans Robot"},{"doi-asserted-by":"crossref","unstructured":"Roberts, R., Nguyen, H., Krishnamurthi, N., Balch, T.: Memory-based learning for visual odometry. In: Proceedings of the IEEE International Conference on robotics and automation (ICRA), pp. 47\u201352 (2008)","key":"7484_CR21","DOI":"10.1109\/ROBOT.2008.4543185"},{"issue":"5","key":"7484_CR22","doi-asserted-by":"publisher","first-page":"526","DOI":"10.1177\/0278364912472245","volume":"32","author":"V Guizilini","year":"2013","unstructured":"Guizilini V, Ramos F (2013) Semi-parametric learning for visual odometry. Int J Rob Res 32(5):526\u2013546. https:\/\/doi.org\/10.1177\/0278364912472245","journal-title":"Int J Rob Res"},{"doi-asserted-by":"crossref","unstructured":"Kendall, A., Grimes, M., Cipolla, R.: Posenet: A convolutional network for real-time 6-dof camera relocalization. In: Proceedings of the IEEE International Conference on computer vision (ICCV), pp. 2938\u20132946 (2015)","key":"7484_CR23","DOI":"10.1109\/ICCV.2015.336"},{"doi-asserted-by":"crossref","unstructured":"Ilg, E., Mayer, N., Saikia, T., Keuper, M., Dosovitskiy, A., Brox, T.: Flownet 2.0: Evolution of optical flow estimation with deep networks. In: Proceedings of the IEEE Conference on computer vision and pattern Recognition (CVPR), pp. 2462\u20132470 (2017)","key":"7484_CR24","DOI":"10.1109\/CVPR.2017.179"},{"doi-asserted-by":"crossref","unstructured":"CS\u00a0Kumar, A., Bhandarkar, S.M., Prasad, M.: Depthnet: A recurrent neural network architecture for monocular depth prediction. In: Proceedings of the IEEE Conference on computer vision and pattern recognition workshops (CVPRW), pp. 283\u2013291 (2018)","key":"7484_CR25","DOI":"10.1109\/CVPRW.2018.00066"},{"issue":"1","key":"7484_CR26","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/LRA.2015.2505717","volume":"1","author":"G Costante","year":"2015","unstructured":"Costante G, Mancini M, Valigi P, Ciarfuglia TA (2015) Exploring representation learning with cnns for frame-to-frame ego-motion estimation. IEEE Robot Autom Lett 1(1):18\u201325. https:\/\/doi.org\/10.1109\/LRA.2015.2505717","journal-title":"IEEE Robot Autom Lett"},{"doi-asserted-by":"crossref","unstructured":"Brox, T., Bruhn, A., Papenberg, N., Weickert, J.: High accuracy optical flow estimation based on a theory for warping. In: Proceedings of the European Conference on computer vision (ECCV), pp. 25\u201336 (2004)","key":"7484_CR27","DOI":"10.1007\/978-3-540-24673-2_3"},{"doi-asserted-by":"publisher","unstructured":"Li, X, Hou, Y, Wang, P, Gao, Z, Xu, M, Li, W Transformer guided geometry model for flow-based unsupervised visual odometry. Neural Comput Appl., 1\u201312 (2021). https:\/\/doi.org\/10.1007\/s00521-020-05545-8","key":"7484_CR28","DOI":"10.1007\/s00521-020-05545-8"},{"doi-asserted-by":"crossref","unstructured":"Muller, P., Savakis, A.: Flowdometry: An optical flow and deep learning based approach to visual odometry. In: Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 624\u2013631 (2017)","key":"7484_CR29","DOI":"10.1109\/WACV.2017.75"},{"issue":"3","key":"7484_CR30","doi-asserted-by":"publisher","first-page":"222","DOI":"10.3390\/electronics10030222","volume":"10","author":"B Zhao","year":"2021","unstructured":"Zhao B, Huang Y, Wei H, Hu X (2021) Ego-motion estimation using recurrent convolutional neural networks through optical flow learning. Electronics 10(3):222. https:\/\/doi.org\/10.3390\/electronics10030222","journal-title":"Electronics"},{"issue":"4","key":"7484_CR31","doi-asserted-by":"publisher","first-page":"1313","DOI":"10.3390\/s21041313","volume":"21","author":"T Pandey","year":"2021","unstructured":"Pandey T, Pena D, Byrne J, Moloney D (2021) Leveraging deep learning for visual odometry using optical flow. Sensors 21(4):1313. https:\/\/doi.org\/10.3390\/s21041313","journal-title":"Sensors"},{"doi-asserted-by":"crossref","unstructured":"Sun, D., Yang, X., Liu, M.-Y., Kautz, J.: Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume. In: Proceedings of the IEEE Conference on computer vision and pattern recognition (CVPR), pp. 8934\u20138943 (2018)","key":"7484_CR32","DOI":"10.1109\/CVPR.2018.00931"},{"doi-asserted-by":"crossref","unstructured":"Hui, T.-W., Tang, X., Loy, C.C.: Liteflownet: A lightweight convolutional neural network for optical flow estimation. In: Proceedings of the IEEE Conference on computer vision and pattern recognition (CVPR), pp. 8981\u20138989 (2018)","key":"7484_CR33","DOI":"10.1109\/CVPR.2018.00936"},{"doi-asserted-by":"crossref","unstructured":"Saputra, M.R.U., Gusm\u00e3o, P.P.B.D., Almalioglu, Y., Markham, A., Trigoni, A.: Distilling knowledge from a deep pose regressor network, pp. 263\u2013272 (2019)","key":"7484_CR34","DOI":"10.1109\/ICCV.2019.00035"},{"key":"7484_CR35","doi-asserted-by":"publisher","first-page":"175220","DOI":"10.1109\/ACCESS.2020.3025557","volume":"8","author":"X Wang","year":"2020","unstructured":"Wang X, Zhang H (2020) Deep monocular visual odometry for ground vehicle. IEEE Access 8:175220\u2013175229. https:\/\/doi.org\/10.1109\/ACCESS.2020.3025557","journal-title":"IEEE Access"},{"unstructured":"Kendall, A., Gal, Y.: What uncertainties do we need in bayesian deep learning for computer vision? In: Proceedings of the Neural Information Processing Systems (NIPS) (2017)","key":"7484_CR36"},{"doi-asserted-by":"crossref","unstructured":"Kendall, A., Cipolla, R.: Geometric loss functions for camera pose regression with deep learning. In: Proceedings of the IEEE Conference on computer vision and pattern recognition (CVPR), pp. 6555\u20136564 (2017)","key":"7484_CR37","DOI":"10.1109\/CVPR.2017.694"},{"doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.-Y., Kweon, I.-S.: Cbam: Convolutional block attention module. In: Proceedings of the European Conference on computer vision (ECCV) (2018)","key":"7484_CR38","DOI":"10.1007\/978-3-030-01234-2_1"},{"unstructured":"Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)","key":"7484_CR39"},{"doi-asserted-by":"crossref","unstructured":"Geiger, A., Ziegler, J., Stiller, C.: Stereoscan: Dense 3d reconstruction in real-time. In: Proceedings of the Intelligent Vehicles Symposium (IV), pp. 963\u2013968 (2011)","key":"7484_CR40","DOI":"10.1109\/IVS.2011.5940405"},{"doi-asserted-by":"crossref","unstructured":"Saputra, M.R.U., Gusm\u00e3o, P.P.B.D., Wang, S., Markham, A., Trigoni, A.: Learning monocular visual odometry through geometry-aware curriculum learning, pp. 3549\u20133555 (2019)","key":"7484_CR41","DOI":"10.1109\/ICRA.2019.8793581"},{"doi-asserted-by":"publisher","unstructured":"Liu, Y., Wang, H., Wang, J., Wang, X.: Unsupervised monocular visual odometry based on confidence evaluation. IEEE trans Intell Transp Syst, 1\u201310 (2021). https:\/\/doi.org\/10.1109\/TITS.2021.3053412","key":"7484_CR42","DOI":"10.1109\/TITS.2021.3053412"},{"doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M.A., Snavely, N., Lowe, D.: Unsupervised learning of depth and ego-motion from video. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6612\u20136619 (2017)","key":"7484_CR43","DOI":"10.1109\/CVPR.2017.700"},{"doi-asserted-by":"crossref","unstructured":"Yin, Z., Shi, J.: GeoNet: Unsupervised Learning of Dense Depth, Optical Flow and Camera Pose. In: Proceedings of the IEEE Conference on Computer vision and pattern recognition (CVPR), pp. 1983\u20131992 (2018)","key":"7484_CR44","DOI":"10.1109\/CVPR.2018.00212"},{"doi-asserted-by":"publisher","unstructured":"Bian, J.-W., Zhan, H., Wang, N., Li, Z., Zhang, L., Shen, C., Cheng, M.-M., Reid, I.: Unsupervised scale-consistent depth learning from video. Int J Comput Vis., 1\u201317 (2021). https:\/\/doi.org\/10.1007\/s11263-021-01484-6","key":"7484_CR45","DOI":"10.1007\/s11263-021-01484-6"},{"issue":"2","key":"7484_CR46","doi-asserted-by":"publisher","first-page":"207","DOI":"10.1177\/0278364913507326","volume":"33","author":"J-L Blanco-Claraco","year":"2014","unstructured":"Blanco-Claraco J-L, Moreno-Duenas F-A, Gonz\u00e1lez-Jim\u00e9nez J (2014) The m\u00e1laga urban dataset: High-rate stereo and lidar in a realistic urban scenario. Int J Rob Res 33(2):207\u2013214. https:\/\/doi.org\/10.1177\/0278364913507326","journal-title":"Int J Rob Res"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-022-07484-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-022-07484-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-022-07484-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,20]],"date-time":"2022-10-20T21:31:08Z","timestamp":1666301468000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-022-07484-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,24]]},"references-count":46,"journal-issue":{"issue":"21","published-print":{"date-parts":[[2022,11]]}},"alternative-id":["7484"],"URL":"https:\/\/doi.org\/10.1007\/s00521-022-07484-y","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"type":"print","value":"0941-0643"},{"type":"electronic","value":"1433-3058"}],"subject":[],"published":{"date-parts":[[2022,6,24]]},"assertion":[{"value":"20 December 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 May 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 June 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}