{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,24]],"date-time":"2025-06-24T06:30:02Z","timestamp":1750746602218,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":25,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819958467"},{"type":"electronic","value":"9789819958474"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-981-99-5847-4_4","type":"book-chapter","created":{"date-parts":[[2023,8,29]],"date-time":"2023-08-29T20:30:00Z","timestamp":1693341000000},"page":"46-59","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["A Lightweight Sensor Fusion for\u00a0Neural Visual Inertial Odometry"],"prefix":"10.1007","author":[{"given":"Yao","family":"Lu","sequence":"first","affiliation":[]},{"given":"Xiaoxu","family":"Yin","sequence":"additional","affiliation":[]},{"given":"Feng","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Ke","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Menghua","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Weijie","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,30]]},"reference":[{"issue":"49","key":"4_CR1","doi-asserted-by":"publisher","first-page":"15601","DOI":"10.1523\/JNEUROSCI.2574-09.2009","volume":"29","author":"CR Fetsch","year":"2009","unstructured":"Fetsch, C.R., Turner, A.H., DeAngelis, G.C., Angelaki, D.E.: Dynamic reweighting of visual and vestibular cues during self-motion perception. J. Neurosci. 29(49), 15601\u201315612 (2009)","journal-title":"J. Neurosci."},{"issue":"1","key":"4_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TRO.2016.2597321","volume":"33","author":"C Forster","year":"2017","unstructured":"Forster, C., Carlone, L., Dellaert, F., Scaramuzza, D.: Onmanifold preintegration for real-time visual Cinertial odometry. IEEE Trans. Rob. 33(1), 1\u201321 (2017)","journal-title":"IEEE Trans. Rob."},{"issue":"3","key":"4_CR3","doi-asserted-by":"publisher","first-page":"314","DOI":"10.1177\/0278364914554813","volume":"34","author":"S Leutenegger","year":"2015","unstructured":"Leutenegger, S., Lynen, S., Bosse, M., Siegwart, R., Furgale, P.: Keyframe-based visual Cinertial odometry using nonlinear optimization. Int. J. Robot. Res. 34(3), 314\u2013334 (2015)","journal-title":"Int. J. Robot. Res."},{"issue":"6","key":"4_CR4","doi-asserted-by":"publisher","first-page":"690","DOI":"10.1177\/0278364913481251","volume":"32","author":"M Li","year":"2013","unstructured":"Li, M., Mourikis, A.I.: High-precision, consistent EKF based visual-inertial odometry. Int. J. Robot. Res. 32(6), 690\u2013711 (2013)","journal-title":"Int. J. Robot. Res."},{"issue":"4","key":"4_CR5","doi-asserted-by":"publisher","first-page":"1004","DOI":"10.1109\/TRO.2018.2853729","volume":"34","author":"T Qin","year":"2018","unstructured":"Qin, T., Li, P., Shen, S.: VINS-MONO: a robust and versatile monocular visual-inertial state estimator. IEEE Trans. Rob. 34(4), 1004\u20131020 (2018)","journal-title":"IEEE Trans. Rob."},{"key":"4_CR6","doi-asserted-by":"crossref","unstructured":"Clark, R., Wang, S., Wen, H., Markham, A., Trigoni, N.: ViNet: visual-inertial odometry as a sequence-to-sequence learning problem. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31 (2017)","DOI":"10.1609\/aaai.v31i1.11215"},{"issue":"6","key":"4_CR7","doi-asserted-by":"publisher","first-page":"1309","DOI":"10.1109\/TRO.2016.2624754","volume":"32","author":"C Cadena","year":"2016","unstructured":"Cadena, C., et al.: Past, present, and future of simultaneous localization and mapping: toward the robust-perception age. IEEE Trans. Rob. 32(6), 1309\u20131332 (2016)","journal-title":"IEEE Trans. Rob."},{"issue":"3","key":"4_CR8","doi-asserted-by":"publisher","first-page":"611","DOI":"10.1109\/TPAMI.2017.2658577","volume":"40","author":"J Engel","year":"2017","unstructured":"Engel, J., Koltun, V., Cremers, D.: Direct sparse odometry. IEEE Trans. Pattern Anal. Mach. Intell. 40(3), 611\u2013625 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4_CR9","doi-asserted-by":"crossref","unstructured":"Mur-Artal, R., Tard\u00aes, J.D.: Orb-slam2: an open-source slam system for monocular, stereo, and RGB-D cameras. IEEE Trans. Robot. 33(5), 1255\u20131262 (2017)","DOI":"10.1109\/TRO.2017.2705103"},{"key":"4_CR10","doi-asserted-by":"crossref","unstructured":"Chen, C., Rosa, S., Miao, Y., et al.: Selective sensor fusion for neural visual-inertial odometry. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10542\u201310551 (2019)","DOI":"10.1109\/CVPR.2019.01079"},{"key":"4_CR11","doi-asserted-by":"crossref","unstructured":"Liu, L., Li, G., Li, T.H.: AtVio: attention guided visual-inertial odometry. In ICASSP 2021\u20132021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4125\u20134129. IEEE (2021)","DOI":"10.1109\/ICASSP39728.2021.9413912"},{"key":"4_CR12","doi-asserted-by":"crossref","unstructured":"Shamwell, E.J., Leung, S., Nothwang, W.D.: Vision-aided absolute trajectory estimation using an unsupervised deep network with online error correction. In: 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2524\u20132531. IEEE (2018)","DOI":"10.1109\/IROS.2018.8593573"},{"key":"4_CR13","doi-asserted-by":"crossref","unstructured":"Han, L., Lin, Y., Du, G., Lian, S.: Deepvio: self-supervised deep learning of monocular visual inertial odometry using 3D geometric constraints. In: 2019 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 6906\u20136913. IEEE (2019)","DOI":"10.1109\/IROS40897.2019.8968467"},{"key":"4_CR14","doi-asserted-by":"crossref","unstructured":"Almalioglu, Yasin, et al.: SelfVIO: self-supervised deep monocular Visual CInertial Odometry and depth estimation, pp. 119\u2013136. Neural Networks, 150 (2022)","DOI":"10.1016\/j.neunet.2022.03.005"},{"issue":"6","key":"4_CR15","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. Commun. ACM 60(6), 84\u201390 (2017)","journal-title":"Commun. ACM"},{"key":"4_CR16","unstructured":"Simonyan K, Zisserman A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"4_CR17","unstructured":"Ren, S., et al.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in Neural Information Processing Systems, vol. 28 (2015)"},{"key":"4_CR18","doi-asserted-by":"crossref","unstructured":"Yang, M., Chen, Y., Kim, H.S.: Efficient deep visual and inertial odometry with adaptive visual modality selection. In: Computer Vision CECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, pp. 233\u2013250. Proceedings, Part XXXVIII (2022)","DOI":"10.1007\/978-3-031-19839-7_14"},{"key":"4_CR19","doi-asserted-by":"crossref","unstructured":"Mourikis, A.I., Roumeliotis, S.I.: A multi-state constraint Kalman filter for vision-aided inertial navigation. In: Proceedings 2007 IEEE International Conference on Robotics and Automation, pp. 3565\u20133572. IEEE (2007)","DOI":"10.1109\/ROBOT.2007.364024"},{"key":"4_CR20","doi-asserted-by":"crossref","unstructured":"Bloesch, M., Omari, S., Hutter, M., Siegwart, R.: Robust visual inertial odometry using a direct EKF-based approach. In: 2015 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 298\u2013304. IEEE (2015)","DOI":"10.1109\/IROS.2015.7353389"},{"key":"4_CR21","doi-asserted-by":"crossref","unstructured":"Leutenegger, S., Furgale, P., Rabaud, V., et al.: Keyframe-based visual-inertial slam using nonlinear optimization. In: Proceedings of Robotis Science and Systems (RSS) 2013 (2013)","DOI":"10.15607\/RSS.2013.IX.037"},{"key":"4_CR22","doi-asserted-by":"crossref","unstructured":"Chen, C., et al.: Selective sensor fusion for neural visual-inertial odometry. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10542\u201310551 (2019)","DOI":"10.1109\/CVPR.2019.01079"},{"key":"4_CR23","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The kitti vision benchmark suite. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3354\u20133361. IEEE (2012)","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"4_CR24","doi-asserted-by":"crossref","unstructured":"Forster, C., Carlone, L., Dellaert, F., Scaramuzza, D.: IMU preintegration on manifold for efficient visual-inertial maximum-a-posteriori estimation. In: Robotics: Science and Systems XI (2015)","DOI":"10.15607\/RSS.2015.XI.006"},{"key":"4_CR25","doi-asserted-by":"crossref","unstructured":"Forster, C., Pizzoli, M., Scaramuzza, D.: SVO: fast semi-direct monocular visual odometry. In: 2014 IEEE International Conference on Robotics and Automation (ICRA), pp. 15\u201322. IEEE (2014)","DOI":"10.1109\/ICRA.2014.6906584"}],"container-title":["Communications in Computer and Information Science","International Conference on Neural Computing for Advanced Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-5847-4_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,8,29]],"date-time":"2023-08-29T20:30:57Z","timestamp":1693341057000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-5847-4_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9789819958467","9789819958474"],"references-count":25,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-5847-4_4","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"30 August 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"NCAA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Computing for Advanced Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hefei","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 July 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 July 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ncaa2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/dl2link.com\/ncaa2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easy chair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"211","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"83","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"39% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.67","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}