{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T09:35:51Z","timestamp":1774604151025,"version":"3.50.1"},"reference-count":15,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,9,23]],"date-time":"2024-09-23T00:00:00Z","timestamp":1727049600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,23]],"date-time":"2024-09-23T00:00:00Z","timestamp":1727049600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["41674036"],"award-info":[{"award-number":["41674036"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"the Science and Technology Development Fund of Wuxi","award":["N20201011"],"award-info":[{"award-number":["N20201011"]}]},{"DOI":"10.13039\/501100004608","name":"Natural Science Foundation of Jiangsu Province","doi-asserted-by":"publisher","award":["BK20211037"],"award-info":[{"award-number":["BK20211037"]}],"id":[{"id":"10.13039\/501100004608","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1007\/s00530-024-01482-x","type":"journal-article","created":{"date-parts":[[2024,9,23]],"date-time":"2024-09-23T17:02:04Z","timestamp":1727110924000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Research on tightly coupled visual inertial SLAM based on point and line features"],"prefix":"10.1007","volume":"30","author":[{"given":"Jiajia","family":"Lu","sequence":"first","affiliation":[]},{"given":"Yue","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Qilin","family":"Hang","sequence":"additional","affiliation":[]},{"given":"Yue","family":"Tang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,23]]},"reference":[{"key":"1482_CR1","doi-asserted-by":"crossref","unstructured":"Lu, Y., Song, D.: Robust RGB-D odometry using point and line features [C]. In: 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, pp. 3934\u20133942. IEEE (2015)","DOI":"10.1109\/ICCV.2015.448"},{"key":"1482_CR2","doi-asserted-by":"crossref","unstructured":"Gomez-Ojeda, R., Gonzalez-Jimenez, J.: Robust stereo visual odometry through a probabilistic combination of points and line segments [C]. In: 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, pp. 2521\u20132526. IEEE (2016)","DOI":"10.1109\/ICRA.2016.7487406"},{"key":"1482_CR3","volume-title":"A binocular vision SLAM method based on integrated point and line features [D]","author":"XJ Xie","year":"2017","unstructured":"Xie, X.J.: A binocular vision SLAM method based on integrated point and line features [D]. Zhejiang University, Hangzhou (2017)"},{"issue":"3","key":"1482_CR4","doi-asserted-by":"publisher","first-page":"734","DOI":"10.1109\/TRO.2019.2899783","volume":"35","author":"R Gomez-Ojeda","year":"2019","unstructured":"Gomez-Ojeda, R., Moreno, F.A., Zuniga-noel, D., et al.: PL-SLAM: a stereo SLAM system through the combination of points and line segments [J\/OL]. IEEE Trans. Rob. 35(3), 734\u2013746 (2019)","journal-title":"IEEE Trans. Rob."},{"key":"1482_CR5","doi-asserted-by":"crossref","unstructured":"Weiss, S., Siegwart, R.: Real-time metric state estimation for modular vision-inertial systems [C]. In: 2011 IEEE International Conference on Robotics and Automation (ICRA), Shanghai, China, pp. 4531\u20134537. IEEE (2011)","DOI":"10.1109\/ICRA.2011.5979982"},{"issue":"3","key":"1482_CR6","doi-asserted-by":"publisher","first-page":"314","DOI":"10.1177\/0278364914554813","volume":"34","author":"S Leutenegger","year":"2015","unstructured":"Leutenegger, S., Lynen, S., Bosse, M., et al.: Keyframe-based visual\u2013inertial odometry using nonlinear optimization [J]. Int. J. Robot. Res. 34(3), 314\u2013334 (2015)","journal-title":"Int. J. Robot. Res."},{"issue":"4","key":"1482_CR7","doi-asserted-by":"publisher","first-page":"1159","DOI":"10.3390\/s18041159","volume":"18","author":"Y He","year":"2018","unstructured":"He, Y., Zhao, J., Guo, Y., et al.: Pl-vio: tightly-coupled monocular visual\u2013inertial odometry using point and line features [J]. Sensors 18(4), 1159 (2018)","journal-title":"Sensors"},{"issue":"6","key":"1482_CR8","doi-asserted-by":"publisher","first-page":"1874","DOI":"10.1109\/TRO.2021.3075644","volume":"37","author":"C Campos","year":"2021","unstructured":"Campos, C., Elvira, R., Rodr\u00edguez, J.J.G., et al.: Orb-slam3: An accurate open-source library for visual, visual\u2013inertial, and multimap slam [J]. IEEE Trans. Rob. 37(6), 1874\u20131890 (2021)","journal-title":"IEEE Trans. Rob."},{"issue":"2","key":"1482_CR9","doi-asserted-by":"publisher","first-page":"3483","DOI":"10.1109\/LRA.2022.3146893","volume":"7","author":"B Xu","year":"2022","unstructured":"Xu, B., Wang, P., He, Y., et al.: Leveraging structural information to improve point line visual-inertial odometry [J]. IEEE Robot. Autom. Lett. 7(2), 3483\u20133490 (2022)","journal-title":"IEEE Robot. Autom. Lett."},{"issue":"4","key":"1482_CR10","doi-asserted-by":"publisher","first-page":"7033","DOI":"10.1109\/LRA.2021.3095518","volume":"6","author":"J Lee","year":"2021","unstructured":"Lee, J., Park, S.Y.: PLF-VINS: real-time monocular visual-inertial SLAM with point-line fusion and parallel-line fusion[J]. IEEE Robot. Autom. Lett. 6(4), 7033\u20137040 (2021)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"1482_CR11","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1016\/j.mechatronics.2017.12.002","volume":"49","author":"MS Bahraini","year":"2018","unstructured":"Bahraini, M.S., Bozorg, M., Rad, A.B.: SLAM in dynamic environments via ML-RANSAC [J]. Mechatronics 49, 105\u2013118 (2018)","journal-title":"Mechatronics"},{"key":"1482_CR12","doi-asserted-by":"crossref","unstructured":"Mair, E., Fleps, M., Suppa, M., et al.: Spatio-temporal initialization for IMU to camera registration [C]. In: 2011 IEEE International Conference on Robotics and Biomimetics, Karon Beach, Thailand, pp. 557\u2013564. IEEE (2012)","DOI":"10.1109\/ROBIO.2011.6181345"},{"key":"1482_CR13","doi-asserted-by":"publisher","first-page":"185408","DOI":"10.1109\/ACCESS.2019.2961266","volume":"7","author":"C Chen","year":"2019","unstructured":"Chen, C., Zhu, H., Wang, L., et al.: A stereo visual-inertial SLAM approach for indoor mobile robots in unknown environments without occlusions [J]. IEEE Access 7, 185408\u2013185421 (2019)","journal-title":"IEEE Access"},{"issue":"6","key":"1482_CR14","doi-asserted-by":"publisher","first-page":"5799","DOI":"10.1109\/TVT.2022.3162155","volume":"71","author":"Z Yu","year":"2022","unstructured":"Yu, Z., Zhu, L., Lu, G.: Tightly-coupled fusion of VINS and motion constraint for autonomous vehicle [J]. IEEE Trans. Veh. Technol. 71(6), 5799\u20135810 (2022)","journal-title":"IEEE Trans. Veh. Technol."},{"key":"1482_CR15","volume-title":"Study of tightly coupled nonlinear optimization methods for visual inertial odometers [D]","author":"QG Su","year":"2021","unstructured":"Su, Q.G.: Study of tightly coupled nonlinear optimization methods for visual inertial odometers [D]. Chongqing University, Chongqing (2021)"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01482-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01482-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01482-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T18:16:39Z","timestamp":1730139399000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01482-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,23]]},"references-count":15,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["1482"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01482-x","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,9,23]]},"assertion":[{"value":"11 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 September 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 September 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"All authors of this manuscript have carefully studied the submission guidelines and have unanimously agreed to publish.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}],"article-number":"286"}}