{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T05:33:43Z","timestamp":1768887223891,"version":"3.49.0"},"reference-count":25,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2022,6,6]],"date-time":"2022-06-06T00:00:00Z","timestamp":1654473600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,6,6]],"date-time":"2022-06-06T00:00:00Z","timestamp":1654473600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"name":"The National Key Research and Development Program of China","award":["2018YFB1307402"],"award-info":[{"award-number":["2018YFB1307402"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Intel Serv Robotics"],"published-print":{"date-parts":[[2022,7]]},"DOI":"10.1007\/s11370-022-00423-9","type":"journal-article","created":{"date-parts":[[2022,6,6]],"date-time":"2022-06-06T03:12:41Z","timestamp":1654485161000},"page":"363-379","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["An efficient loop closure detection method based on spatially constrained feature matching"],"prefix":"10.1007","volume":"15","author":[{"given":"Hong","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Tao","family":"Zhao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8046-1904","authenticated-orcid":false,"given":"Yuzhong","family":"Zhong","sequence":"additional","affiliation":[]},{"given":"Yanjie","family":"Yin","sequence":"additional","affiliation":[]},{"given":"Haobin","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Songyi","family":"Dian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,6,6]]},"reference":[{"key":"423_CR1","doi-asserted-by":"publisher","unstructured":"An S, Che G, Zhou F, Liu X, Ma X, Chen Y (2019) Fast and incremental loop closure detection using proximity graphs. https:\/\/doi.org\/10.1109\/IROS40897.2019.8968043","DOI":"10.1109\/IROS40897.2019.8968043"},{"key":"423_CR2","doi-asserted-by":"publisher","unstructured":"Arroyo R, Alcantarilla PF, Bergasa LM, Yebes JJ, G\u00e1mez S (2014) Bidirectional loop closure detection on panoramas for visual navigation. In: 2014 IEEE intelligent vehicles symposium proceedings. IEEE, pp 1378\u20131383 https:\/\/doi.org\/10.1109\/IVS.2014.6856457","DOI":"10.1109\/IVS.2014.6856457"},{"key":"423_CR3","doi-asserted-by":"publisher","unstructured":"Calonder M, Lepetit V, Strecha C, Fua P (2010) Brief: binary robust independent elementary features. In: European conference on computer vision. Springer, pp 778\u2013792 https:\/\/doi.org\/10.1007\/978-3-642-15561-1_56","DOI":"10.1007\/978-3-642-15561-1_56"},{"issue":"6","key":"423_CR4","doi-asserted-by":"publisher","first-page":"1120","DOI":"10.3390\/app9061120","volume":"9","author":"B Chen","year":"2019","unstructured":"Chen B, Yuan D, Liu C, Wu Q (2019) Loop closure detection based on multi-scale deep feature fusion. Appl Sci 9(6):1120. https:\/\/doi.org\/10.3390\/app9061120","journal-title":"Appl Sci"},{"issue":"6","key":"423_CR5","doi-asserted-by":"publisher","first-page":"647","DOI":"10.1177\/0278364908090961","volume":"27","author":"M Cummins","year":"2008","unstructured":"Cummins M, Newman P (2008) Fab-map: probabilistic localization and mapping in the space of appearance. Int J Robot Res 27(6):647\u2013665. https:\/\/doi.org\/10.1177\/0278364908090961","journal-title":"Int J Robot Res"},{"issue":"9","key":"423_CR6","doi-asserted-by":"publisher","first-page":"1100","DOI":"10.1177\/0278364910385483","volume":"30","author":"M Cummins","year":"2011","unstructured":"Cummins M, Newman P (2011) Appearance-only slam at large scale with fab-map 2.0. Int J Robot Res 30(9):1100\u20131123. https:\/\/doi.org\/10.1177\/0278364910385483","journal-title":"Int J Robot Res"},{"key":"423_CR7","doi-asserted-by":"publisher","unstructured":"Fritsch J, Kuehnl T, Geiger A (2013) A new performance measure and evaluation benchmark for road detection algorithms. In: 16th international ieee conference on intelligent transportation systems (ITSC 2013). IEEE, pp 1693\u20131700 https:\/\/doi.org\/10.1109\/ITSC.2013.6728473","DOI":"10.1109\/ITSC.2013.6728473"},{"issue":"1","key":"423_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10514-015-9516-2","volume":"41","author":"X Gao","year":"2017","unstructured":"Gao X, Zhang T (2017) Unsupervised learning to detect loops using deep neural networks for visual slam system. Auton Robot 41(1):1\u201318. https:\/\/doi.org\/10.1007\/s10514-015-9516-2","journal-title":"Auton Robot"},{"issue":"4","key":"423_CR9","doi-asserted-by":"publisher","first-page":"3051","DOI":"10.1109\/LRA.2018.2849609","volume":"3","author":"E Garcia-Fidalgo","year":"2018","unstructured":"Garcia-Fidalgo E, Ortiz A (2018) ibow-lcd: an appearance-based loop-closure detection approach using incremental bags of binary words. IEEE Robot Autom Lett 3(4):3051\u20133057. https:\/\/doi.org\/10.1109\/LRA.2018.2849609","journal-title":"IEEE Robot Autom Lett"},{"key":"423_CR10","doi-asserted-by":"publisher","unstructured":"Glover A, Maddern W, Warren M, Reid S, Milford M, Wyeth G (2012) Openfabmap: an open source toolbox for appearance-based loop closure detection. In: 2012 IEEE international conference on robotics and automation, pp 4730\u20134735. https:\/\/doi.org\/10.1109\/ICRA.2012.6224843","DOI":"10.1109\/ICRA.2012.6224843"},{"key":"423_CR11","doi-asserted-by":"publisher","unstructured":"G\u00e1lvez-L\u00f3pez D, Tard\u00f3s JD (2011) Real-time loop detection with bags of binary words. In: 2011 IEEE\/RSJ international conference on intelligent robots and systems, pp 51\u201358 https:\/\/doi.org\/10.1109\/IROS.2011.6094885","DOI":"10.1109\/IROS.2011.6094885"},{"key":"423_CR12","doi-asserted-by":"publisher","unstructured":"Han J, Dong R, Kan J (2020) A novel loop closure detection method with the combination of points and lines based on information entropy. J Field Robot. https:\/\/doi.org\/10.1002\/rob.21992","DOI":"10.1002\/rob.21992"},{"key":"423_CR13","doi-asserted-by":"publisher","unstructured":"Hou Y, Zhang H, Zhou S (2015) Convolutional neural network-based image representation for visual loop closure detection. In: 2015 IEEE international conference on information and automation. IEEE, pp 2238\u20132245. https:\/\/doi.org\/10.1109\/ICInfA.2015.7279659","DOI":"10.1109\/ICInfA.2015.7279659"},{"key":"423_CR14","doi-asserted-by":"publisher","unstructured":"Khan S, Wollherr D (2015) Ibuild: Incremental bag of binary words for appearance based loop closure detection. In: 2015 IEEE international conference on robotics and automation (ICRA). IEEE, pp 5441\u20135447 https:\/\/doi.org\/10.1109\/ICRA.2015.7139959","DOI":"10.1109\/ICRA.2015.7139959"},{"key":"423_CR15","doi-asserted-by":"publisher","unstructured":"Liu Y, Zhang H (2012) Visual loop closure detection with a compact image descriptor. In: 2012 IEEE\/RSJ international conference on intelligent robots and systems. IEEE, pp 1051\u20131056 https:\/\/doi.org\/10.1109\/IROS.2012.6386145","DOI":"10.1109\/IROS.2012.6386145"},{"issue":"1","key":"423_CR16","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TRO.2015.2496823","volume":"32","author":"S Lowry","year":"2015","unstructured":"Lowry S, S\u00fcnderhauf N, Newman P, Leonard JJ, Cox D, Corke P, Milford MJ (2015) Visual place recognition: a survey. IEEE Trans Robot 32(1):1\u201319. https:\/\/doi.org\/10.1109\/TRO.2015.2496823","journal-title":"IEEE Trans Robot"},{"key":"423_CR17","doi-asserted-by":"publisher","unstructured":"Milford MJ, Wyeth GF (2012) Seqslam: visual route-based navigation for sunny summer days and stormy winter nights. In: 2012 IEEE international conference on robotics and automation. IEEE, pp 1643\u20131649 https:\/\/doi.org\/10.1109\/ICRA.2012.6224623","DOI":"10.1109\/ICRA.2012.6224623"},{"issue":"5","key":"423_CR18","doi-asserted-by":"publisher","first-page":"1255","DOI":"10.1109\/TRO.2017.2705103","volume":"33","author":"R Mur-Artal","year":"2017","unstructured":"Mur-Artal R, Tard\u00f3s JD (2017) Orb-slam2: an open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Trans Robot 33(5):1255\u20131262. https:\/\/doi.org\/10.1109\/TRO.2017.2705103","journal-title":"IEEE Trans Robot"},{"issue":"4","key":"423_CR19","doi-asserted-by":"publisher","first-page":"886","DOI":"10.1109\/TRO.2012.2192013","volume":"28","author":"T Nicosevici","year":"2012","unstructured":"Nicosevici T, Garcia R (2012) Automatic visual bag-of-words for online robot navigation and mapping. IEEE Trans Robot 28(4):886\u2013898. https:\/\/doi.org\/10.1109\/TRO.2012.2192013","journal-title":"IEEE Trans Robot"},{"key":"423_CR20","unstructured":"rmsalinas (2017) DBoW3 dbow3. https:\/\/github.com\/rmsalinas\/DBow3"},{"issue":"4","key":"423_CR21","doi-asserted-by":"publisher","first-page":"861","DOI":"10.1109\/TRO.2009.2022424","volume":"25","author":"C Siagian","year":"2009","unstructured":"Siagian C, Itti L (2009) Biologically inspired mobile robot vision localization. IEEE Trans Robot 25(4):861\u2013873. https:\/\/doi.org\/10.1109\/TRO.2009.2022424","journal-title":"IEEE Trans Robot"},{"key":"423_CR22","doi-asserted-by":"publisher","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. https:\/\/doi.org\/10.1109\/SLT.2016.7846307","DOI":"10.1109\/SLT.2016.7846307"},{"issue":"2","key":"423_CR23","doi-asserted-by":"publisher","first-page":"1737","DOI":"10.1109\/LRA.2019.2897151","volume":"4","author":"KA Tsintotas","year":"2019","unstructured":"Tsintotas KA, Bampis L, Gasteratos A (2019) Probabilistic appearance-based place recognition through bag of tracked words. IEEE Robot Autom Lett 4(2):1737\u20131744. https:\/\/doi.org\/10.1109\/LRA.2019.2897151","journal-title":"IEEE Robot Autom Lett"},{"key":"423_CR24","doi-asserted-by":"publisher","unstructured":"Viswanathan DG (2009) Features from accelerated segment test (fast). In: Proceedings of the 10th workshop on image analysis for multimedia interactive services, London, UK, pp 6\u20138. https:\/\/doi.org\/10.1109\/SLT.2016.7846307","DOI":"10.1109\/SLT.2016.7846307"},{"issue":"2","key":"423_CR25","doi-asserted-by":"publisher","first-page":"389","DOI":"10.1007\/s10846-018-0917-2","volume":"95","author":"X Zhang","year":"2019","unstructured":"Zhang X, Wang L, Zhao Y, Su Y (2019) Graph-based place recognition in image sequences with cnn features. J Intell Robot Syst 95(2):389\u2013403","journal-title":"J Intell Robot Syst"}],"container-title":["Intelligent Service Robotics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11370-022-00423-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11370-022-00423-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11370-022-00423-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,13]],"date-time":"2022-07-13T08:35:25Z","timestamp":1657701325000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11370-022-00423-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,6]]},"references-count":25,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2022,7]]}},"alternative-id":["423"],"URL":"https:\/\/doi.org\/10.1007\/s11370-022-00423-9","relation":{},"ISSN":["1861-2776","1861-2784"],"issn-type":[{"value":"1861-2776","type":"print"},{"value":"1861-2784","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,6,6]]},"assertion":[{"value":"4 September 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 March 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 June 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Not applicable.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}]}}