{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,17]],"date-time":"2026-04-17T16:19:05Z","timestamp":1776442745442,"version":"3.51.2"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,4,30]],"date-time":"2024-04-30T00:00:00Z","timestamp":1714435200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,4,30]],"date-time":"2024-04-30T00:00:00Z","timestamp":1714435200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U20A20220"],"award-info":[{"award-number":["U20A20220"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100018694","name":"HORIZON EUROPE Marie Sklodowska-Curie Actions","doi-asserted-by":"publisher","award":["778062"],"award-info":[{"award-number":["778062"]}],"id":[{"id":"10.13039\/100018694","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2024,8]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>With the development of deep learning, a higher level of perception of the environment such as the semantic level can be achieved in the simultaneous localization and mapping (SLAM) domain. However, previous works did not achieve a natural-language level of perception. Therefore, LP-SLAM (Language-Perceptive RGB-D SLAM) is proposed that leverages large language models (LLMs). The texts in the scene can be detected by scene text recognition (STR) and mapped as landmarks with a task-driven selection. A text error correction chain (TECC) is designed with a similarity classification method, a two-stage memory strategy, and a text clustering method. The proposed architecture is designed to deal with the mis-detection and mis-recognition cases of STR and to provide accurate text information to the framework. The proposed framework takes input images and generates a 3D map with sparse point cloud and task-related texts. Finally, a natural user interface (NUI) is designed based on the constructed map and LLM, which gives position instructions based on users\u2019 natural queries. The experimental results validated the proposed TECC design and the overall framework. We publish the virtual dataset with ground truth, as well as the source code for further research. <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" ext-link-type=\"uri\" xlink:href=\"https:\/\/github.com\/GroupOfLPSLAM\/LP_SLAM\">https:\/\/github.com\/GroupOfLPSLAM\/LP_SLAM<\/jats:ext-link>.<\/jats:p>","DOI":"10.1007\/s40747-024-01408-0","type":"journal-article","created":{"date-parts":[[2024,5,1]],"date-time":"2024-05-01T12:51:58Z","timestamp":1714567918000},"page":"5391-5409","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Lp-slam: language-perceptive RGB-D SLAM framework exploiting large language model"],"prefix":"10.1007","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1296-3579","authenticated-orcid":false,"given":"Weiyi","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Yushi","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Liting","family":"Niu","sequence":"additional","affiliation":[]},{"given":"Peijun","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zeyu","family":"Wan","sequence":"additional","affiliation":[]},{"given":"Fei","family":"Shao","sequence":"additional","affiliation":[]},{"given":"Cheng","family":"Nian","sequence":"additional","affiliation":[]},{"given":"Fasih Ud Din","family":"Farrukh","sequence":"additional","affiliation":[]},{"given":"Debing","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Chun","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Qiang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jianwei","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,30]]},"reference":[{"issue":"2","key":"1408_CR1","doi-asserted-by":"publisher","first-page":"2000","DOI":"10.1109\/LRA.2022.3142739","volume":"7","author":"K Chen","year":"2022","unstructured":"Chen K, Lopez BT, Agha-mohammadi A-A, Mehta A (2022) Direct lidar odometry: Fast localization with dense point clouds. IEEE Robot Autom Lett 7(2):2000\u20132007","journal-title":"IEEE Robot Autom Lett"},{"issue":"5","key":"1408_CR2","doi-asserted-by":"publisher","first-page":"1255","DOI":"10.1109\/TRO.2017.2705103","volume":"33","author":"R Mur-Artal","year":"2017","unstructured":"Mur-Artal R, Tard\u00f3s JD (2017) Orb-slam2: an open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Trans Robot 33(5):1255\u20131262","journal-title":"IEEE Trans Robot"},{"issue":"3","key":"1408_CR3","doi-asserted-by":"publisher","first-page":"734","DOI":"10.1109\/TRO.2013.2242375","volume":"29","author":"M Labbe","year":"2013","unstructured":"Labbe M, Michaud F (2013) Appearance-based loop closure detection for online large-scale and long-term operation. IEEE Trans Robot 29(3):734\u2013745","journal-title":"IEEE Trans Robot"},{"issue":"1","key":"1408_CR4","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1109\/TRO.2013.2279412","volume":"30","author":"F Endres","year":"2013","unstructured":"Endres F, Hess J, Sturm J, Cremers D, Burgard W (2013) 3-d mapping with an rgb-d camera. IEEE Trans Robot 30(1):177\u2013187","journal-title":"IEEE Trans Robot"},{"key":"1408_CR5","doi-asserted-by":"crossref","unstructured":"Harris C, Stephens M et\u00a0al (1988) A combined corner and edge detector. In: Alvey Vision Conference, vol 15, pp 10\u20135244","DOI":"10.5244\/C.2.23"},{"key":"1408_CR6","unstructured":"Shi J et\u00a0al (1994) Good features to track. In: 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp 593\u2013600"},{"key":"1408_CR7","doi-asserted-by":"crossref","unstructured":"Rublee E, Rabaud V, Konolige K, Bradski G (2011) Orb: an efficient alternative to sift or surf. In: 2011 International Conference on Computer Vision, pp 2564\u20132571","DOI":"10.1109\/ICCV.2011.6126544"},{"key":"1408_CR8","doi-asserted-by":"crossref","unstructured":"Hirose K, Saito H (2012) Fast line description for line-based slam. In: BMVC, pp 1\u201311","DOI":"10.5244\/C.26.83"},{"key":"1408_CR9","doi-asserted-by":"crossref","unstructured":"Zuo X, Xie X, Liu Y, Huang G (2017) Robust visual slam with point and line features. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp 1775\u20131782","DOI":"10.1109\/IROS.2017.8205991"},{"key":"1408_CR10","doi-asserted-by":"crossref","unstructured":"Yang S, Song Y, Kaess M, Scherer S (2016) Pop-up slam: semantic monocular plane slam for low-texture environments. In: 2016 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp 1222\u20131229","DOI":"10.1109\/IROS.2016.7759204"},{"issue":"4","key":"1408_CR11","doi-asserted-by":"publisher","first-page":"3145","DOI":"10.1109\/LRA.2019.2924848","volume":"4","author":"S Yang","year":"2019","unstructured":"Yang S, Scherer S (2019) Monocular object and plane slam in structured environments. IEEE Robot Autom Lett 4(4):3145\u20133152","journal-title":"IEEE Robot Autom Lett"},{"key":"1408_CR12","doi-asserted-by":"crossref","unstructured":"Liu B, Zhao W, Sun Q (2017) Study of object detection based on faster r-cnn. In: 2017 Chinese Automation Congress (CAC), pp 6233\u20136236","DOI":"10.1109\/CAC.2017.8243900"},{"key":"1408_CR13","doi-asserted-by":"crossref","unstructured":"He K, Gkioxari G, Doll\u00e1r P, Girshick R (2017) Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp 2961\u20132969","DOI":"10.1109\/ICCV.2017.322"},{"key":"1408_CR14","doi-asserted-by":"crossref","unstructured":"S\u00fcnderhauf N, Pham TT, Latif Y, Milford M, Reid I (2017) Meaningful maps with object-oriented semantic mapping. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5079\u20135085","DOI":"10.1109\/IROS.2017.8206392"},{"key":"1408_CR15","unstructured":"Liu M, Zhu M, White M, Li Y, Kalenichenko D (2019) Looking fast and slow: memory-guided mobile video object detection. arXiv preprint arXiv:1903.10172"},{"key":"1408_CR16","doi-asserted-by":"crossref","unstructured":"Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4510\u20134520","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1408_CR17","doi-asserted-by":"crossref","unstructured":"Wu W, Guo L, Gao H, You Z, Liu Y, Chen Z (2022) Yolo-slam: A semantic slam system towards dynamic environment with geometric constraint. Neural Comput Appl 1\u201316","DOI":"10.1007\/s00521-021-06764-3"},{"issue":"4","key":"1408_CR18","doi-asserted-by":"publisher","first-page":"925","DOI":"10.1109\/TRO.2019.2909168","volume":"35","author":"S Yang","year":"2019","unstructured":"Yang S, Scherer S (2019) Cubeslam: monocular 3-d object slam. IEEE Trans Robot 35(4):925\u2013938","journal-title":"IEEE Trans Robot"},{"key":"1408_CR19","doi-asserted-by":"crossref","unstructured":"Yu C, Liu Z, Liu X-J, Xie F, Yang Y, Wei Q, Fei Q (2018) Ds-slam: a semantic visual slam towards dynamic environments. In: 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp 1168\u20131174","DOI":"10.1109\/IROS.2018.8593691"},{"key":"1408_CR20","doi-asserted-by":"crossref","unstructured":"Bao SY, Bagra M, Chao Y-W, Savarese S (2012) Semantic structure from motion with points, regions, and objects. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp 2703\u20132710","DOI":"10.1109\/CVPR.2012.6247992"},{"issue":"2","key":"1408_CR21","doi-asserted-by":"publisher","first-page":"2455","DOI":"10.1109\/LRA.2022.3145066","volume":"7","author":"Z Qian","year":"2022","unstructured":"Qian Z, Fu J, Xiao J (2022) Towards accurate loop closure detection in semantic slam with 3d semantic covisibility graphs. IEEE Robot Autom Lett 7(2):2455\u20132462","journal-title":"IEEE Robot Autom Lett"},{"key":"1408_CR22","doi-asserted-by":"crossref","unstructured":"Rong X, Li B, Munoz JP, Xiao J, Arditi A, Tian Y (2016) Guided text spotting for assistive blind navigation in unfamiliar indoor environments. In: Advances in Visual Computing: 12th International Symposium, ISVC 2016, Las Vegas, NV, USA, December 12-14, 2016, Proceedings, Part II 12, Springer, pp 11\u201322","DOI":"10.1007\/978-3-319-50832-0_2"},{"key":"1408_CR23","doi-asserted-by":"crossref","unstructured":"Wang H-C, Landa Y, Fallon M, Teller S (2014) Spatially prioritized and persistent text detection and decoding. In: camera-based document analysis and recognition: 5th International Workshop, CBDAR 2013, Washington, DC, USA, August 23, 2013, Revised Selected Papers 5, Springer, pp 3\u201317","DOI":"10.1007\/978-3-319-05167-3_1"},{"key":"1408_CR24","doi-asserted-by":"crossref","unstructured":"Wang H-C, Finn C, Paull L, Kaess M, Rosenholtz R, Teller S, Leonard J (2015) Bridging text spotting and slam with junction features. In: 2015 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp 3701\u20133708","DOI":"10.1109\/IROS.2015.7353895"},{"key":"1408_CR25","doi-asserted-by":"crossref","unstructured":"Li B, Zou D, Sartori D, Pei L, Yu W (2020) Textslam: visual slam with planar text features. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp 2102\u20132108","DOI":"10.1109\/ICRA40945.2020.9197233"},{"key":"1408_CR26","doi-asserted-by":"crossref","unstructured":"Tong Q, Ma L, Chen K, Liu J, Qian J, Zhang J (2022) Txslam: a monocular semantic slam tightly coupled with planar text features. In: 2022 IEEE 25th International Conference on Computer Supported Cooperative Work in Design (CSCWD), pp 728\u2013733","DOI":"10.1109\/CSCWD54268.2022.9776293"},{"key":"1408_CR27","unstructured":"Scao TL, Fan A, Akiki C, Pavlick E, Ili\u0107 S, Hesslow D, Castagn\u00e9 R. Luccioni AS, Yvon F, Gall\u00e9 M et al (2022) Bloom: a 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100"},{"key":"1408_CR28","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A et al (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877\u20131901","journal-title":"Adv Neural Inf Process Syst"},{"key":"1408_CR29","first-page":"20","volume":"2","author":"S Vemprala","year":"2023","unstructured":"Vemprala S, Bonatti R, Bucker A, Kapoor A (2023) Chatgpt for robotics: design principles and model abilities. Microsoft Auton Syst Robot Res 2:20","journal-title":"Microsoft Auton Syst Robot Res"},{"issue":"6","key":"1408_CR30","doi-asserted-by":"publisher","first-page":"1052","DOI":"10.1109\/TPAMI.2007.1049","volume":"29","author":"AJ Davison","year":"2007","unstructured":"Davison AJ, Reid ID, Molton ND, Stasse O (2007) Monoslam: Real-time single camera slam. IEEE Trans Pattern Anal Mach Intell 29(6):1052\u20131067","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1408_CR31","doi-asserted-by":"crossref","unstructured":"Klein G, Murray D (2007) Parallel tracking and mapping for small ar workspaces. In: 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality, pp 225\u2013234","DOI":"10.1109\/ISMAR.2007.4538852"},{"issue":"5","key":"1408_CR32","doi-asserted-by":"publisher","first-page":"1147","DOI":"10.1109\/TRO.2015.2463671","volume":"31","author":"R Mur-Artal","year":"2015","unstructured":"Mur-Artal R, Montiel JMM, Tardos JD (2015) Orb-slam: a versatile and accurate monocular slam system. IEEE Trans Robot 31(5):1147\u20131163","journal-title":"IEEE Trans Robot"},{"issue":"6","key":"1408_CR33","doi-asserted-by":"publisher","first-page":"1874","DOI":"10.1109\/TRO.2021.3075644","volume":"37","author":"C Campos","year":"2021","unstructured":"Campos C, Elvira R, Rodr\u00edguez JJG, Montiel JM, Tard\u00f3s JD (2021) Orb-slam3: An accurate open-source library for visual, visual-inertial, and multimap slam. IEEE Trans Robot 37(6):1874\u20131890","journal-title":"IEEE Trans Robot"},{"key":"1408_CR34","doi-asserted-by":"crossref","unstructured":"Engel J, Sch\u00f6ps T, Cremers D (2014) Lsd-slam: Large-scale direct monocular slam. In: European Conference on Computer Vision, Springer, pp 834\u2013849","DOI":"10.1007\/978-3-319-10605-2_54"},{"issue":"4","key":"1408_CR35","doi-asserted-by":"publisher","first-page":"1363","DOI":"10.1109\/TRO.2020.2991614","volume":"36","author":"J Zubizarreta","year":"2020","unstructured":"Zubizarreta J, Aguinaga I, Montiel JMM (2020) Direct sparse mapping. IEEE Trans Robot 36(4):1363\u20131370","journal-title":"IEEE Trans Robot"},{"key":"1408_CR36","doi-asserted-by":"crossref","unstructured":"DeTone D, Malisiewicz T, Rabinovich A (2018) Superpoint: Self-supervised interest point detection and description. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp 224\u2013236","DOI":"10.1109\/CVPRW.2018.00060"},{"key":"1408_CR37","doi-asserted-by":"crossref","unstructured":"Sarlin P-E, DeTone D, Malisiewicz T, Rabinovich A (2020) Superglue: learning feature matching with graph neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 4938\u20134947","DOI":"10.1109\/CVPR42600.2020.00499"},{"issue":"11","key":"1408_CR38","doi-asserted-by":"publisher","first-page":"2298","DOI":"10.1109\/TPAMI.2016.2646371","volume":"39","author":"B Shi","year":"2016","unstructured":"Shi B, Bai X, Yao C (2016) An end-to-end trainable neural network for image-based sequence recognition and its application to scene text recognition. IEEE Trans Pattern Anal Mach Intell 39(11):2298\u20132304","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"9","key":"1408_CR39","doi-asserted-by":"publisher","first-page":"2035","DOI":"10.1109\/TPAMI.2018.2848939","volume":"41","author":"B Shi","year":"2018","unstructured":"Shi B, Yang M, Wang X, Lyu P, Yao C, Bai X (2018) Aster: an attentional scene text recognizer with flexible rectification. IEEE Trans Pattern Anal Mach Intell 41(9):2035\u20132048","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1408_CR40","doi-asserted-by":"crossref","unstructured":"Yu D, Li X, Zhang C, Liu T, Han J, Liu J, Ding E (2020) Towards accurate scene text recognition with semantic reasoning networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 12113\u201312122","DOI":"10.1109\/CVPR42600.2020.01213"},{"issue":"7","key":"1408_CR41","doi-asserted-by":"publisher","first-page":"1001","DOI":"10.1109\/29.32278","volume":"37","author":"LR Bahl","year":"1989","unstructured":"Bahl LR, Brown PF, Souza PV, Mercer RL (1989) A tree-based statistical language model for natural language speech recognition. IEEE Trans Acoust Speech Signal Process 37(7):1001\u20131008","journal-title":"IEEE Trans Acoust Speech Signal Process"},{"issue":"4","key":"1408_CR42","first-page":"467","volume":"18","author":"PF Brown","year":"1992","unstructured":"Brown PF, Della Pietra VJ, Desouza PV, Lai JC, Mercer RL (1992) Class-based n-gram models of natural language. Comput Linguist 18(4):467\u2013480","journal-title":"Comput Linguist"},{"key":"1408_CR43","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser \u0141, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst 30"},{"key":"1408_CR44","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1146\/annurev-control-101119-071628","volume":"3","author":"S Tellex","year":"2020","unstructured":"Tellex S, Gopalan N, Kress-Gazit H, Matuszek C (2020) Robots that use language. Ann Rev Control Robot Auton Syst 3:25\u201355","journal-title":"Ann Rev Control Robot Auton Syst"},{"key":"1408_CR45","doi-asserted-by":"crossref","unstructured":"Pan J, Chou G, Berenson D (2023) Data-efficient learning of natural language to linear temporal logic translators for robot task specification. arXiv preprint arXiv:2303.08006","DOI":"10.1109\/ICRA48891.2023.10161125"},{"key":"1408_CR46","doi-asserted-by":"crossref","unstructured":"Mees O, Borja-Diaz J, Burgard W (2023) Grounding language with visual affordances over unstructured data. In: 2023 IEEE International Conference on Robotics and Automation (ICRA), pp 11576\u201311582","DOI":"10.1109\/ICRA48891.2023.10160396"},{"key":"1408_CR47","doi-asserted-by":"crossref","unstructured":"Chen B, Xia F, Ichter B, Rao K, Gopalakrishnan K, Ryoo MS, Stone A, Kappler D (2023) Open-vocabulary queryable scene representations for real world planning. In: 2023 IEEE International Conference on Robotics and Automation (ICRA), pp 11509\u201311522","DOI":"10.1109\/ICRA48891.2023.10161534"},{"key":"1408_CR48","doi-asserted-by":"crossref","unstructured":"Singh I, Blukis V, Mousavian A, Goyal A, Xu D, Tremblay J, Fox D, Thomason J, Garg A (2023) Progprompt: generating situated robot task plans using large language models. In: 2023 IEEE International Conference on Robotics and Automation (ICRA), pp 11523\u201311530","DOI":"10.1109\/ICRA48891.2023.10161317"},{"key":"1408_CR49","doi-asserted-by":"crossref","unstructured":"Wang W, Li X, Dong Y, Xie J, Guo D, Liu H (2023) Natural language instruction understanding for robotic manipulation: a multisensory perception approach. In: 2023 IEEE International Conference on Robotics and Automation (ICRA), pp 9800\u20139806","DOI":"10.1109\/ICRA48891.2023.10160906"},{"key":"1408_CR50","doi-asserted-by":"crossref","unstructured":"Mo Y, Zhang H, Kong T (2023) Towards open-world interactive disambiguation for robotic grasping. In: 2023 IEEE International Conference on Robotics and Automation (ICRA), pp 8061\u20138067","DOI":"10.1109\/ICRA48891.2023.10161333"},{"key":"1408_CR51","unstructured":"Levenshtein VI et\u00a0al (1966) Binary codes capable of correcting deletions, insertions, and reversals. In: Soviet Physics Doklady, vol 10, Soviet Union, pp 707\u2013710"},{"key":"1408_CR52","unstructured":"Du Y, Li C, Guo R, Cui C, Liu W, Zhou J, Lu B, Yang Y, Liu Q, Hu X et al (2021) Pp-ocrv2: Bag of tricks for ultra lightweight ocr system. arXiv preprint arXiv:2109.03144"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01408-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-024-01408-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01408-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,17]],"date-time":"2024-07-17T17:23:29Z","timestamp":1721237009000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-024-01408-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,30]]},"references-count":52,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["1408"],"URL":"https:\/\/doi.org\/10.1007\/s40747-024-01408-0","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,4,30]]},"assertion":[{"value":"7 December 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 March 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 April 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All authors have no conflicts of interest to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"All authors declare that this work does not include humans and animals, as well as did not collect data from human subjects.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}