{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T10:09:17Z","timestamp":1764842957866,"version":"3.35.0"},"reference-count":50,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2025,1,2]],"date-time":"2025-01-02T00:00:00Z","timestamp":1735776000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,2]],"date-time":"2025-01-02T00:00:00Z","timestamp":1735776000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"National Science and Technology Innovation 2030 of China-New Generation Artificial Intelligence","award":["2022ZD0115603"],"award-info":[{"award-number":["2022ZD0115603"]}]},{"DOI":"10.13039\/501100013058","name":"Jiangsu Provincial Key Research and Development Program","doi-asserted-by":"publisher","award":["BE2022053-5"],"award-info":[{"award-number":["BE2022053-5"]}],"id":[{"id":"10.13039\/501100013058","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2025,2]]},"DOI":"10.1007\/s10489-024-06054-0","type":"journal-article","created":{"date-parts":[[2025,1,3]],"date-time":"2025-01-03T01:27:58Z","timestamp":1735867678000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Mapless autonomous navigation for UGV in cluttered off-road environment with the guidance of wayshowers using deep reinforcement learning"],"prefix":"10.1007","volume":"55","author":[{"given":"Zhijian","family":"Li","sequence":"first","affiliation":[]},{"given":"Xu","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jinchao","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Xixiang","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,2]]},"reference":[{"issue":"6","key":"6054_CR1","doi-asserted-by":"publisher","first-page":"4909","DOI":"10.1109\/TITS.2021.3054625","volume":"23","author":"BR Kiran","year":"2021","unstructured":"Kiran BR, Sobh I, Talpaert V et al (2021) Deep reinforcement learning for autonomous driving: A survey. IEEE Trans Intell Transp Syst 23(6):4909\u20134926. https:\/\/doi.org\/10.1109\/TITS.2021.3054625","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"2","key":"6054_CR2","doi-asserted-by":"publisher","first-page":"1216","DOI":"10.1109\/LRA.2021.3056028","volume":"6","author":"M Thoresen","year":"2021","unstructured":"Thoresen M, Nielsen NH, Mathiassen K et al (2021) Path planning for ugvs based on traversability hybrid a. IEEE Robot Autom Lett 6(2):1216\u20131223. https:\/\/doi.org\/10.1109\/LRA.2021.3056028","journal-title":"IEEE Robot Autom Lett"},{"issue":"2","key":"6054_CR3","doi-asserted-by":"publisher","first-page":"496","DOI":"10.1109\/LRA.2022.3228159","volume":"8","author":"Y Qi","year":"2022","unstructured":"Qi Y, He B, Wang R et al (2022) Hierarchical motion planning for autonomous vehicles in unstructured dynamic environments. IEEE Robot Autom Lett 8(2):496\u2013503. https:\/\/doi.org\/10.1109\/LRA.2022.3228159","journal-title":"IEEE Robot Autom Lett"},{"key":"6054_CR4","doi-asserted-by":"publisher","first-page":"108453","DOI":"10.1016\/j.compag.2023.108453","volume":"215","author":"L Ye","year":"2023","unstructured":"Ye L, Wu F, Zou X et al (2023) Path planning for mobile robots in unstructured orchard environments: An improved kinematically constrained bi-directional rrt approach. Comput Electron Agric 215:108453. https:\/\/doi.org\/10.1016\/j.compag.2023.108453","journal-title":"Comput Electron Agric"},{"issue":"7","key":"6054_CR5","doi-asserted-by":"publisher","first-page":"6282","DOI":"10.1109\/TITS.2021.3086804","volume":"23","author":"R Roriz","year":"2022","unstructured":"Roriz R, Cabral J, Gomes T (2022) Automotive lidar technology: A survey. IEEE Trans Intell Transp Syst 23(7):6282\u20136297. https:\/\/doi.org\/10.1109\/TITS.2021.3086804","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"1","key":"6054_CR6","doi-asserted-by":"publisher","first-page":"2145","DOI":"10.1109\/TIP.2024.3364511","volume":"33","author":"D Kong","year":"2024","unstructured":"Kong D, Li X, Xu Q et al (2024) Sc_lpr: Semantically consistent lidar place recognition based on chained cascade network in long-term dynamic environments. IEEE Trans Image Process 33(1):2145\u20132157. https:\/\/doi.org\/10.1109\/TIP.2024.3364511","journal-title":"IEEE Trans Image Process"},{"issue":"1","key":"6054_CR7","doi-asserted-by":"publisher","first-page":"103","DOI":"10.1109\/TIV.2023.3318070","volume":"9","author":"PS Chib","year":"2023","unstructured":"Chib PS, Singh P (2023) Recent advancements in end-to-end autonomous driving using deep learning: A survey. IEEE Trans Intell Veh 9(1):103\u2013118. https:\/\/doi.org\/10.1109\/TIV.2023.3318070","journal-title":"IEEE Trans Intell Veh"},{"key":"6054_CR8","doi-asserted-by":"publisher","first-page":"368","DOI":"10.1016\/j.neucom.2021.03.084","volume":"449","author":"F Li","year":"2021","unstructured":"Li F, Guo C, Luo B et al (2021) Multi goals and multi scenes visual mapless navigation in indoor using meta-learning and scene priors. Neurocomputing 449:368\u2013377. https:\/\/doi.org\/10.1016\/j.neucom.2021.03.084","journal-title":"Neurocomputing"},{"key":"6054_CR9","doi-asserted-by":"publisher","first-page":"118","DOI":"10.1016\/j.neucom.2022.06.102","volume":"503","author":"H Jiang","year":"2022","unstructured":"Jiang H, Esfahani MA, Wu K et al (2022) itd3-cln: Learn to navigate in dynamic scene through deep reinforcement learning. Neurocomputing 503:118\u2013128. https:\/\/doi.org\/10.1016\/j.neucom.2022.06.102","journal-title":"Neurocomputing"},{"key":"6054_CR10","doi-asserted-by":"publisher","first-page":"195608","DOI":"10.1109\/ACCESS.2020.3034524","volume":"8","author":"A Staroverov","year":"2020","unstructured":"Staroverov A, Yudin DA, Belkin I et al (2020) Real-time object navigation with deep neural networks and hierarchical reinforcement learning. IEEE Access 8:195608\u2013195621. https:\/\/doi.org\/10.1109\/ACCESS.2020.3034524","journal-title":"IEEE Access"},{"issue":"20","key":"6054_CR11","doi-asserted-by":"publisher","first-page":"23244","DOI":"10.1007\/s10489-023-04754-7","volume":"53","author":"Z Li","year":"2023","unstructured":"Li Z, Zhou A (2023) Rddrl: a recurrent deduction deep reinforcement learning model for multimodal vision-robot navigation. Appl Intell 53(20):23244\u201323270. https:\/\/doi.org\/10.1007\/s10489-023-04754-7","journal-title":"Appl Intell"},{"issue":"3","key":"6054_CR12","doi-asserted-by":"publisher","first-page":"2488","DOI":"10.1007\/s10489-022-03357-y","volume":"53","author":"MS Rais","year":"2023","unstructured":"Rais MS, Boudour R, Zouaidia K et al (2023) Decision making for autonomous vehicles in highway scenarios using harmonic sk deep sarsa. Appl Intell 53(3):2488\u20132505. https:\/\/doi.org\/10.1007\/s10489-022-03357-y","journal-title":"Appl Intell"},{"issue":"9","key":"6054_CR13","doi-asserted-by":"publisher","first-page":"12669","DOI":"10.1109\/TITS.2024.3386200","volume":"25","author":"J Hu","year":"2024","unstructured":"Hu J, Li X, Hu W et al (2024) A cooperative control methodology considering dynamic interaction for multiple connected and automated vehicles in the merging zone. IEEE Trans Intell Transp Syst 25(9):12669\u201312681. https:\/\/doi.org\/10.1109\/TITS.2024.3386200","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"10","key":"6054_CR14","doi-asserted-by":"publisher","first-page":"18620","DOI":"10.1109\/TITS.2022.3157910","volume":"23","author":"J Hu","year":"2022","unstructured":"Hu J, Li X, Cen Y et al (2022) A roadside decision-making methodology based on deep reinforcement learning to simultaneously improve the safety and efficiency of merging zone. IEEE Trans Intell Transp Syst 23(10):18620\u201318631. https:\/\/doi.org\/10.1109\/TITS.2022.3157910","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"2","key":"6054_CR15","doi-asserted-by":"publisher","first-page":"1312","DOI":"10.1109\/LRA.2021.3057023","volume":"6","author":"G Kahn","year":"2021","unstructured":"Kahn G, Abbeel P, Levine S (2021) Badgr: An autonomous self-supervised learning-based navigation system. IEEE Robot Autom Lett 6(2):1312\u20131319. https:\/\/doi.org\/10.1109\/LRA.2021.3057023","journal-title":"IEEE Robot Autom Lett"},{"issue":"4","key":"6054_CR16","doi-asserted-by":"publisher","first-page":"6569","DOI":"10.1109\/LRA.2021.3093551","volume":"6","author":"H Hu","year":"2021","unstructured":"Hu H, Zhang K, Tan AH et al (2021) A sim-to-real pipeline for deep reinforcement learning for autonomous robot navigation in cluttered rough terrain. IEEE Robot Autom Lett 6(4):6569\u20136576. https:\/\/doi.org\/10.1109\/LRA.2021.3093551","journal-title":"IEEE Robot Autom Lett"},{"issue":"4","key":"6054_CR17","doi-asserted-by":"publisher","first-page":"6748","DOI":"10.1109\/LRA.2020.3011912","volume":"5","author":"S Josef","year":"2020","unstructured":"Josef S, Degani A (2020) Deep reinforcement learning for safe local planning of a ground vehicle in unknown rough terrain. IEEE Robot Autom Lett 5(4):6748\u20136755. https:\/\/doi.org\/10.1109\/LRA.2020.3011912","journal-title":"IEEE Robot Autom Lett"},{"issue":"5","key":"6054_CR18","doi-asserted-by":"publisher","first-page":"4962","DOI":"10.1109\/TIE.2022.3190850","volume":"70","author":"W Zhu","year":"2023","unstructured":"Zhu W, Hayashibe M (2023) A hierarchical deep reinforcement learning framework with high efficiency and generalization for fast and safe navigation. IEEE Trans Ind Electron 70(5):4962\u20134971. https:\/\/doi.org\/10.1109\/TIE.2022.3190850","journal-title":"IEEE Trans Ind Electron"},{"key":"6054_CR19","doi-asserted-by":"publisher","unstructured":"Zhang K, Niroui F, Ficocelli M et\u00a0al (2018) Robot navigation of environments with unknown rough terrain using deep reinforcement learning. In: 2018 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), IEEE, pp 1\u20137. https:\/\/doi.org\/10.1109\/SSRR.2018.8468643","DOI":"10.1109\/SSRR.2018.8468643"},{"issue":"2","key":"6054_CR20","doi-asserted-by":"publisher","first-page":"730","DOI":"10.1109\/LRA.2021.3133591","volume":"7","author":"R Cimurs","year":"2022","unstructured":"Cimurs R, Suh IH, Lee JH (2022) Goal-driven autonomous exploration through deep reinforcement learning. IEEE Robot Autom Lett 7(2):730\u2013737. https:\/\/doi.org\/10.1109\/LRA.2021.3133591","journal-title":"IEEE Robot Autom Lett"},{"key":"6054_CR21","doi-asserted-by":"publisher","unstructured":"Weerakoon K, Sathyamoorthy AJ, Patel U et\u00a0al (2022) Terp: Reliable planning in uneven outdoor environments using deep reinforcement learning. In: 2022 International Conference on Robotics and Automation (ICRA), pp 9447\u20139453. https:\/\/doi.org\/10.1109\/ICRA46639.2022.9812238","DOI":"10.1109\/ICRA46639.2022.9812238"},{"key":"6054_CR22","doi-asserted-by":"publisher","first-page":"2147","DOI":"10.1007\/s40747-022-00906-3","volume":"9","author":"SS Samsani","year":"2022","unstructured":"Samsani SS, Mutahira H, Muhammad MS (2022) Memory-based crowd-aware robot navigation using deep reinforcement learning. Complex Intell Syst 9:2147\u20132158. https:\/\/doi.org\/10.1007\/s40747-022-00906-3","journal-title":"Complex Intell Syst"},{"issue":"2","key":"6054_CR23","doi-asserted-by":"publisher","first-page":"1832","DOI":"10.1109\/TITS.2023.3312453","volume":"25","author":"W Huang","year":"2024","unstructured":"Huang W, Zhou Y, He X et al (2024) Goal-guided transformer-enabled reinforcement learning for efficient autonomous navigation. IEEE Trans Intell Transp Syst 25(2):1832\u20131845. https:\/\/doi.org\/10.1109\/TITS.2023.3312453","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"6054_CR24","doi-asserted-by":"publisher","first-page":"101889","DOI":"10.1016\/j.aei.2023.101889","volume":"55","author":"K Zhou","year":"2023","unstructured":"Zhou K, Guo C, Zhang H et al (2023) Optimal graph transformer viterbi knowledge inference network for more successful visual navigation. Adv Eng Inf 55:101889. https:\/\/doi.org\/10.1016\/j.aei.2023.101889","journal-title":"Adv Eng Inf"},{"issue":"2","key":"6054_CR25","doi-asserted-by":"publisher","first-page":"912","DOI":"10.1109\/LRA.2023.3234815","volume":"8","author":"Y Yang","year":"2023","unstructured":"Yang Y, Jiang J, Zhang J et al (2023) St$$^{2}$$: Spatial-temporal state transformer for crowd-aware autonomous navigation. IEEE Robot Autom Lett 8(2):912\u2013919. https:\/\/doi.org\/10.1109\/LRA.2023.3234815","journal-title":"IEEE Robot Autom Lett"},{"issue":"3","key":"6054_CR26","doi-asserted-by":"publisher","first-page":"4405","DOI":"10.1109\/TIV.2024.3372625","volume":"9","author":"H Liu","year":"2024","unstructured":"Liu H, Huang Z, Mo X et al (2024) Augmenting reinforcement learning with transformer-based scene representation learning for decision-making of autonomous driving. IEEE Trans Intell Veh 9(3):4405\u20134421. https:\/\/doi.org\/10.1109\/TIV.2024.3372625","journal-title":"IEEE Trans Intell Veh"},{"issue":"6","key":"6054_CR27","doi-asserted-by":"publisher","first-page":"6013","DOI":"10.1109\/TIE.2023.3290244","volume":"71","author":"VRF Miranda","year":"2024","unstructured":"Miranda VRF, Neto AA, Freitas GM et al (2024) Generalization in deep reinforcement learning for robotic navigation by reward shaping. IEEE Trans Ind Electron 71(6):6013\u20136020. https:\/\/doi.org\/10.1109\/TIE.2023.3290244","journal-title":"IEEE Trans Ind Electron"},{"issue":"12","key":"6054_CR28","doi-asserted-by":"publisher","first-page":"14745","DOI":"10.1109\/TPAMI.2023.3314762","volume":"45","author":"J Wu","year":"2023","unstructured":"Wu J, Zhou Y, Yang H et al (2023) Human-guided reinforcement learning with sim-to-real transfer for autonomous navigation. IEEE Trans Pattern Anal Mach Intell 45(12):14745\u201314759. https:\/\/doi.org\/10.1109\/TPAMI.2023.3314762","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"6","key":"6054_CR29","doi-asserted-by":"publisher","first-page":"10778","DOI":"10.1109\/JIOT.2023.3327753","volume":"11","author":"L Zhang","year":"2024","unstructured":"Zhang L, Peng J, Yi W et al (2024) A state-decomposition ddpg algorithm for uav autonomous navigation in 3-d complex environments. IEEE Internet Things J 11(6):10778\u201310790. https:\/\/doi.org\/10.1109\/JIOT.2023.3327753","journal-title":"IEEE Internet Things J"},{"issue":"4","key":"6054_CR30","doi-asserted-by":"publisher","first-page":"6458","DOI":"10.1109\/LRA.2021.3093009","volume":"6","author":"H Lim","year":"2021","unstructured":"Lim H, Oh M, Myung H (2021) Patchwork: Concentric zone-based region-wise ground segmentation with ground likelihood estimation using a 3d lidar sensor. IEEE Robot Autom Lett 6(4):6458\u20136465. https:\/\/doi.org\/10.1109\/LRA.2021.3093009","journal-title":"IEEE Robot Autom Lett"},{"key":"6054_CR31","unstructured":"Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need. In: Guyon I, Luxburg UV, Bengio S et al (eds) Advances in Neural Information Processing Systems, Long Beach, CA, USA"},{"issue":"6","key":"6054_CR32","doi-asserted-by":"publisher","first-page":"3078","DOI":"10.3390\/app12063078","volume":"12","author":"SY Han","year":"2022","unstructured":"Han SY, Liang T (2022) Reinforcement-learning-based vibration control for a vehicle semi-active suspension system via the ppo approach. Appl Sci 12(6):3078. https:\/\/doi.org\/10.3390\/app12063078","journal-title":"Appl Sci"},{"issue":"3","key":"6054_CR33","doi-asserted-by":"publisher","first-page":"4986","DOI":"10.1109\/LRA.2021.3064461","volume":"6","author":"X Huang","year":"2021","unstructured":"Huang X, Deng H, Zhang W et al (2021) Towards multi-modal perception-based navigation: A deep reinforcement learning method. IEEE Robot Autom Lett 6(3):4986\u20134993. https:\/\/doi.org\/10.1109\/LRA.2021.3064461","journal-title":"IEEE Robot Autom Lett"},{"key":"6054_CR34","unstructured":"Raffin A, Hill A, Gleave A et\u00a0al (2021) Stable-baselines3: Reliable reinforcement learning implementations. J Mach Learn Res 22(268):1\u20138. http:\/\/jmlr.org\/papers\/v22\/20-1364.html"},{"issue":"6","key":"6054_CR35","doi-asserted-by":"publisher","first-page":"3692","DOI":"10.1109\/TIV.2023.3274536","volume":"8","author":"S Teng","year":"2023","unstructured":"Teng S, Hu X, Deng P et al (2023) Motion planning for autonomous driving: The state of the art and future perspectives. IEEE Trans Intell Veh 8(6):3692\u20133711. https:\/\/doi.org\/10.1109\/TIV.2023.3274536","journal-title":"IEEE Trans Intell Veh"},{"key":"6054_CR36","unstructured":"Dosovitskiy A, Ros G, Codevilla F et\u00a0al (2017) CARLA: An open urban driving simulator. In: Proceedings of the 1st annual conference on robot learning, proceedings of machine learning research, vol\u00a078. PMLR, pp 1\u201316. https:\/\/proceedings.mlr.press\/v78\/dosovitskiy17a.html"},{"key":"6054_CR37","doi-asserted-by":"publisher","unstructured":"Yang F, Cao C, Zhu H et\u00a0al (2022) Far planner: Fast, attemptable route planner using dynamic visibility update. In: 2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp 9\u201316. https:\/\/doi.org\/10.1109\/IROS47612.2022.9981574","DOI":"10.1109\/IROS47612.2022.9981574"},{"issue":"2","key":"6054_CR38","doi-asserted-by":"publisher","first-page":"922","DOI":"10.1109\/TRO.2021.3100142","volume":"38","author":"J Tordesillas","year":"2022","unstructured":"Tordesillas J, Lopez BT, Everett M et al (2022) Faster: Fast and safe trajectory planner for navigation in unknown environments. IEEE Trans Robot 38(2):922\u2013938. https:\/\/doi.org\/10.1109\/TRO.2021.3100142","journal-title":"IEEE Trans Robot"},{"issue":"4","key":"6054_CR39","doi-asserted-by":"publisher","first-page":"399","DOI":"10.1007\/s11044-022-09816-1","volume":"54","author":"S Benatti","year":"2022","unstructured":"Benatti S, Young A, Elmquist A et al (2022) End-to-end learning for off-road terrain navigation using the chrono open-source simulation platform. Multibody Syst Dyn 54(4):399\u2013414. https:\/\/doi.org\/10.1007\/s11044-022-09816-1","journal-title":"Multibody Syst Dyn"},{"key":"6054_CR40","doi-asserted-by":"publisher","unstructured":"Wang Y, Wang J, Yang Y et\u00a0al (2022) An end-to-end deep reinforcement learning model based on proximal policy optimization algorithm for autonomous driving of off-road vehicle. In: International conference on autonomous unmanned systems. Springer, pp 2692\u20132704. https:\/\/doi.org\/10.1007\/978-981-99-0479-2_248","DOI":"10.1007\/978-981-99-0479-2_248"},{"key":"6054_CR41","doi-asserted-by":"publisher","first-page":"1501","DOI":"10.1007\/s00366-023-01852-5","volume":"40","author":"A Abbaszadeh Shahri","year":"2024","unstructured":"Abbaszadeh Shahri A, Chunling S, Larsson S (2024) A hybrid ensemble-based automated deep learning approach to generate 3d geo-models and uncertainty analysis. Eng Comput 40:1501\u20131516. https:\/\/doi.org\/10.1007\/s00366-023-01852-5","journal-title":"Eng Comput"},{"key":"6054_CR42","doi-asserted-by":"publisher","first-page":"1351","DOI":"10.1007\/s11053-022-10051-w","volume":"31","author":"A Abbaszadeh Shahri","year":"2022","unstructured":"Abbaszadeh Shahri A, Shan C, Larsson S (2022) A novel approach to uncertainty quantification in groundwater table modeling by automated predictive deep learning. Nat Resour Res 31:1351\u20131373. https:\/\/doi.org\/10.1007\/s11053-022-10051-w","journal-title":"Nat Resour Res"},{"issue":"7","key":"6054_CR43","doi-asserted-by":"publisher","first-page":"4600","DOI":"10.1109\/TSMC.2021.3098451","volume":"52","author":"Y Gu","year":"2022","unstructured":"Gu Y, Cheng Y, Chen CLP et al (2022) Proximal policy optimization with policy feedback. IEEE Trans Syst Man Cybern Syst 52(7):4600\u20134610. https:\/\/doi.org\/10.1109\/TSMC.2021.3098451","journal-title":"IEEE Trans Syst Man Cybern Syst"},{"key":"6054_CR44","unstructured":"Hiraoka T, Imagawa T, Hashimoto T et\u00a0al (2022) Dropout q-functions for doubly efficient reinforcement learning. In: International conference on learning representations. https:\/\/openreview.net\/forum?id=xCVJMsPv3RT"},{"key":"6054_CR45","unstructured":"Bhatt A, Palenicek D, Belousov B et\u00a0al (2024) Crossq: Batch normalization in deep reinforcement learning for greater sample efficiency and simplicity. In: The twelfth international conference on learning representations, Vienna, Austria. https:\/\/openreview.net\/forum?id=PczQtTsTIX"},{"key":"6054_CR46","unstructured":"Fujimoto S, Chang WD, Smith E et\u00a0al (2023) For sale: State-action representation learning for deep reinforcement learning. In: Advances in neural information processing systems, New Orleans, LA, USA, pp 61573\u201361624. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2023\/file\/c20ac0df6c213db6d3a930fe9c7296c8-Paper-Conference.pdf"},{"key":"6054_CR47","doi-asserted-by":"publisher","unstructured":"Abbaszadeh Shahri A, Shan C, Larsson S et al (2024) Normalizing large scale sensor-based mwd data: An automated method toward a unified database. Sensors 24(4):1209. https:\/\/doi.org\/10.3390\/s24041209","DOI":"10.3390\/s24041209"},{"issue":"7","key":"6054_CR48","doi-asserted-by":"publisher","first-page":"10515","DOI":"10.1109\/JSEN.2024.3358341","volume":"24","author":"X Yan","year":"2024","unstructured":"Yan X, Yang J, Zhu X et al (2024) Denoising framework based on multiframe continuous point clouds for autonomous driving lidar in snowy weather. IEEE Sens J 24(7):10515\u201310527. https:\/\/doi.org\/10.1109\/JSEN.2024.3358341","journal-title":"IEEE Sens J"},{"issue":"1","key":"6054_CR49","doi-asserted-by":"publisher","first-page":"855","DOI":"10.1109\/TNNLS.2022.3177685","volume":"35","author":"J Wu","year":"2024","unstructured":"Wu J, Huang Z, Huang W et al (2024) Prioritized experience-based reinforcement learning with human guidance for autonomous driving. IEEE Trans Neural Netw Learn Syst 35(1):855\u2013869. https:\/\/doi.org\/10.1109\/TNNLS.2022.3177685","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"6054_CR50","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1016\/j.eng.2022.05.017","volume":"21","author":"J Wu","year":"2023","unstructured":"Wu J, Huang Z, Hu Z et al (2023) Toward human-in-the-loop ai: Enhancing deep reinforcement learning via real-time human guidance for autonomous driving. Engineering 21:75\u201391. https:\/\/doi.org\/10.1016\/j.eng.2022.05.017","journal-title":"Engineering"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-024-06054-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-024-06054-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-024-06054-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,31]],"date-time":"2025-01-31T14:53:23Z","timestamp":1738335203000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-024-06054-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,2]]},"references-count":50,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["6054"],"URL":"https:\/\/doi.org\/10.1007\/s10489-024-06054-0","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"type":"print","value":"0924-669X"},{"type":"electronic","value":"1573-7497"}],"subject":[],"published":{"date-parts":[[2025,1,2]]},"assertion":[{"value":"6 September 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 January 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"254"}}