{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T09:07:32Z","timestamp":1777626452888,"version":"3.51.4"},"reference-count":37,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,4,26]],"date-time":"2024-04-26T00:00:00Z","timestamp":1714089600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,4,26]],"date-time":"2024-04-26T00:00:00Z","timestamp":1714089600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Sci. China Inf. Sci."],"published-print":{"date-parts":[[2024,5]]},"DOI":"10.1007\/s11432-022-3904-9","type":"journal-article","created":{"date-parts":[[2024,4,29]],"date-time":"2024-04-29T15:01:57Z","timestamp":1714402917000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Sampling-efficient path planning and improved actor-critic-based obstacle avoidance for autonomous robots"],"prefix":"10.1007","volume":"67","author":[{"given":"Yefeng","family":"Yang","sequence":"first","affiliation":[]},{"given":"Tao","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Tianqi","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Wenyu","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Han","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Boyang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Chih-yung","family":"Wen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,26]]},"reference":[{"key":"3904_CR1","doi-asserted-by":"publisher","first-page":"610","DOI":"10.1109\/LRA.2019.2891991","volume":"4","author":"F Niroui","year":"2019","unstructured":"Niroui F, Zhang K C, Kashino Z, et al. Deep reinforcement learning robot for search and rescue applications: exploration in unknown cluttered environments. IEEE Robot Autom Lett, 2019, 4: 610\u2013617","journal-title":"IEEE Robot Autom Lett"},{"key":"3904_CR2","doi-asserted-by":"publisher","first-page":"139203","DOI":"10.1007\/s11432-020-3188-4","volume":"66","author":"X L Hou","year":"2023","unstructured":"Hou X L, Li Z Y, Pan Q. Autonomous navigation of a multirotor robot in GNSS-denied environments for search and rescue. Sci China Inf Sci, 2023, 66: 139203","journal-title":"Sci China Inf Sci"},{"key":"3904_CR3","doi-asserted-by":"publisher","first-page":"132","DOI":"10.1080\/01691864.2019.1691941","volume":"34","author":"T Kamegawa","year":"2020","unstructured":"Kamegawa T, Akiyama T, Sakai S, et al. Development of a separable search-and-rescue robot composed of a mobile robot and a snake robot. Adv Robotics, 2020, 34: 132\u2013139","journal-title":"Adv Robotics"},{"key":"3904_CR4","doi-asserted-by":"publisher","first-page":"140306","DOI":"10.1007\/s11432-020-3013-1","volume":"64","author":"T Ma","year":"2021","unstructured":"Ma T, Zhou H B, Qian B, et al. A large-scale clustering and 3D trajectory optimization approach for UAV swarms. Sci China Inf Sci, 2021, 64: 140306","journal-title":"Sci China Inf Sci"},{"key":"3904_CR5","doi-asserted-by":"publisher","first-page":"1748","DOI":"10.1109\/TASE.2020.2976560","volume":"17","author":"J K Wang","year":"2020","unstructured":"Wang J K, Chi W Z, Li C M, et al. Neural RRT*: learning-based optimal path planning. IEEE Trans Automat Sci Eng, 2020, 17: 1748\u20131758","journal-title":"IEEE Trans Automat Sci Eng"},{"key":"3904_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/LRA.2018.2888947","volume":"4","author":"M Kleinbort","year":"2019","unstructured":"Kleinbort M, Solovey K, Littlefield Z, et al. Probabilistic completeness of RRT for geometric and kinodynamic planning with forward propagation. IEEE Robot Autom Lett, 2019, 4: 1\u20137","journal-title":"IEEE Robot Autom Lett"},{"key":"3904_CR7","doi-asserted-by":"crossref","unstructured":"Kuffner J J, LaValle S M. RRT-connect: an efficient approach to single-query path planning. In: Proceedings of the IEEE International Conference on Robotics and Automation, San Francisco, 2000. 995\u20131001","DOI":"10.1109\/ROBOT.2000.844730"},{"key":"3904_CR8","doi-asserted-by":"publisher","first-page":"846","DOI":"10.1177\/0278364911406761","volume":"30","author":"S Karaman","year":"2011","unstructured":"Karaman S, Frazzoli E. Sampling-based algorithms for optimal motion planning. Int J Robotics Res, 2011, 30: 846\u2013894","journal-title":"Int J Robotics Res"},{"key":"3904_CR9","doi-asserted-by":"publisher","first-page":"299","DOI":"10.5772\/56718","volume":"10","author":"J Nasir","year":"2013","unstructured":"Nasir J, Islam F, Malik U, et al. RRT*-Smart: a rapid convergence implementation of RRT*. Int J Adv Robotic Syst, 2013, 10: 299","journal-title":"Int J Adv Robotic Syst"},{"key":"3904_CR10","doi-asserted-by":"publisher","first-page":"8718","DOI":"10.1109\/TIE.2018.2816000","volume":"65","author":"Y Li","year":"2018","unstructured":"Li Y, Cui R X, Li Z J, et al. Neural network approximation based near-optimal motion planning with kinodynamic constraints using RRT. IEEE Trans Ind Electron, 2018, 65: 8718\u20138729","journal-title":"IEEE Trans Ind Electron"},{"key":"3904_CR11","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1016\/j.robot.2018.06.013","volume":"108","author":"Z Tahir","year":"2018","unstructured":"Tahir Z, Qureshi A H, Ayaz Y, et al. Potentially guided bidirectionalized RRT* for fast optimal path planning in cluttered environments. Robotics Autonomous Syst, 2018, 108: 13\u201327","journal-title":"Robotics Autonomous Syst"},{"key":"3904_CR12","doi-asserted-by":"publisher","first-page":"2063","DOI":"10.1109\/TASE.2020.2987397","volume":"17","author":"J K Wang","year":"2020","unstructured":"Wang J K, Meng M Q H, Khatib O. EB-RRT: optimal motion planning for mobile robots. IEEE Trans Automat Sci Eng, 2020, 17: 2063\u20132073","journal-title":"IEEE Trans Automat Sci Eng"},{"key":"3904_CR13","doi-asserted-by":"publisher","first-page":"7244","DOI":"10.1109\/TIE.2020.2998740","volume":"68","author":"J Qi","year":"2021","unstructured":"Qi J, Yang H, Sun H X. MOD-RRT*: a sampling-based algorithm for robot path planning in dynamic environment. IEEE Trans Ind Electron, 2021, 68: 7244\u20137251","journal-title":"IEEE Trans Ind Electron"},{"key":"3904_CR14","doi-asserted-by":"publisher","first-page":"1271","DOI":"10.1109\/TASE.2018.2877963","volume":"16","author":"W Z Chi","year":"2019","unstructured":"Chi W Z, Wang C Q, Wang J K, et al. Risk-DTRRT-based optimal motion planning algorithm for mobile robots. IEEE Trans Automat Sci Eng, 2019, 16: 1271\u20131288","journal-title":"IEEE Trans Automat Sci Eng"},{"key":"3904_CR15","doi-asserted-by":"publisher","first-page":"172209","DOI":"10.1007\/s11432-020-3056-5","volume":"64","author":"L L Xi","year":"2021","unstructured":"Xi L L, Peng Z H, Jiao L, et al. Smooth quadrotor trajectory generation for tracking a moving target in cluttered environments. Sci China Inf Sci, 2021, 64: 172209","journal-title":"Sci China Inf Sci"},{"key":"3904_CR16","doi-asserted-by":"publisher","first-page":"51","DOI":"10.1007\/s10514-020-09947-4","volume":"45","author":"L Chang","year":"2021","unstructured":"Chang L, Shan L, Jiang C, et al. Reinforcement based mobile robot path planning with improved dynamic window approach in unknown environment. Auton Robot, 2021, 45: 51\u201376","journal-title":"Auton Robot"},{"key":"3904_CR17","doi-asserted-by":"publisher","first-page":"1376","DOI":"10.1109\/TIE.2019.2898599","volume":"67","author":"Y J Huang","year":"2019","unstructured":"Huang Y J, Ding H T, Zhang Y B, et al. A motion planning and tracking framework for autonomous vehicles based on artificial potential field elaborated resistance network approach. IEEE Trans Ind Electron, 2019, 67: 1376\u20131386","journal-title":"IEEE Trans Ind Electron"},{"key":"3904_CR18","doi-asserted-by":"crossref","unstructured":"Zhang Y Z, Ma B, Wai C K. A practical study of time-elastic-band planning method for driverless vehicle for auto-parking. In: Proceedings of the International Conference on Intelligent Autonomous Systems, Singapore, 2018. 196\u2013200","DOI":"10.1109\/ICoIAS.2018.8494025"},{"key":"3904_CR19","doi-asserted-by":"crossref","unstructured":"Ames A, Coogan S, Egerstedt M, et al. Control barrier functions: theory and applications. In: Proceedings of the 18th European Control Conference (ECC), Naples, 2019. 3420\u20133431","DOI":"10.23919\/ECC.2019.8796030"},{"key":"3904_CR20","doi-asserted-by":"crossref","unstructured":"Yang G, Vang B, Serlin Z, et al. Sampling-based motion planning via control barrier functions. In: Proceedings of the 3rd International Conference on Automation, Control and Robots, Beijing, 2019. 22\u201329","DOI":"10.1145\/3365265.3365282"},{"key":"3904_CR21","doi-asserted-by":"crossref","unstructured":"Saveriano M, Lee D. Learning barrier functions for constrained motion planning with dynamical systems. In: Proceedings of the IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), Macao, 2019. 112\u2013119","DOI":"10.1109\/IROS40897.2019.8967981"},{"key":"3904_CR22","doi-asserted-by":"publisher","first-page":"529","DOI":"10.1038\/nature14236","volume":"518","author":"V Mnih","year":"2015","unstructured":"Mnih V, Kavukcuoglu K, Silver D, et al. Human-level control through deep reinforcement learning. Nature, 2015, 518: 529\u2013533","journal-title":"Nature"},{"key":"3904_CR23","doi-asserted-by":"publisher","first-page":"9800","DOI":"10.1109\/JIOT.2020.3020067","volume":"8","author":"W S Shi","year":"2021","unstructured":"Shi W S, Li J L, Wu H Q, et al. Drone-cell trajectory planning and resource allocation for highly mobile networks: a hierarchical DRL approach. IEEE Internet Things J, 2021, 8: 9800\u20139813","journal-title":"IEEE Internet Things J"},{"key":"3904_CR24","doi-asserted-by":"publisher","first-page":"102505","DOI":"10.1016\/j.sysarc.2022.102505","volume":"126","author":"Y F Peng","year":"2022","unstructured":"Peng Y F, Tan G Z, Si H W, et al. DRL-GAT-SA: deep reinforcement learning for autonomous driving planning based on graph attention networks and simplex architecture. J Syst Architecture, 2022, 126: 102505","journal-title":"J Syst Architecture"},{"key":"3904_CR25","doi-asserted-by":"publisher","first-page":"5435","DOI":"10.1109\/TNNLS.2021.3084685","volume":"32","author":"L X Zhang","year":"2021","unstructured":"Zhang L X, Zhang R X, Wu T, et al. Safe reinforcement learning with stability guarantee for motion planning of autonomous vehicles. IEEE Trans Neural Netw Learn Syst, 2021, 32: 5435\u20135444","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"3904_CR26","doi-asserted-by":"publisher","first-page":"172102","DOI":"10.1007\/s11432-019-2644-8","volume":"63","author":"J Wang","year":"2020","unstructured":"Wang J, An J, Chen M S, et al. From model to implementation: a network algorithm programming language. Sci China Inf Sci, 2020, 63: 172102","journal-title":"Sci China Inf Sci"},{"key":"3904_CR27","doi-asserted-by":"publisher","first-page":"152303","DOI":"10.1007\/s11432-020-3014-6","volume":"65","author":"H L You","year":"2022","unstructured":"You H L, Hu Y Y, Pan Z W, et al. Density-based user clustering in downlink NOMA systems. Sci China Inf Sci, 2022, 65: 152303","journal-title":"Sci China Inf Sci"},{"key":"3904_CR28","doi-asserted-by":"publisher","first-page":"101445","DOI":"10.1016\/j.jocs.2021.101445","volume":"55","author":"A Fahim","year":"2021","unstructured":"Fahim A. K and starting means for k-means algorithm. J Comput Sci, 2021, 55: 101445","journal-title":"J Comput Sci"},{"key":"3904_CR29","doi-asserted-by":"publisher","first-page":"110241","DOI":"10.1016\/j.knosys.2022.110241","volume":"262","author":"H Z Li","year":"2023","unstructured":"Li H Z, Wang J. CAPKM++2.0: an upgraded version of the collaborative annealing power k-means++ clustering algorithm. Knowledge-Based Syst, 2023, 262: 110241","journal-title":"Knowledge-Based Syst"},{"key":"3904_CR30","doi-asserted-by":"publisher","first-page":"269","DOI":"10.1007\/BF01386390","volume":"1","author":"E W Dijkstra","year":"1959","unstructured":"Dijkstra E W. A note on two problems in connexion with graphs. Numer Math, 1959, 1: 269\u2013271","journal-title":"Numer Math"},{"key":"3904_CR31","doi-asserted-by":"publisher","first-page":"150210","DOI":"10.1007\/s11432-019-2663-y","volume":"63","author":"L Dong","year":"2020","unstructured":"Dong L, Yuan X, Sun C Y. Event-triggered receding horizon control via actor-critic design. Sci China Inf Sci, 2020, 63: 150210","journal-title":"Sci China Inf Sci"},{"key":"3904_CR32","unstructured":"Sutton R S, McAllester D, Singh S, et al. Policy gradient methods for reinforcement learning with function approximation. In: Proceedings of the Advances in Neural Information Processing Systems, 1999. 1057\u20131063"},{"key":"3904_CR33","doi-asserted-by":"publisher","first-page":"1387","DOI":"10.1109\/LRA.2019.2895892","volume":"4","author":"M Pflueger","year":"2019","unstructured":"Pflueger M, Agha A, Sukhatme G S. Rover-IRL: inverse reinforcement learning with soft value iteration networks for planetary rover path planning. IEEE Robot Autom Lett, 2019, 4: 1387\u20131394","journal-title":"IEEE Robot Autom Lett"},{"key":"3904_CR34","unstructured":"Islam F, Nasir J, Malik U, et al. RRT*-Smart: rapid convergence implementation of RRT* towards optimal solution. In: Proceedings of the International Conference on Mechatronics and Automation, Chengdu, 2012. 1651\u20131656"},{"key":"3904_CR35","doi-asserted-by":"publisher","first-page":"628","DOI":"10.1049\/iet-its.2018.5065","volume":"13","author":"Z Tang","year":"2019","unstructured":"Tang Z, Xu X, Wang F, et al. Coordinated control for path following of two-wheel independently actuated autonomous ground vehicle. IET Intelligent Transp Syst, 2019, 13: 628\u2013635","journal-title":"IET Intelligent Transp Syst"},{"key":"3904_CR36","doi-asserted-by":"crossref","unstructured":"Dankwa S, Zheng W F. Twin-delayed DDPG: a deep reinforcement learning technique to model a continuous movement of an intelligent robot agent. In: Proceedings of the 3rd International Conference on Vision, Image and Signal Processing, 2019. 1\u20135","DOI":"10.1145\/3387168.3387199"},{"key":"3904_CR37","doi-asserted-by":"publisher","first-page":"8577","DOI":"10.1109\/JIOT.2019.2921159","volume":"6","author":"C R Qiu","year":"2019","unstructured":"Qiu C R, Hu Y, Chen Y, et al. Deep deterministic policy gradient (DDPG)-based energy harvesting wireless communications. IEEE Internet Things J, 2019, 6: 8577\u20138588","journal-title":"IEEE Internet Things J"}],"container-title":["Science China Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11432-022-3904-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11432-022-3904-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11432-022-3904-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T19:52:33Z","timestamp":1750362753000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11432-022-3904-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,26]]},"references-count":37,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,5]]}},"alternative-id":["3904"],"URL":"https:\/\/doi.org\/10.1007\/s11432-022-3904-9","relation":{},"ISSN":["1674-733X","1869-1919"],"issn-type":[{"value":"1674-733X","type":"print"},{"value":"1869-1919","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,4,26]]},"assertion":[{"value":"20 December 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 August 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 November 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 April 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"152204"}}