{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T14:17:22Z","timestamp":1769696242999,"version":"3.49.0"},"reference-count":37,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T00:00:00Z","timestamp":1769644800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T00:00:00Z","timestamp":1769644800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"General Program of the Xinjiang Natural Science Foundation under Grant","award":["2024D01C55"],"award-info":[{"award-number":["2024D01C55"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52275003"],"award-info":[{"award-number":["52275003"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Intel Serv Robotics"],"published-print":{"date-parts":[[2026,3]]},"DOI":"10.1007\/s11370-026-00693-7","type":"journal-article","created":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T02:40:37Z","timestamp":1769654437000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Autonomous navigation via dual-priority experience replay with adaptive hybrid weighting"],"prefix":"10.1007","volume":"19","author":[{"given":"Yuewei","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Wendong","family":"Xiao","sequence":"additional","affiliation":[]},{"given":"Liang","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Teng","family":"Ran","sequence":"additional","affiliation":[]},{"given":"Jianping","family":"Cui","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Lv","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,29]]},"reference":[{"key":"693_CR1","unstructured":"Quinlan S, Khatib O (1993) Elastic bands: Connecting path planning and control. In: [1993] Proceedings IEEE International Conference on Robotics and Automation, pp. 802\u2013807. IEEE"},{"issue":"1","key":"693_CR2","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1109\/100.580977","volume":"4","author":"D Fox","year":"1997","unstructured":"Fox D, Burgard W, Thrun S (1997) The dynamic window approach to collision avoidance. IEEE Robotics & Automation Magazine 4(1):23\u201333","journal-title":"IEEE Robotics & Automation Magazine"},{"issue":"5","key":"693_CR3","doi-asserted-by":"publisher","first-page":"569","DOI":"10.1007\/s10514-022-10039-8","volume":"46","author":"X Xiao","year":"2022","unstructured":"Xiao X, Liu B, Warnell G, Stone P (2022) Motion planning and control for mobile robot navigation using machine learning: a survey. Auton Robot 46(5):569\u2013597","journal-title":"Auton Robot"},{"issue":"1","key":"693_CR4","doi-asserted-by":"publisher","first-page":"411","DOI":"10.1146\/annurev-control-042920-020211","volume":"5","author":"L Brunke","year":"2022","unstructured":"Brunke L, Greeff M, Hall AW, Yuan Z, Zhou S, Panerati J, Schoellig AP (2022) Safe learning in robotics: From learning-based control to safe reinforcement learning. Annual Review of Control Robotics and Autonomous Systems 5(1):411\u2013444","journal-title":"Annual Review of Control Robotics and Autonomous Systems"},{"issue":"4\u20135","key":"693_CR5","doi-asserted-by":"publisher","first-page":"698","DOI":"10.1177\/0278364920987859","volume":"40","author":"J Ibarz","year":"2021","unstructured":"Ibarz J, Tan J, Finn C, Kalakrishnan M, Pastor P, Levine S (2021) How to train your robot with deep reinforcement learning: lessons we have learned. The International Journal of Robotics Research 40(4\u20135):698\u2013721","journal-title":"The International Journal of Robotics Research"},{"issue":"7","key":"693_CR6","doi-asserted-by":"publisher","first-page":"3762","DOI":"10.3390\/s23073762","volume":"23","author":"D Han","year":"2023","unstructured":"Han D, Mulyana B, Stankovic V, Cheng S (2023) A survey on deep reinforcement learning algorithms for robotic manipulation. Sensors 23(7):3762","journal-title":"Sensors"},{"issue":"6","key":"693_CR7","doi-asserted-by":"publisher","first-page":"4909","DOI":"10.1109\/TITS.2021.3054625","volume":"23","author":"BR Kiran","year":"2021","unstructured":"Kiran BR, Sobh I, Talpaert V, Mannion P, Al Sallab AA, Yogamani S, P\u00e9rez P (2021) Deep reinforcement learning for autonomous driving: A survey. IEEE Trans Intell Transp Syst 23(6):4909\u20134926","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"1","key":"693_CR8","doi-asserted-by":"publisher","first-page":"194","DOI":"10.1109\/TIV.2022.3185159","volume":"8","author":"J Wu","year":"2022","unstructured":"Wu J, Huang Z, Lv C (2022) Uncertainty-aware model-based reinforcement learning: Methodology and application in autonomous driving. IEEE Transactions on Intelligent Vehicles 8(1):194\u2013203","journal-title":"IEEE Transactions on Intelligent Vehicles"},{"issue":"2","key":"693_CR9","doi-asserted-by":"publisher","first-page":"401","DOI":"10.1016\/j.ejor.2021.07.016","volume":"298","author":"RN Boute","year":"2022","unstructured":"Boute RN, Gijsbrechts J, Van Jaarsveld W, Vanvuchelen N (2022) Deep reinforcement learning for inventory control: A roadmap. Eur J Oper Res 298(2):401\u2013412","journal-title":"Eur J Oper Res"},{"issue":"3","key":"693_CR10","doi-asserted-by":"publisher","first-page":"1349","DOI":"10.1287\/msom.2021.1064","volume":"24","author":"J Gijsbrechts","year":"2022","unstructured":"Gijsbrechts J, Boute RN, Van Mieghem JA, Zhang DJ (2022) Can deep reinforcement learning improve inventory management? performance on lost sales, dual-sourcing, and multi-echelon problems. Manufacturing & Service Operations Management 24(3):1349\u20131368","journal-title":"Manufacturing & Service Operations Management"},{"key":"693_CR11","doi-asserted-by":"crossref","unstructured":"Rupp F, Eberhardinger M, Eckert K (2023) Balancing of competitive two-player game levels with reinforcement learning. In: 2023 IEEE Conference on Games (CoG), 1\u20138. IEEE","DOI":"10.1109\/CoG57401.2023.10333248"},{"issue":"4","key":"693_CR12","doi-asserted-by":"publisher","first-page":"2443","DOI":"10.3390\/app13042443","volume":"13","author":"K Souchleris","year":"2023","unstructured":"Souchleris K, Sidiropoulos GK, Papakostas GA (2023) Reinforcement learning in game industry$$-$$review, prospects and challenges. Appl Sci 13(4):2443","journal-title":"Appl Sci"},{"key":"693_CR13","unstructured":"Watkins CJCH, et\u00a0al (1989) Learning from delayed rewards"},{"key":"693_CR14","doi-asserted-by":"publisher","first-page":"229","DOI":"10.1023\/A:1022672621406","volume":"8","author":"RJ Williams","year":"1992","unstructured":"Williams RJ (1992) Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach Learn 8:229\u2013256","journal-title":"Mach Learn"},{"issue":"2","key":"693_CR15","doi-asserted-by":"publisher","first-page":"604","DOI":"10.1109\/TCYB.2019.2939174","volume":"51","author":"Z Zhang","year":"2019","unstructured":"Zhang Z, Chen J, Chen Z, Li W (2019) Asynchronous episodic deep deterministic policy gradient: Toward continuous control in computationally complex environments. IEEE transactions on cybernetics 51(2):604\u2013613","journal-title":"IEEE transactions on cybernetics"},{"key":"693_CR16","unstructured":"Mnih V, Kavukcuoglu K, Silver D, Graves A, Antonoglou I, Wierstra D, Riedmiller M (2013) Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602"},{"key":"693_CR17","unstructured":"Lillicrap TP, Hunt JJ, Pritzel A, Heess N, Erez T, Tassa Y, Silver D, Wierstra D (2015) Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971"},{"issue":"5","key":"693_CR18","doi-asserted-by":"publisher","first-page":"3476","DOI":"10.1109\/JIOT.2020.3022861","volume":"8","author":"Z Shi","year":"2020","unstructured":"Shi Z, Xie X, Lu H, Yang H, Kadoch M, Cheriet M (2020) Deep-reinforcement-learning-based spectrum resource management for industrial internet of things. IEEE Internet Things J 8(5):3476\u20133489","journal-title":"IEEE Internet Things J"},{"key":"693_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2022.105642","volume":"118","author":"Y Zhang","year":"2023","unstructured":"Zhang Y, Rao X, Liu C, Zhang X, Zhou Y (2023) A cooperative ev charging scheduling strategy based on double deep q-network and prioritized experience replay. Eng Appl Artif Intell 118:105642","journal-title":"Eng Appl Artif Intell"},{"issue":"2","key":"693_CR20","doi-asserted-by":"publisher","first-page":"871","DOI":"10.1109\/TNSE.2022.3224028","volume":"10","author":"G Pang","year":"2022","unstructured":"Pang G, Wang X, Wang L, Hao F, Lin Y, Wan P, Min G (2022) Efficient deep reinforcement learning-enabled recommendation. IEEE Transactions on Network Science and Engineering 10(2):871\u2013886","journal-title":"IEEE Transactions on Network Science and Engineering"},{"key":"693_CR21","doi-asserted-by":"publisher","first-page":"44","DOI":"10.1016\/j.neucom.2022.12.020","volume":"523","author":"L Yu","year":"2023","unstructured":"Yu L, Huo S, Wang Z, Li K (2023) Hybrid attention-oriented experience replay for deep reinforcement learning and its application to a multi-robot cooperative hunting problem. Neurocomputing 523:44\u201357","journal-title":"Neurocomputing"},{"key":"693_CR22","doi-asserted-by":"crossref","unstructured":"Ruan X, Ren D, Zhu X, Huang J (2019) Mobile robot navigation based on deep reinforcement learning. In: 2019 Chinese Control and Decision Conference (CCDC), 6174\u20136178. IEEE","DOI":"10.1109\/CCDC.2019.8832393"},{"key":"693_CR23","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s42979-021-00817-z","volume":"2","author":"G Chen","year":"2021","unstructured":"Chen G, Pan L, Chen Y, Xu P, Wang Z, Wu P, Ji J, Chen X (2021) Deep reinforcement learning of map-based obstacle avoidance for mobile robot navigation. SN Computer Science 2:1\u201314","journal-title":"SN Computer Science"},{"issue":"5","key":"693_CR24","doi-asserted-by":"publisher","first-page":"4962","DOI":"10.1109\/TIE.2022.3190850","volume":"70","author":"W Zhu","year":"2022","unstructured":"Zhu W, Hayashibe M (2022) A hierarchical deep reinforcement learning framework with high efficiency and generalization for fast and safe navigation. IEEE Trans Industr Electron 70(5):4962\u20134971","journal-title":"IEEE Trans Industr Electron"},{"key":"693_CR25","doi-asserted-by":"publisher","DOI":"10.1016\/j.aei.2024.102920","volume":"62","author":"L Bo","year":"2024","unstructured":"Bo L, Zhang T, Zhang H, Hong J, Liu M, Zhang C, Liu B (2024) 3d uav path planning in unknown environment: A transfer reinforcement learning method based on low-rank adaption. Adv Eng Inform 62:102920","journal-title":"Adv Eng Inform"},{"key":"693_CR26","doi-asserted-by":"crossref","unstructured":"Mnih V, Kavukcuoglu K, Silver D, Rusu AA, Veness J, Bellemare MG, Graves A, Riedmiller M, Fidjeland AK, Ostrovski G, et\u00a0al. (2015) Human-level control through deep reinforcement learning. nature 518(7540), 529\u2013533","DOI":"10.1038\/nature14236"},{"key":"693_CR27","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1023\/A:1022628806385","volume":"8","author":"L-J Lin","year":"1992","unstructured":"Lin L-J (1992) Self-improving reactive agents based on reinforcement learning, planning and teaching. Mach Learn 8:293\u2013321","journal-title":"Mach Learn"},{"key":"693_CR28","unstructured":"Schaul T, Quan J, Antonoglou I, Silver D (2015) Prioritized experience replay. arXiv preprint arXiv:1511.05952"},{"key":"693_CR29","unstructured":"Andrychowicz M, Wolski F, Ray A, Schneider J, Fong R, Welinder P, McGrew B, Tobin J, Pieter\u00a0Abbeel O, Zaremba W (2017) Hindsight experience replay. Advances in neural information processing systems 30"},{"key":"693_CR30","unstructured":"Liu H, Trott A, Socher R, Xiong C (2019) Competitive experience replay. arXiv preprint arXiv:1902.00528"},{"key":"693_CR31","doi-asserted-by":"publisher","first-page":"60296","DOI":"10.1109\/ACCESS.2021.3074535","volume":"9","author":"C Kang","year":"2021","unstructured":"Kang C, Rong C, Ren W, Huo F, Liu P (2021) Deep deterministic policy gradient based on double network prioritized experience replay. IEEE Access 9:60296\u201360308","journal-title":"IEEE Access"},{"issue":"11","key":"693_CR32","doi-asserted-by":"publisher","first-page":"11253","DOI":"10.1109\/JSEN.2023.3237206","volume":"23","author":"J Cui","year":"2023","unstructured":"Cui J, Yuan L, He L, Xiao W, Ran T, Zhang J (2023) Multi-input autonomous driving based on deep reinforcement learning with double bias experience replay. IEEE Sens J 23(11):11253\u201311261","journal-title":"IEEE Sens J"},{"key":"693_CR33","doi-asserted-by":"crossref","unstructured":"Liu X, Yu M, Yang C, Zhou L, Wang H, Zhou H (2024) Value distribution ddpg with dual-prioritized experience replay for coordinated control of coal-fired power generation systems. IEEE Transactions on Industrial Informatics","DOI":"10.1109\/TII.2024.3369712"},{"key":"693_CR34","unstructured":"Horgan D, Quan J, Budden D, Barth-Maron G, Hessel M, Hasselt HV, Silver D (2018) Distributed prioritized experience replay. arXiv: 1803.00933"},{"key":"693_CR35","unstructured":"Fujimoto S, Hoof H, Meger D (2018) Addressing function approximation error in actor-critic methods. In: International Conference on Machine Learning, pp. 1587\u20131596. PMLR"},{"issue":"2","key":"693_CR36","doi-asserted-by":"publisher","first-page":"730","DOI":"10.1109\/LRA.2021.3133591","volume":"7","author":"R Cimurs","year":"2022","unstructured":"Cimurs R, Suh IH, Lee JH (2022) Goal-driven autonomous exploration through deep reinforcement learning. IEEE Robotics and Automation Letters 7(2):730\u2013737. https:\/\/doi.org\/10.1109\/LRA.2021.3133591","journal-title":"IEEE Robotics and Automation Letters"},{"key":"693_CR37","doi-asserted-by":"crossref","unstructured":"Lee K, Kim S, Choi J (2023) Adaptive and explainable deployment of navigation skills via hierarchical deep reinforcement learning. In: 2023 IEEE International Conference on Robotics and Automation (ICRA), pp. 1673\u20131679","DOI":"10.1109\/ICRA48891.2023.10160371"}],"container-title":["Intelligent Service Robotics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11370-026-00693-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11370-026-00693-7","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11370-026-00693-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T02:40:45Z","timestamp":1769654445000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11370-026-00693-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,29]]},"references-count":37,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,3]]}},"alternative-id":["693"],"URL":"https:\/\/doi.org\/10.1007\/s11370-026-00693-7","relation":{},"ISSN":["1861-2776","1861-2784"],"issn-type":[{"value":"1861-2776","type":"print"},{"value":"1861-2784","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,29]]},"assertion":[{"value":"31 May 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"29"}}