{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T19:59:24Z","timestamp":1764014364914,"version":"3.45.0"},"reference-count":24,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"12","license":[{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Guangdong S&amp;T Program","award":["2024B1111090001"],"award-info":[{"award-number":["2024B1111090001"]}]},{"name":"Shenzhen Science and Technology Research Foundation","award":["JCYJ20230807110802005"],"award-info":[{"award-number":["JCYJ20230807110802005"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Robot. Autom. Lett."],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1109\/lra.2025.3616648","type":"journal-article","created":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T17:40:27Z","timestamp":1759340427000},"page":"13201-13208","source":"Crossref","is-referenced-by-count":0,"title":["Lifelong Multi-Agent Path Finding Based on Reinforcement Learning and GPU"],"prefix":"10.1109","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2267-4666","authenticated-orcid":false,"given":"JinSheng","family":"Yang","sequence":"first","affiliation":[{"name":"School of Intelligent Systems Engineering, Sun Yat-sen University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2646-9897","authenticated-orcid":false,"given":"Zhi","family":"Li","sequence":"additional","affiliation":[{"name":"School of Intelligent Systems Engineering, Sun Yat-sen University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2748-8953","authenticated-orcid":false,"given":"JunBo","family":"Wang","sequence":"additional","affiliation":[{"name":"School of Intelligent Systems Engineering, Sun Yat-sen University, Shenzhen, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/783"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i13.17344"},{"key":"ref3","first-page":"608","article-title":"Planning, scheduling and monitoring for airport surface operations","volume-title":"Proc. Workshops 13th AAAI Conf. Artif. Intell.","author":"Morris","year":"2016"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/j.jmsy.2019.12.002"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v27i1.8541"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2024\/35"},{"key":"ref7","article-title":"Learn to follow: Lifelong multi-agent pathfinding with decentralized replanning","volume-title":"Proc. PRL Workshop Ser. Bridging Gap Between AI Plan. Reinforcement Learn.","author":"Skrynnik","year":"2023"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i16.29703"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/IROS55552.2023.10342305"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9341668"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2021.3077863"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN60899.2024.10651091"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.3390\/drones7040236"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/tiv.2024.3363179"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA46639.2022.9811643"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2019.2903261"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9340876"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-77220-0_21"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2021.3062803"},{"key":"ref20","first-page":"709","article-title":"Dynamic programming for partially observable stochastic games","volume-title":"Proc. AAAI Conf. Artif. Intell.","author":"Hansen","year":"2004"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11796"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1609\/socs.v10i1.18510"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2021.3139145"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i18.30054"}],"container-title":["IEEE Robotics and Automation Letters"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/7083369\/11215960\/11185356.pdf?arnumber=11185356","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T19:02:40Z","timestamp":1764010960000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11185356\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12]]},"references-count":24,"journal-issue":{"issue":"12"},"URL":"https:\/\/doi.org\/10.1109\/lra.2025.3616648","relation":{},"ISSN":["2377-3766","2377-3774"],"issn-type":[{"type":"electronic","value":"2377-3766"},{"type":"electronic","value":"2377-3774"}],"subject":[],"published":{"date-parts":[[2025,12]]}}}