{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T05:23:13Z","timestamp":1772774593995,"version":"3.50.1"},"reference-count":55,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,2,22]],"date-time":"2024-02-22T00:00:00Z","timestamp":1708560000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,2,22]],"date-time":"2024-02-22T00:00:00Z","timestamp":1708560000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52175251"],"award-info":[{"award-number":["52175251"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52205268"],"award-info":[{"award-number":["52205268"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"National Basic Scientific Research Program of China","award":["JCKY2021206B005"],"award-info":[{"award-number":["JCKY2021206B005"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2024,6]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Autonomous region protection is a significant research area in multi-agent systems, aiming to empower defenders in preventing intruders from accessing specific regions. This paper presents a Multi-agent Region Protection Environment (MRPE) featuring fewer defenders, defender damages, and intruder evasion strategies targeting defenders. MRPE poses challenges for traditional protection methods due to its high nonstationarity and limited interception time window. To surmount these hurdles, we modify evolutionary reinforcement learning, giving rise to the corresponding multi-agent region protection method (MRPM). MRPM amalgamates the merits of evolutionary algorithms and deep reinforcement learning, specifically leveraging Differential Evolution (DE) and Multi-Agent Deep Deterministic Policy Gradient (MADDPG). DE facilitates diverse sample exploration and overcomes sparse rewards, while MADDPG trains defenders and expedites the DE convergence process. Additionally, an elite selection strategy tailored for multi-agent systems is devised to enhance defender collaboration. The paper also presents ingenious designs for the fitness and reward functions to effectively drive policy optimizations. Finally, extensive numerical simulations are conducted to validate the effectiveness of MRPM.<\/jats:p>","DOI":"10.1007\/s40747-024-01385-4","type":"journal-article","created":{"date-parts":[[2024,2,22]],"date-time":"2024-02-22T18:02:40Z","timestamp":1708624960000},"page":"3727-3742","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["A modified evolutionary reinforcement learning for multi-agent region protection with fewer defenders"],"prefix":"10.1007","volume":"10","author":[{"given":"Siqing","family":"Sun","sequence":"first","affiliation":[]},{"given":"Huachao","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Tianbo","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,22]]},"reference":[{"issue":"2","key":"1385_CR1","doi-asserted-by":"publisher","first-page":"1121","DOI":"10.1109\/TII.2022.3201589","volume":"19","author":"B Ning","year":"2023","unstructured":"Ning B, Han Q-L, Zuo Z, Ding L, Lu Q, Ge X (2023) Fixed-time and prescribed-time consensus control of multiagent systems and its applications: a survey of recent trends and methodologies. IEEE Trans Ind Inform 19(2):1121\u20131135. https:\/\/doi.org\/10.1109\/TII.2022.3201589","journal-title":"IEEE Trans Ind Inform"},{"key":"1385_CR2","doi-asserted-by":"publisher","first-page":"108031","DOI":"10.1016\/j.ast.2022.108031","volume":"132","author":"W Chen","year":"2023","unstructured":"Chen W, Gao C, Jing W (2023) Proximal policy optimization guidance algorithm for intercepting near-space maneuvering targets. Aerosp Sci Technol 132:108031","journal-title":"Aerosp Sci Technol"},{"key":"1385_CR3","first-page":"6382","volume":"30","author":"R Lowe","year":"2017","unstructured":"Lowe R, Wu Y, Tamar A, Harb J, Pieter Abbeel O, Mordatch I (2017) Multi-agent actor-critic for mixed cooperative-competitive environments. Adv Neural Inf Process Syst 30:6382\u20136393","journal-title":"Adv Neural Inf Process Syst"},{"key":"1385_CR4","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1007\/s10514-014-9409-9","volume":"38","author":"E Raboin","year":"2015","unstructured":"Raboin E, \u0160vec P, Nau DS, Gupta SK (2015) Model-predictive asset guarding by team of autonomous surface vehicles in environment with civilian boats. Auton Robot 38:261\u2013282","journal-title":"Auton Robot"},{"key":"1385_CR5","doi-asserted-by":"publisher","first-page":"299","DOI":"10.1016\/j.ast.2015.08.002","volume":"46","author":"X Wang","year":"2015","unstructured":"Wang X, Zhang Y, Wu H (2015) Distributed cooperative guidance of multiple anti-ship missiles with arbitrary impact angle constraint. Aerosp Sci Technol 46:299\u2013311","journal-title":"Aerosp Sci Technol"},{"key":"1385_CR6","doi-asserted-by":"publisher","first-page":"108268","DOI":"10.1016\/j.oceaneng.2020.108268","volume":"219","author":"X Meng","year":"2021","unstructured":"Meng X, Sun B, Zhu D (2021) Harbour protection: moving invasion target interception for multi-AUV based on prediction planning interception method. Ocean Eng 219:108268","journal-title":"Ocean Eng"},{"issue":"5","key":"1385_CR7","doi-asserted-by":"publisher","first-page":"1072","DOI":"10.1007\/s11431-021-1951-9","volume":"65","author":"Y Yu","year":"2022","unstructured":"Yu Y, Liu J, Wei C (2022) Hawk and pigeon\u2019s intelligence for UAV swarm dynamic combat game via competitive learning pigeon-inspired optimization. Sci China Technol Sci 65(5):1072\u20131086","journal-title":"Sci China Technol Sci"},{"issue":"2","key":"1385_CR8","doi-asserted-by":"publisher","first-page":"732","DOI":"10.1109\/TCYB.2022.3163739","volume":"53","author":"S Sui","year":"2023","unstructured":"Sui S, Tong S (2023) Finite-time fuzzy adaptive PPC for nonstrict-feedback nonlinear MIMO systems. IEEE Trans Cybern 53(2):732\u2013742. https:\/\/doi.org\/10.1109\/TCYB.2022.3163739","journal-title":"IEEE Trans Cybern"},{"issue":"3","key":"1385_CR9","doi-asserted-by":"publisher","first-page":"514","DOI":"10.1109\/TCDS.2018.2840971","volume":"10","author":"Y Rizk","year":"2018","unstructured":"Rizk Y, Awad M, Tunstel EW (2018) Decision making in multiagent systems: a survey. IEEE Trans Cognit Dev Syst 10(3):514\u2013529","journal-title":"IEEE Trans Cognit Dev Syst"},{"issue":"1","key":"1385_CR10","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1109\/TCDS.2020.3030571","volume":"14","author":"F Khemakhem","year":"2022","unstructured":"Khemakhem F, Ellouzi H, Ltifi H, Ayed MB (2022) Agent-based intelligent decision support systems: a systematic review. IEEE Trans Cognit Dev Syst 14(1):20\u201334","journal-title":"IEEE Trans Cognit Dev Syst"},{"issue":"4","key":"1385_CR11","doi-asserted-by":"publisher","first-page":"2568","DOI":"10.1109\/TAC.2022.3200962","volume":"68","author":"S Sui","year":"2023","unstructured":"Sui S, Chen CLP, Tong S (2023) A novel full errors fixed-time control for constraint nonlinear systems. IEEE Trans Autom Control 68(4):2568\u20132575. https:\/\/doi.org\/10.1109\/TAC.2022.3200962","journal-title":"IEEE Trans Autom Control"},{"key":"1385_CR12","doi-asserted-by":"publisher","DOI":"10.1007\/s40747-023-01135-y","author":"Z Peng","year":"2023","unstructured":"Peng Z, Song X, Song S, Stojanovic V (2023) Hysteresis quantified control for switched reaction\u2013diffusion systems and its application. Complex Intell Syst. https:\/\/doi.org\/10.1007\/s40747-023-01135-y","journal-title":"Complex Intell Syst"},{"key":"1385_CR13","doi-asserted-by":"publisher","first-page":"107491","DOI":"10.1016\/j.compeleceng.2021.107491","volume":"96","author":"C Liu","year":"2021","unstructured":"Liu C, Sun S, Tao C, Shou Y, Xu B (2021) Sliding mode control of multi-agent system with application to UAV air combat. Comput Electr Eng 96:107491","journal-title":"Comput Electr Eng"},{"issue":"1","key":"1385_CR14","doi-asserted-by":"publisher","first-page":"706","DOI":"10.1109\/TAES.2020.3029624","volume":"57","author":"H Duan","year":"2020","unstructured":"Duan H, Zhao J, Deng Y, Shi Y, Ding X (2020) Dynamic discrete pigeon-inspired optimization for multi-UAV cooperative search-attack mission planning. IEEE Trans Aerosp Electron Syst 57(1):706\u2013720","journal-title":"IEEE Trans Aerosp Electron Syst"},{"issue":"5","key":"1385_CR15","first-page":"5274","volume":"53","author":"B Wang","year":"2023","unstructured":"Wang B, Li S, Gao X, Xie T (2023) Weighted mean field reinforcement learning for large-scale UAV swarm confrontation. Appl Intell 53(5):5274\u20135289","journal-title":"Appl Intell"},{"key":"1385_CR16","doi-asserted-by":"crossref","unstructured":"Duan T, Wang W, Wang T (2022) A review for unmanned swarm gaming: framework, model and algorithm. In: 2022 8th International conference on big data and information analytics (BigDIA), IEEE. pp 164\u2013170","DOI":"10.1109\/BigDIA56350.2022.9874133"},{"issue":"4","key":"1385_CR17","doi-asserted-by":"publisher","first-page":"342","DOI":"10.1109\/TG.2021.3075065","volume":"13","author":"E Antonioni","year":"2021","unstructured":"Antonioni E, Suriani V, Riccio F, Nardi D (2021) Game strategies for physical robot soccer players: a survey. IEEE Trans Games 13(4):342\u2013357","journal-title":"IEEE Trans Games"},{"issue":"4","key":"1385_CR18","doi-asserted-by":"publisher","first-page":"2392","DOI":"10.1109\/TNSE.2022.3163447","volume":"9","author":"F Liu","year":"2022","unstructured":"Liu F, Dong X, Yu J, Hua Y, Li Q, Ren Z (2022) Distributed Nash equilibrium seeking of $$n$$-coalition noncooperative games with application to UAV swarms. IEEE Trans Netw Sci Eng 9(4):2392\u20132405","journal-title":"IEEE Trans Netw Sci Eng"},{"key":"1385_CR19","doi-asserted-by":"publisher","first-page":"332","DOI":"10.1007\/s10957-015-0706-z","volume":"168","author":"V Stojanovic","year":"2016","unstructured":"Stojanovic V, Nedic N (2016) A nature inspired parameter tuning approach to cascade control for hydraulically driven parallel robot platform. J Optim Theory Appl 168:332\u2013347","journal-title":"J Optim Theory Appl"},{"issue":"3","key":"1385_CR20","doi-asserted-by":"publisher","first-page":"1772","DOI":"10.3390\/su15031772","volume":"15","author":"C Chen","year":"2023","unstructured":"Chen C, Li Y, Cao G, Zhang J (2023) Research on dynamic scheduling model of plant protection UAV based on levy simulated annealing algorithm. Sustainability 15(3):1772","journal-title":"Sustainability"},{"issue":"2","key":"1385_CR21","doi-asserted-by":"publisher","first-page":"1260","DOI":"10.1109\/TITS.2020.3023819","volume":"23","author":"S Sun","year":"2022","unstructured":"Sun S, Song B, Wang P, Dong H, Chen X (2022) Real-time mission-motion planner for multi-UUVS cooperative work using tri-level programing. IEEE Trans Intell Transp Syst 23(2):1260\u20131273. https:\/\/doi.org\/10.1109\/TITS.2020.3023819","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"1","key":"1385_CR22","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1177\/0954406216662367","volume":"231","author":"D Pr\u0161i\u0107","year":"2017","unstructured":"Pr\u0161i\u0107 D, Nedi\u0107 N, Stojanovi\u0107 V (2017) A nature inspired optimal control of pneumatic-driven parallel robot platform. Proc Inst Mech Eng Part C J Mech Eng Sci 231(1):59\u201371","journal-title":"Proc Inst Mech Eng Part C J Mech Eng Sci"},{"key":"1385_CR23","doi-asserted-by":"publisher","first-page":"1085","DOI":"10.1007\/s00170-014-5735-5","volume":"72","author":"N Nedic","year":"2014","unstructured":"Nedic N, Prsic D, Dubonjic L, Stojanovic V, Djordjevic V (2014) Optimal cascade hydraulic control for a parallel robot platform by PSO. Int J Adv Manuf Technol 72:1085\u20131098","journal-title":"Int J Adv Manuf Technol"},{"key":"1385_CR24","doi-asserted-by":"crossref","unstructured":"Lei Y, Huo M, Deng Y, Duan H (2022) Multiple UAVS target allocation via stochastic dominant learning pigeon-inspired optimization in beyond-visual-range air combat. In: 2022 12th International conference on CYBER technology in automation, control, and intelligent systems (CYBER), pp 1269\u20131274","DOI":"10.1109\/CYBER55403.2022.9907711"},{"key":"1385_CR25","doi-asserted-by":"publisher","first-page":"110121","DOI":"10.1016\/j.oceaneng.2021.110121","volume":"242","author":"J Cai","year":"2021","unstructured":"Cai J, Zhang F, Sun S, Li T (2021) A meta-heuristic assisted underwater glider path planning method. Ocean Eng 242:110121","journal-title":"Ocean Eng"},{"key":"1385_CR26","doi-asserted-by":"publisher","first-page":"101169","DOI":"10.1016\/j.swevo.2022.101169","volume":"75","author":"C Chen","year":"2022","unstructured":"Chen C, Wang X, Dong H, Wang P (2022) Surrogate-assisted hierarchical learning water cycle algorithm for high-dimensional expensive optimization. Swarm Evol Comput 75:101169","journal-title":"Swarm Evol Comput"},{"issue":"18","key":"1385_CR27","doi-asserted-by":"publisher","first-page":"10139","DOI":"10.1002\/rnc.6354","volume":"32","author":"C Zhou","year":"2022","unstructured":"Zhou C, Tao H, Chen Y, Stojanovic V, Paszke W (2022) Robust point-to-point iterative learning control for constrained systems: a minimum energy approach. Int J Robust Nonlinear Control 32(18):10139\u201310161","journal-title":"Int J Robust Nonlinear Control"},{"issue":"15","key":"1385_CR28","doi-asserted-by":"publisher","first-page":"3177","DOI":"10.1080\/00207721.2022.2076171","volume":"53","author":"P Cheng","year":"2022","unstructured":"Cheng P, Wang H, Stojanovic V, Liu F, He S, Shi K (2022) Dissipativity-based finite-time asynchronous output feedback control for wind turbine system via a hidden Markov model. Int J Syst Sci 53(15):3177\u20133189","journal-title":"Int J Syst Sci"},{"issue":"7540","key":"1385_CR29","doi-asserted-by":"publisher","first-page":"529","DOI":"10.1038\/nature14236","volume":"518","author":"V Mnih","year":"2015","unstructured":"Mnih V, Kavukcuoglu K, Silver D, Rusu AA, Veness J, Bellemare MG, Graves A, Riedmiller M, Fidjeland AK, Ostrovski G et al (2015) Human-level control through deep reinforcement learning. Nature 518(7540):529\u2013533","journal-title":"Nature"},{"issue":"7676","key":"1385_CR30","doi-asserted-by":"publisher","first-page":"354","DOI":"10.1038\/nature24270","volume":"550","author":"D Silver","year":"2017","unstructured":"Silver D, Schrittwieser J, Simonyan K, Antonoglou I, Huang A, Guez A, Hubert T, Baker L, Lai M, Bolton A et al (2017) Mastering the game of go without human knowledge. Nature 550(7676):354\u2013359","journal-title":"Nature"},{"issue":"7782","key":"1385_CR31","doi-asserted-by":"publisher","first-page":"350","DOI":"10.1038\/s41586-019-1724-z","volume":"575","author":"O Vinyals","year":"2019","unstructured":"Vinyals O, Babuschkin I, Czarnecki WM, Mathieu M, Dudzik A, Chung J, Choi DH, Powell R, Ewalds T, Georgiev P et al (2019) Grandmaster level in Starcraft II using multi-agent reinforcement learning. Nature 575(7782):350\u2013354","journal-title":"Nature"},{"key":"1385_CR32","doi-asserted-by":"publisher","first-page":"67887","DOI":"10.1109\/ACCESS.2020.2985576","volume":"8","author":"Y Li","year":"2020","unstructured":"Li Y, Han W, Wang Y (2020) Deep reinforcement learning with application to air confrontation intelligent decision-making of manned\/unmanned aerial vehicle cooperative system. IEEE Access 8:67887\u201367898","journal-title":"IEEE Access"},{"issue":"10","key":"1385_CR33","doi-asserted-by":"publisher","first-page":"7900","DOI":"10.1109\/TNNLS.2022.3146976","volume":"34","author":"R Zhang","year":"2023","unstructured":"Zhang R, Zong Q, Zhang X, Dou L, Tian B (2023) Game of drones: Multi-UAV pursuit-evasion game with online motion planning by deep reinforcement learning. IEEE Trans Neural Netw Learn Syst 34(10):7900\u20137909. https:\/\/doi.org\/10.1109\/TNNLS.2022.3146976","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"issue":"12","key":"1385_CR34","doi-asserted-by":"publisher","first-page":"9477","DOI":"10.1109\/TPAMI.2021.3127674","volume":"44","author":"O Tutsoy","year":"2022","unstructured":"Tutsoy O (2022) Pharmacological, non-pharmacological policies and mutation: an artificial intelligence based multi-dimensional policy making algorithm for controlling the casualties of the pandemic diseases. IEEE Trans Pattern Anal Mach Intell 44(12):9477\u20139488. https:\/\/doi.org\/10.1109\/TPAMI.2021.3127674","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1385_CR35","unstructured":"Colas C, Sigaud O, Oudeyer P-Y (2018) Gep-pg: Decoupling exploration and exploitation in deep reinforcement learning algorithms. In: International conference on machine learning, PMLR. pp 1039\u20131048"},{"key":"1385_CR36","unstructured":"Lillicrap TP, Hunt JJ, Pritzel A, Heess N, Erez T, Tassa Y, Silver D, Wierstra D (2015)Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971"},{"key":"1385_CR37","unstructured":"Khadka S, Tumer K (2018) Evolution-guided policy gradient in reinforcement learning. In: Proceedings of the 32nd international conference on neural information processing systems. NIPS\u201918, Curran Associates Inc., Red Hook, NY, USA. pp. 1196\u20131208"},{"key":"1385_CR38","doi-asserted-by":"publisher","first-page":"105798","DOI":"10.1016\/j.engappai.2022.105798","volume":"120","author":"L Nugroho","year":"2023","unstructured":"Nugroho L, Andiarti R, Akmeliawati R, Kutay AT, Larasati DK, Wijaya SK (2023) Optimization of reward shaping function based on genetic algorithm applied to a cross validated deep deterministic policy gradient in a powered landing guidance problem. Eng Appl Artif Intell 120:105798","journal-title":"Eng Appl Artif Intell"},{"key":"1385_CR39","doi-asserted-by":"publisher","first-page":"228","DOI":"10.1016\/j.swevo.2018.03.011","volume":"44","author":"MM Drugan","year":"2019","unstructured":"Drugan MM (2019) Reinforcement learning versus evolutionary computation: a survey on hybrid algorithms. Swarm Evol Comput 44:228\u2013246","journal-title":"Swarm Evol Comput"},{"key":"1385_CR40","unstructured":"Khadka S, Majumdar S, Nassar T, Dwiel Z, Tumer E, Miret S, Liu Y, Tumer K (2019) Collaborative evolutionary reinforcement learning. In: International conference on machine learning, PMLR. pp 3341\u20133350"},{"key":"1385_CR41","doi-asserted-by":"publisher","unstructured":"Wang Z, Liu F, Guo J, Hong C, Chen M, Wang E, Zhao Y (2022) UAV swarm confrontation based on multi-agent deep reinforcement learning. In: 2022 41st Chinese control conference (CCC), pp 4996\u20135001 . https:\/\/doi.org\/10.23919\/CCC55666.2022.9902663","DOI":"10.23919\/CCC55666.2022.9902663"},{"key":"1385_CR42","doi-asserted-by":"crossref","unstructured":"Zhou H, Zhang X, Zhang Z (2022) Reinforcement learning technology for air combat confrontation of unmanned aerial vehicle. In: International conference on computer graphics, artificial intelligence, and data processing (ICCAID 2021), vol 12168. SPIE. pp 454\u2013459","DOI":"10.1117\/12.2631651"},{"key":"1385_CR43","doi-asserted-by":"crossref","unstructured":"Oroojlooy A, Hajinezhad D (2022) A review of cooperative multi-agent deep reinforcement learning. Appl Intell 53:13677\u201313722","DOI":"10.1007\/s10489-022-04105-y"},{"key":"1385_CR44","doi-asserted-by":"publisher","first-page":"945","DOI":"10.1007\/s10462-021-09997-9","volume":"55","author":"B Singh","year":"2022","unstructured":"Singh B, Kumar R, Singh VP (2022) Reinforcement learning in robotic applications: a comprehensive survey. Artif Intell Rev 55:945\u2013990","journal-title":"Artif Intell Rev"},{"key":"1385_CR45","unstructured":"Iqbal S, Sha F (2019) Actor-attention-critic for multi-agent reinforcement learning. In: International conference on machine learning, PMLR. pp 2961\u20132970"},{"key":"1385_CR46","unstructured":"Majumdar S, Khadka S, Miret S, McAleer S, Tumer K (2020) Evolutionary reinforcement learning for sample-efficient multiagent coordination. In: International conference on machine learning, PMLR. pp. 6651\u20136660"},{"key":"1385_CR47","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1016\/j.neunet.2022.05.013","volume":"153","author":"T Rupprecht","year":"2022","unstructured":"Rupprecht T, Wang Y (2022) A survey for deep reinforcement learning in Markovian cyber-physical systems: common problems and solutions. Neural Netw 153:13\u201336","journal-title":"Neural Netw"},{"key":"1385_CR48","doi-asserted-by":"publisher","first-page":"114896","DOI":"10.1016\/j.eswa.2021.114896","volume":"176","author":"L Huang","year":"2021","unstructured":"Huang L, Fu M, Qu H, Wang S, Hu S (2021) A deep reinforcement learning-based method applied for solving multi-agent defense and attack problems. Expert Syst Appl 176:114896","journal-title":"Expert Syst Appl"},{"key":"1385_CR49","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2021\/3360116","volume":"2021","author":"B Wang","year":"2021","unstructured":"Wang B, Li S, Gao X, Xie T (2021) UAV swarm confrontation using hierarchical multiagent reinforcement learning. Int J Aerosp Eng 2021:1\u201312","journal-title":"Int J Aerosp Eng"},{"key":"1385_CR50","doi-asserted-by":"crossref","unstructured":"Xuan S, Ke L (2022) UAV swarm attack-defense confrontation based on multi-agent reinforcement learning. In: Advances in guidance, navigation and control: proceedings of 2020 international conference on guidance, navigation and control, ICGNC 2020, Tianjin, China, October 23\u201325, 2020. Springer, pp 5599\u20135608","DOI":"10.1007\/978-981-15-8155-7_464"},{"issue":"2","key":"1385_CR51","doi-asserted-by":"publisher","first-page":"763","DOI":"10.1109\/TR.2022.3158279","volume":"71","author":"T Zhang","year":"2022","unstructured":"Zhang T, Chai L, Wang S, Jin J, Liu X, Song A, Lan Y (2022) Improving autonomous behavior strategy learning in an unmanned swarm system through knowledge enhancement. IEEE Trans Reliab 71(2):763\u2013774","journal-title":"IEEE Trans Reliab"},{"issue":"3","key":"1385_CR52","doi-asserted-by":"publisher","first-page":"401","DOI":"10.1109\/TAC.2005.864190","volume":"51","author":"R Olfati-Saber","year":"2006","unstructured":"Olfati-Saber R (2006) Flocking for multi-agent dynamic systems: algorithms and theory. IEEE Trans Autom Control 51(3):401\u2013420","journal-title":"IEEE Trans Autom Control"},{"issue":"3","key":"1385_CR53","doi-asserted-by":"publisher","first-page":"2927","DOI":"10.1109\/TIE.2022.3169825","volume":"70","author":"C Tang","year":"2023","unstructured":"Tang C, Zhang H-T, Wang J (2023) Flexible formation tracking control of multiple unmanned surface vessels for navigating through narrow channels with unknown curvatures. IEEE Trans Ind Electron 70(3):2927\u20132938. https:\/\/doi.org\/10.1109\/TIE.2022.3169825","journal-title":"IEEE Trans Ind Electron"},{"issue":"1","key":"1385_CR54","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1109\/TRO.2022.3200138","volume":"39","author":"M Lauri","year":"2023","unstructured":"Lauri M, Hsu D, Pajarinen J (2023) Partially observable Markov decision processes in robotics: a survey. IEEE Trans Robot 39(1):21\u201340","journal-title":"IEEE Trans Robot"},{"key":"1385_CR55","doi-asserted-by":"publisher","first-page":"108086","DOI":"10.1016\/j.asoc.2021.108086","volume":"115","author":"S Sun","year":"2022","unstructured":"Sun S, Song B, Wang P, Dong H, Chen X (2022) An adaptive bi-level task planning strategy for multi-USVS target visitation. Appl Soft Comput 115:108086","journal-title":"Appl Soft Comput"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01385-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-024-01385-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01385-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,16]],"date-time":"2024-05-16T18:18:34Z","timestamp":1715883514000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-024-01385-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,2,22]]},"references-count":55,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,6]]}},"alternative-id":["1385"],"URL":"https:\/\/doi.org\/10.1007\/s40747-024-01385-4","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,2,22]]},"assertion":[{"value":"29 June 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 January 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 February 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}