{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,26]],"date-time":"2025-08-26T07:00:26Z","timestamp":1756191626074,"version":"3.32.0"},"reference-count":32,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,12,12]],"date-time":"2024-12-12T00:00:00Z","timestamp":1733961600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,12,12]],"date-time":"2024-12-12T00:00:00Z","timestamp":1733961600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62373273,62073234,62373268,62103417"],"award-info":[{"award-number":["62373273,62073234,62373268,62103417"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,12,12]]},"DOI":"10.1109\/icarcv63323.2024.10821702","type":"proceedings-article","created":{"date-parts":[[2025,1,9]],"date-time":"2025-01-09T19:36:27Z","timestamp":1736451387000},"page":"226-231","source":"Crossref","is-referenced-by-count":1,"title":["Unmanned Aerial Vehicle Defense Penetration on a Digital-Twin-City Based on an Improved MASAC Reinforcement Learning"],"prefix":"10.1109","author":[{"given":"Da","family":"Liu","sequence":"first","affiliation":[{"name":"School of Electrical and Information Engineering, Tianjin University,Tianjin,China,300072"}]},{"given":"Ruilong","family":"Zhang","sequence":"additional","affiliation":[{"name":"Beijing Aerospace Automatic Control Institute,Beijing,China,100143"}]},{"given":"Ping","family":"Song","sequence":"additional","affiliation":[{"name":"China Helicopter Research and Development Institute,Jingdezhen,Jiangxi,China,333000"}]},{"given":"Xiuyun","family":"Zhang","sequence":"additional","affiliation":[{"name":"School of Electrical and Information Engineering, Tianjin University,Tianjin,China,300072"}]},{"given":"Weilun","family":"Wang","sequence":"additional","affiliation":[{"name":"School of Electrical and Information Engineering, Tianjin University,Tianjin,China,300072"}]},{"given":"Qun","family":"Zong","sequence":"additional","affiliation":[{"name":"School of Electrical and Information Engineering, Tianjin University,Tianjin,China,300072"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2019.2924143"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/SECon.2012.6196930"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/LCOMM.2016.2524405"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-022-10281-7"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.2514\/6.2008-7134"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICCCS52626.2021.9449283"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TCSII.2021.3112787"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TIV.2024.3399231"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/s10458-019-09421-1"},{"key":"ref10","article-title":"Playing atari with deep reinforcement learning","author":"Mnih","year":"2013","journal-title":"arXiv preprint"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.13140\/RG.2.2.18893.74727"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-019-1724-z"},{"key":"ref13","article-title":"Starcraft ii: A new challenge for reinforcement learning","author":"Vinyals","year":"2017","journal-title":"arXiv preprint"},{"key":"ref14","article-title":"Continuous control with deep reinforce-ment learning","author":"Lillicrap","year":"2015","journal-title":"arXiv preprint"},{"key":"ref15","first-page":"1587","article-title":"Addressing function approximation error in actor-critic methods","volume-title":"International conference on machine learning","author":"Fujimoto"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1812.05905"},{"key":"ref17","article-title":"Multi-agent actor-critic for mixed cooperative-competitive environments","volume":"30","author":"Lowe","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11794"},{"key":"ref19","article-title":"Learning multiagent communication with backpropagation","volume":"29","author":"Sukhbaatar","year":"2016","journal-title":"Advances in neural information processing systems"},{"key":"ref20","first-page":"4295","article-title":"Qmix: Monotonic value function factorisation for deep multi-agent reinforcement learning","volume-title":"International Conference on Machine Learning","author":"Rashid"},{"key":"ref21","first-page":"5887","article-title":"Qtran: Learning to factorize with transformation for cooperative multi-agent reinforcement learning","volume-title":"International Conference on Machine Learning","author":"Son"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1016\/j.ast.2019.06.024"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/s42979-020-00326-5"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TNSE.2023.3330266"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.2514\/6.2024-0121"},{"article-title":"Dop: Off-policy multi-agent decomposed policy gradients","volume-title":"International Conference on Learning Representations","author":"Wang","key":"ref26"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1016\/j.conb.2012.05.008"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/FUZZ-IEEE.2016.7737744"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3146976"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TETCI.2024.3360282"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2009.03.012"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/IMCEC51613.2021.9482235"}],"event":{"name":"2024 18th International Conference on Control, Automation, Robotics and Vision (ICARCV)","start":{"date-parts":[[2024,12,12]]},"location":"Dubai, United Arab Emirates","end":{"date-parts":[[2024,12,15]]}},"container-title":["2024 18th International Conference on Control, Automation, Robotics and Vision (ICARCV)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10821514\/10821497\/10821702.pdf?arnumber=10821702","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,10]],"date-time":"2025-01-10T05:46:45Z","timestamp":1736488005000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10821702\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,12]]},"references-count":32,"URL":"https:\/\/doi.org\/10.1109\/icarcv63323.2024.10821702","relation":{},"subject":[],"published":{"date-parts":[[2024,12,12]]}}}