{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,29]],"date-time":"2024-10-29T20:45:49Z","timestamp":1730234749426,"version":"3.28.0"},"reference-count":10,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,9,28]],"date-time":"2020-09-28T00:00:00Z","timestamp":1601251200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,9,28]],"date-time":"2020-09-28T00:00:00Z","timestamp":1601251200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,9,28]],"date-time":"2020-09-28T00:00:00Z","timestamp":1601251200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,9,28]]},"DOI":"10.1109\/icce-taiwan49838.2020.9258271","type":"proceedings-article","created":{"date-parts":[[2020,11,23]],"date-time":"2020-11-23T18:51:00Z","timestamp":1606157460000},"page":"1-2","source":"Crossref","is-referenced-by-count":0,"title":["Position-Aware Communication via Self-Attention for Multi-Agent Reinforcement Learning"],"prefix":"10.1109","author":[{"given":"Tsan-Hua","family":"Shih","sequence":"first","affiliation":[]},{"given":"Hsien-I","family":"Lin","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref4","article-title":"Learning multiagent communication with backpropagation","author":"sukhbaatar","year":"0","journal-title":"NIPS"},{"key":"ref3","article-title":"Learning to communicate with deep multi-agent reinforcement learning","author":"foerster","year":"0","journal-title":"NIPS"},{"key":"ref10","article-title":"Multi-agent actor-critic for mixed cooperative-competitive environments","author":"lowe","year":"0","journal-title":"NIPS"},{"key":"ref6","article-title":"Attention is all you need","author":"vaswani","year":"0","journal-title":"NIPS"},{"key":"ref5","article-title":"Learning Attentional Communication for Multi-Agent Cooperation","author":"jiang","year":"0","journal-title":"NIPS"},{"journal-title":"Value-Decomposition Networks For Cooperative Multi-Agent Learning","year":"2017","author":"sunehag","key":"ref8"},{"key":"ref7","article-title":"Deep recurrent Q-learning for partially observable MDPs","volume":"abs 1507 6527","author":"hausknecht","year":"2015","journal-title":"CoRR"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1017\/S0269888912000057"},{"year":"0","key":"ref9"},{"key":"ref1","article-title":"A unified game-theoretic approach to multiagent reinforcement learning","author":"lanctot","year":"0","journal-title":"NIPS"}],"event":{"name":"2020 IEEE International Conference on Consumer Electronics - Taiwan (ICCE-Taiwan)","start":{"date-parts":[[2020,9,28]]},"location":"Taoyuan, Taiwan","end":{"date-parts":[[2020,9,30]]}},"container-title":["2020 IEEE International Conference on Consumer Electronics - Taiwan (ICCE-Taiwan)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9257985\/9257988\/09258271.pdf?arnumber=9258271","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,27]],"date-time":"2022-06-27T20:27:40Z","timestamp":1656361660000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9258271\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,9,28]]},"references-count":10,"URL":"https:\/\/doi.org\/10.1109\/icce-taiwan49838.2020.9258271","relation":{},"subject":[],"published":{"date-parts":[[2020,9,28]]}}}