{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T01:13:04Z","timestamp":1740100384924,"version":"3.37.3"},"reference-count":31,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,11,4]],"date-time":"2021-11-04T00:00:00Z","timestamp":1635984000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,11,4]],"date-time":"2021-11-04T00:00:00Z","timestamp":1635984000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,11,4]],"date-time":"2021-11-04T00:00:00Z","timestamp":1635984000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["IIS-1815886"],"award-info":[{"award-number":["IIS-1815886"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000181","name":"Air Force Office of Scientific Research","doi-asserted-by":"publisher","award":["FA9550-19-1-0195"],"award-info":[{"award-number":["FA9550-19-1-0195"]}],"id":[{"id":"10.13039\/100000181","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,11,4]]},"DOI":"10.1109\/mrs50823.2021.9620689","type":"proceedings-article","created":{"date-parts":[[2021,12,7]],"date-time":"2021-12-07T20:49:53Z","timestamp":1638910193000},"page":"92-100","source":"Crossref","is-referenced-by-count":1,"title":["Reactive Multi-Fitness Learning for Robust Multiagent Teaming"],"prefix":"10.1109","author":[{"given":"Connor","family":"Yates","sequence":"first","affiliation":[]},{"given":"Ayhan Alp","family":"Aydeniz","sequence":"additional","affiliation":[]},{"given":"Kagan","family":"Tumer","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"journal-title":"Neat python","year":"0","author":"mcintyre","key":"ref31"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/s10458-008-9046-9"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1001"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2013.6639012"},{"key":"ref12","first-page":"527","article-title":"Multi-task learning as multi-objective optimization","author":"sener","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref13","first-page":"54","volume":"113","author":"parisi","year":"0","journal-title":"Continual lifelong learning with neural networks A review"},{"key":"ref14","first-page":"3796","volume":"33","author":"hessel","year":"0","journal-title":"Multi-task deep reinforcement learning with PopArt"},{"key":"ref15","first-page":"9","author":"rusu","year":"0","journal-title":"Sim-to-real robot learning from pixels with progressive nets"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/70.294207"},{"key":"ref17","first-page":"1311","article-title":"Automated curriculum learning for neural networks","author":"graves","year":"0","journal-title":"Int Conference on Machine Learning"},{"journal-title":"Neural programmer-interpreters","year":"0","author":"reed","key":"ref18"},{"key":"ref19","first-page":"249","article-title":"Transfer learning in multi-agent reinforcement learning domains","author":"boutsioukis","year":"2011","journal-title":"Proc of European Workshop on Reinforcement Learning"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2019.09.060"},{"key":"ref4","article-title":"Aligning social welfare and agent preferences to alleviate traffic congestion","author":"tumer","year":"2008","journal-title":"Proc Int Joint Conf Autonomous Agents and Multiagent Systems"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1016\/j.ress.2005.11.018"},{"key":"ref3","first-page":"5","article-title":"Efficient objective functions for coordinated learning in large-scale distributed osa systems","volume":"12","author":"oliaee","year":"2013","journal-title":"IEEE Transactions on Mobile Computing"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1145\/3377930.3390220"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2018.2881227"},{"key":"ref5","article-title":"Regulating air traffic flow with coupled agents","author":"agogino","year":"2008","journal-title":"Proc Int Joint Conf Autonomous Agents and Multiagent Systems"},{"key":"ref8","first-page":"181","volume":"112","author":"sutton","year":"0","journal-title":"Between MDPs and Semi-MDPs A Framework for Temporal Abstraction in Reinforcement Learning"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-49774-5_16"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-05816-6_33"},{"key":"ref1","article-title":"First results of the robex analogue mission campaign: Robotic deployment of seismic networks for future lunar missions","volume":"68","author":"wedler","year":"0","journal-title":"Proceedings of the International Astronautical Congress IAC"},{"key":"ref9","first-page":"502","volume":"50","author":"shoeleh","year":"0","journal-title":"Skill based transfer learning with domain adaptation for continuous reinforcement learning domains"},{"key":"ref20","first-page":"2681","article-title":"Deep decentralized multi-task multi-agent reinforcement learning under partial observability","author":"omidshafiei","year":"0","journal-title":"Proceedings of the 34th International Conference on Machine Learning-Volume 70 JMLR org"},{"key":"ref22","first-page":"1","article-title":"Tunable dynamics in agent-based simulation using multi-objective reinforcement learning","author":"kallstrom","year":"2019","journal-title":"Adaptive and Learning Agents Workshop (ALA-19) at AAMAS Montreal Canada"},{"key":"ref21","article-title":"Learning and transferring roles in multi-agent reinforcement","author":"wilson","year":"0","journal-title":"Proc AAAI-08 Workshop on Transfer Learning for Complex Tasks"},{"key":"ref24","first-page":"5392","article-title":"Hybrid reward architecture for reinforcement learning","author":"van seijen","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref23","article-title":"Value-Decomposition Networks For Cooperative Multi-Agent Learning Based On Team Reward","author":"sunehag","year":"0","journal-title":"Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems ser AAMAS &#x2018;18 Richland SC International Foundation for Autonomous Agents and Multiagent Systems 2018 pp 2085&#x2013;2087 event-place Stockholm Sweden"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CEC.2008.4631247"},{"key":"ref25","volume":"31","author":"tessler","year":"0","journal-title":"A deep hierarchical approach to lifelong learning in minecraft"}],"event":{"name":"2021 International Symposium on Multi-Robot and Multi-Agent Systems (MRS)","start":{"date-parts":[[2021,11,4]]},"location":"Cambridge, United Kingdom","end":{"date-parts":[[2021,11,5]]}},"container-title":["2021 International Symposium on Multi-Robot and Multi-Agent Systems (MRS)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9620207\/9620540\/09620689.pdf?arnumber=9620689","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T16:53:33Z","timestamp":1652201613000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9620689\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,11,4]]},"references-count":31,"URL":"https:\/\/doi.org\/10.1109\/mrs50823.2021.9620689","relation":{},"subject":[],"published":{"date-parts":[[2021,11,4]]}}}