{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T16:25:58Z","timestamp":1775665558335,"version":"3.50.1"},"reference-count":27,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,12,10]],"date-time":"2023-12-10T00:00:00Z","timestamp":1702166400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,12,10]],"date-time":"2023-12-10T00:00:00Z","timestamp":1702166400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,12,10]]},"DOI":"10.1109\/wsc60868.2023.10408232","type":"proceedings-article","created":{"date-parts":[[2024,1,31]],"date-time":"2024-01-31T18:30:14Z","timestamp":1706725814000},"page":"3035-3046","source":"Crossref","is-referenced-by-count":2,"title":["Learning Environment For The Air Domain (Lead)"],"prefix":"10.1109","author":[{"given":"Andreas","family":"Strand","sequence":"first","affiliation":[{"name":"FFI &#x2013; Norwegian Defence Research Establishment,Kjeller,Norway,2007"}]},{"given":"Patrick","family":"Gorton","sequence":"additional","affiliation":[{"name":"FFI &#x2013; Norwegian Defence Research Establishment,Kjeller,Norway,2007"}]},{"given":"Martin","family":"Asprusten","sequence":"additional","affiliation":[{"name":"FFI &#x2013; Norwegian Defence Research Establishment,Kjeller,Norway,2007"}]},{"given":"Karsten","family":"Brathen","sequence":"additional","affiliation":[{"name":"FFI &#x2013; Norwegian Defence Research Establishment,Kjeller,Norway,2007"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330701"},{"issue":"06680","key":"ref2","first-page":"1","article-title":"Dota 2 with Large Scale Deep Reinforcement Learning","volume":"1912","author":"Berner","year":"2019"},{"key":"ref3","article-title":"OpenAI Gym","volume-title":"Technical report, OpenAI","author":"Brockman","year":"2016"},{"key":"ref4","volume-title":"Reinforcement Learning with TensorFlow","author":"Dutta","year":"2018"},{"key":"ref5","volume-title":"Parallel and Distribution Simulation Systems","author":"Fujimoto","year":"1999"},{"key":"ref6","article-title":"Imitation Learning for Modelling Air Combat Behaviour","volume-title":"FFI-RAPPORT","author":"Gorton","year":"2022"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1145\/3512290.3528827"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3060426"},{"key":"ref9","article-title":"STD-1516 Standard for Modeling and Simulation (M&S) High Level Architecture (HLA) \u2013 Framework and Rules","volume-title":"Standard","year":"2010"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1017\/aer.2022.8"},{"key":"ref11","volume-title":"Creating Computer Simulation Systems","author":"Kuhl","year":"1999"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CoG51982.2022.9893690"},{"key":"ref13","first-page":"3053","article-title":"RLlib: Abstractions for Distributed Reinforcement Learning","volume-title":"Proceedings of the 35th International Conference on Machine Learning","volume":"80","author":"Liang"},{"key":"ref14","article-title":"Data-Driven Behavior Modeling for Computer Generated Forces","volume-title":"FFI-RAPPORT","author":"L\u00f8vlid","year":"2017"},{"key":"ref15","article-title":"VR-Forces Capabilities","volume-title":"Technical Report","year":"2021"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN48605.2020.9207088"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICUAS51884.2021.9476700"},{"issue":"268","key":"ref18","first-page":"1","article-title":"Stable-Baselines3: Reliable Reinforcement Learning Implementations","volume":"22","author":"Raffin","year":"2021","journal-title":"Journal of Machine Learning Research"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICUAS54217.2022.9836131"},{"issue":"06347","key":"ref20","first-page":"1","article-title":"Proximal Policy Optimization Algorithms","volume":"1707","author":"Schulman","year":"2017"},{"key":"ref21","article-title":"STD-001.1 Standard for Real-time Platform Reference Federation Object Model","volume-title":"Standard","year":"2015"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/LARS\/SBR\/WRE51543.2020.9306947"},{"key":"ref23","article-title":"Calculated Moves","volume-title":"Ph. D. thesis","author":"Toubman","year":"2020"},{"key":"ref24","first-page":"4015","article-title":"On the Importance of Hyperparameter Optimization for Model-based Reinforcement Learning","volume-title":"Proceedings of The 24th International Conference on Artificial Intelligence and Statistics","volume":"130","author":"Zhang"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.3390\/app122010230"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.7249\/rr4311"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2023.3292075"}],"event":{"name":"2023 Winter Simulation Conference (WSC)","location":"San Antonio, TX, USA","start":{"date-parts":[[2023,12,10]]},"end":{"date-parts":[[2023,12,13]]}},"container-title":["2023 Winter Simulation Conference (WSC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10406299\/10407113\/10408232.pdf?arnumber=10408232","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,2]],"date-time":"2024-02-02T00:14:21Z","timestamp":1706832861000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10408232\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,10]]},"references-count":27,"URL":"https:\/\/doi.org\/10.1109\/wsc60868.2023.10408232","relation":{},"subject":[],"published":{"date-parts":[[2023,12,10]]}}}