{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T19:12:58Z","timestamp":1762024378110,"version":"build-2065373602"},"reference-count":12,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,7,1]],"date-time":"2020-07-01T00:00:00Z","timestamp":1593561600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,7,1]],"date-time":"2020-07-01T00:00:00Z","timestamp":1593561600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,7,1]],"date-time":"2020-07-01T00:00:00Z","timestamp":1593561600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,7]]},"DOI":"10.1109\/aim43001.2020.9158839","type":"proceedings-article","created":{"date-parts":[[2020,8,5]],"date-time":"2020-08-05T21:18:31Z","timestamp":1596662311000},"page":"1335-1340","source":"Crossref","is-referenced-by-count":6,"title":["Reinforcement Learning with Imitation for Cavity Filter Tuning"],"prefix":"10.1109","author":[{"given":"Simon","family":"Lindstah","sequence":"first","affiliation":[]},{"given":"Xiaoyu","family":"Lan","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref4","article-title":"Fast policy learning through imitation and reinforcement","author":"cheng","year":"2018","journal-title":"In arXiv preprint arXiv 1805 10413"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989385"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/WCICA.2016.7578618"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2018\/687"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.2528\/PIERM10053105"},{"key":"ref5","article-title":"Reinforcement and imitation learning via interactive no-regret learning","author":"ross","year":"2014","journal-title":"arXiv preprint arXiv 1406 5979"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.2307\/2276774"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CBS.2018.8612197"},{"key":"ref7","article-title":"Continuous control with deep reinforcement learning","author":"lillicrap","year":"2016","journal-title":"In International Conference on Learning Representations"},{"key":"ref2","first-page":"354","article-title":"Mastering the game of go without human knowledge","author":"silver","year":"2017","journal-title":"Nature 550 7676"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ROBIO.2015.7419091"},{"key":"ref1","article-title":"Playing atari with deep reinforcement learning","author":"mnih","year":"2013","journal-title":"arXiv preprint arXiv 1312 5602"}],"event":{"name":"2020 IEEE\/ASME International Conference on Advanced Intelligent Mechatronics (AIM)","start":{"date-parts":[[2020,7,6]]},"location":"Boston, MA, USA","end":{"date-parts":[[2020,7,9]]}},"container-title":["2020 IEEE\/ASME International Conference on Advanced Intelligent Mechatronics (AIM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9149748\/9158642\/09158839.pdf?arnumber=9158839","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,27]],"date-time":"2022-06-27T15:47:11Z","timestamp":1656344831000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9158839\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,7]]},"references-count":12,"URL":"https:\/\/doi.org\/10.1109\/aim43001.2020.9158839","relation":{},"subject":[],"published":{"date-parts":[[2020,7]]}}}