{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T03:20:39Z","timestamp":1730344839938,"version":"3.28.0"},"reference-count":20,"publisher":"IEEE","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,9]]},"DOI":"10.23919\/sice.2019.8859883","type":"proceedings-article","created":{"date-parts":[[2019,10,8]],"date-time":"2019-10-08T00:45:33Z","timestamp":1570495533000},"page":"1456-1461","source":"Crossref","is-referenced-by-count":4,"title":["Self Training Autonomous Driving Agent"],"prefix":"10.23919","author":[{"given":"Shashank","family":"Kotyan","sequence":"first","affiliation":[]},{"given":"Danilo Vasconcellos","family":"Vargas","sequence":"additional","affiliation":[]},{"given":"U.","family":"Venkanna","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"journal-title":"The CMA evolution strategy A tutorial","year":"2016","author":"hansen","key":"ref10"},{"journal-title":"OpenAI Gym","year":"2016","author":"brockman","key":"ref11"},{"journal-title":"Open Source Self-Driving Car","year":"2017","key":"ref12"},{"journal-title":"CARLA An Open Urban Driving Simulator","year":"2017","author":"dosovitskiy","key":"ref13"},{"key":"ref14","volume":"4","author":"wymann","year":"2000","journal-title":"TORCS The Open Racing Car Simulator"},{"journal-title":"Recurrent world models facilitate policy evolution","year":"2018","author":"ha","key":"ref15"},{"journal-title":"Task-Relevant Embeddings for Robust Perception in Reinforcement Learning","year":"2018","author":"liang","key":"ref16"},{"journal-title":"Deep-Q Learning for racecar reinforcement learning problem","year":"2017","author":"prieur","key":"ref17"},{"journal-title":"Reinforcement car racing with A3C","year":"2017","author":"jang","key":"ref18"},{"journal-title":"Car racing using reinforcement learning","year":"2016","author":"khan","key":"ref19"},{"key":"ref4","article-title":"A fuzzy controller with supervised learning assisted reinforcement learning algorithm for obstacle avoidance","author":"ye","year":"2003","journal-title":"IEEE Transactions on Systems Man and Cybernetics Part B Cybernetics"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.5194\/ars-3-205-2005"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1613\/jair.301"},{"journal-title":"Reinforcement Learning An Introduction","year":"2018","author":"sutton","key":"ref5"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/2463372.2463509"},{"journal-title":"Continuous control with deep reinforcement learning","year":"2015","author":"lillicrap","key":"ref7"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/IVS.2013.6629559"},{"journal-title":"self-driving car Definition from PC Magazine Encyclopedia","year":"0","key":"ref1"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1145\/2576768.2598358"},{"journal-title":"CarRacing-v0","year":"2016","author":"klimov","key":"ref20"}],"event":{"name":"2019 58th Annual Conference of the Society of Instrument and Control Engineers of Japan (SICE)","start":{"date-parts":[[2019,9,10]]},"location":"Hiroshima, Japan","end":{"date-parts":[[2019,9,13]]}},"container-title":["2019 58th Annual Conference of the Society of Instrument and Control Engineers of Japan (SICE)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8852508\/8859743\/08859883.pdf?arnumber=8859883","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,10,29]],"date-time":"2019-10-29T00:35:16Z","timestamp":1572309316000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8859883\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,9]]},"references-count":20,"URL":"https:\/\/doi.org\/10.23919\/sice.2019.8859883","relation":{},"subject":[],"published":{"date-parts":[[2019,9]]}}}