{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,22]],"date-time":"2025-03-22T04:20:07Z","timestamp":1742617207455,"version":"3.40.2"},"reference-count":29,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,9,24]],"date-time":"2024-09-24T00:00:00Z","timestamp":1727136000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,9,24]],"date-time":"2024-09-24T00:00:00Z","timestamp":1727136000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,9,24]]},"DOI":"10.1109\/itsc58415.2024.10919757","type":"proceedings-article","created":{"date-parts":[[2025,3,21]],"date-time":"2025-03-21T19:00:11Z","timestamp":1742583611000},"page":"3895-3900","source":"Crossref","is-referenced-by-count":0,"title":["Hardware-Compatible Deep Reinforcement Learning-Based Lateral Trajectory Controller"],"prefix":"10.1109","author":[{"given":"Mohamed","family":"Abouelnaga","sequence":"first","affiliation":[{"name":"TU Chemnitz,Department of Computer Science,Germany"}]},{"given":"Mathias","family":"Haberjahn","sequence":"additional","affiliation":[{"name":"2Infineon Technologies Dresden GmbH,Germany"}]},{"given":"Daniel","family":"Markert","sequence":"additional","affiliation":[{"name":"TU Chemnitz,Department of Computer Science,Germany"}]},{"given":"Alejandro","family":"Masrur","sequence":"additional","affiliation":[{"name":"TU Chemnitz,Department of Computer Science,Germany"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-662-49275-8_39"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-72892-6_3"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/j.conengprac.2011.12.004"},{"journal-title":"Dream to control: Learning behaviors by latent imagination","year":"2020","author":"Hafner","key":"ref4"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/1-84628-148-2"},{"key":"ref6","first-page":"1587","article-title":"Addressing function approximation error in actor-critic methods","volume-title":"International conference on machine learning","author":"Fujimoto"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1812.05905"},{"volume-title":"AURIX\u2122 TC4x","year":"2023","author":"Technologies","key":"ref8"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/IVS.2018.8500556"},{"key":"ref10","article-title":"Deep reinforcement learning for simulated autonomous vehicle control","volume-title":"Course Project Reports: Winter","volume":"2016","author":"Yu","year":"2016"},{"key":"ref11","article-title":"End-to-end deep reinforcement learning for lane keeping assist","author":"Sallab","year":"2016","journal-title":"arXiv preprint"},{"key":"ref12","volume-title":"Introduction to reinforcement learning","volume":"135","author":"Sutton","year":"1998"},{"key":"ref13","article-title":"Playing atari with deep reinforcement learning","author":"Mnih","year":"2013","journal-title":"arXiv preprint"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/MCI.2019.2901089"},{"key":"ref15","first-page":"387","article-title":"Deterministic policy gradient algorithms","volume-title":"International conference on machine learning","author":"Silver"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2020.2967299"},{"key":"ref17","article-title":"Continuous control with deep reinforcement learning","author":"Lillicrap","year":"2015","journal-title":"arXiv preprint"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2020.103799"},{"key":"ref19","first-page":"277","article-title":"Optimising pid control with residual policy reinforcement learning","author":"Hynes","year":"2020","journal-title":"AICS"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.3390\/sym14010031"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/BF00992698"},{"key":"ref22","article-title":"Policy gradient methods for reinforcement learning with function approximation","volume":"12","author":"Sutton","year":"1999","journal-title":"Advances in neural information processing systems"},{"volume-title":"CarMaker","year":"2023","author":"Automotive","key":"ref23"},{"volume-title":"MATLAB\/Simulink","year":"2023","author":"Inc","key":"ref24"},{"article-title":"MetaWare NN SDK","volume-title":"Synopsys","year":"2023","key":"ref25"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.17148\/IARJSET.2015.2305"},{"volume-title":"Hockenheimring","year":"2023","key":"ref27"},{"volume-title":"N\u00fcrburgring","year":"2023","key":"ref28"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1115\/1.4019264"}],"event":{"name":"2024 IEEE 27th International Conference on Intelligent Transportation Systems (ITSC)","start":{"date-parts":[[2024,9,24]]},"location":"Edmonton, AB, Canada","end":{"date-parts":[[2024,9,27]]}},"container-title":["2024 IEEE 27th International Conference on Intelligent Transportation Systems (ITSC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10919469\/10919190\/10919757.pdf?arnumber=10919757","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,21]],"date-time":"2025-03-21T23:16:10Z","timestamp":1742598970000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10919757\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,24]]},"references-count":29,"URL":"https:\/\/doi.org\/10.1109\/itsc58415.2024.10919757","relation":{},"subject":[],"published":{"date-parts":[[2024,9,24]]}}}