{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,17]],"date-time":"2026-03-17T19:59:40Z","timestamp":1773777580916,"version":"3.50.1"},"reference-count":28,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"2","license":[{"start":{"date-parts":[[2023,2,1]],"date-time":"2023-02-01T00:00:00Z","timestamp":1675209600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2023,2,1]],"date-time":"2023-02-01T00:00:00Z","timestamp":1675209600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,2,1]],"date-time":"2023-02-01T00:00:00Z","timestamp":1675209600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61906138"],"award-info":[{"award-number":["61906138"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shanghai Municipal Science and Technology Major Project","award":["2018SHZDZX01"],"award-info":[{"award-number":["2018SHZDZX01"]}]},{"name":"ZJ Lab"},{"DOI":"10.13039\/100020441","name":"Shanghai Center for Brain Science and Brain-Inspired Technology","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100020441","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shanghai Rising Star Program","award":["21QC1400900"],"award-info":[{"award-number":["21QC1400900"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Veh. Technol."],"published-print":{"date-parts":[[2023,2]]},"DOI":"10.1109\/tvt.2022.3212996","type":"journal-article","created":{"date-parts":[[2022,10,10]],"date-time":"2022-10-10T20:13:57Z","timestamp":1665432837000},"page":"2593-2598","source":"Crossref","is-referenced-by-count":34,"title":["A Discrete Soft Actor-Critic Decision-Making Strategy With Sample Filter for Freeway Autonomous Driving"],"prefix":"10.1109","volume":"72","author":[{"given":"Jiayi","family":"Guan","sequence":"first","affiliation":[{"name":"Tongji University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7416-592X","authenticated-orcid":false,"given":"Guang","family":"Chen","sequence":"additional","affiliation":[{"name":"Tongji University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8774-2936","authenticated-orcid":false,"given":"Jin","family":"Huang","sequence":"additional","affiliation":[{"name":"School of Vehicle and Mobility, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3909-488X","authenticated-orcid":false,"given":"Zhijun","family":"Li","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1673-2658","authenticated-orcid":false,"given":"Lu","family":"Xiong","sequence":"additional","affiliation":[{"name":"Tongji University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4778-137X","authenticated-orcid":false,"given":"Jing","family":"Hou","sequence":"additional","affiliation":[{"name":"Tongji University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4840-076X","authenticated-orcid":false,"given":"Alois","family":"Knoll","sequence":"additional","affiliation":[{"name":"Chair of Robotics, Artificial Intelligence and Real-time Systems, Technical University of Munich, Munich, Germany"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2020.2986005"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2018.2822762"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8461233"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/s42154-021-00157-x"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2021.3059050"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2020.3024655"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/JAS.2020.1003294"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1007\/s42154-021-00151-3"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.23919\/ACC45564.2020.9147626"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2021.3102407"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.3027047"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1049\/iet-its.2019.0826"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48506.2021.9562006"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TCDS.2021.3092715"},{"key":"ref15","article-title":"Soft actor-critic for discrete action settings","author":"Christodoulou","year":"2019"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.23919\/ICCAS47443.2019.8971629"},{"issue":"1","key":"ref17","article-title":"Accelerating reinforcement learning with prioritized experience replay for maze game","volume":"3","author":"Hu","year":"2020","journal-title":"SMU Data Sci. Rev."},{"key":"ref18","first-page":"449","article-title":"A distributional perspective on reinforcement learning","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Bellemare","year":"2017"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3082568"},{"key":"ref20","first-page":"1861","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Haarnoja","year":"2018"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1115\/DSCC2017-5099"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/IV47402.2020.9304744"},{"key":"ref23","first-page":"264","article-title":"Smarts: Scalable multi-agent reinforcement learning training school for autonomous driving","volume-title":"Proc. 4th Conf. Robot Learn.","author":"Zhou","year":"2020"},{"key":"ref24","article-title":"Proximal policy optimization algorithms","author":"Schulman","year":"2017"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v30i1.10295"},{"key":"ref26","first-page":"1","article-title":"Prioritized experience replay","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Schaul","year":"2016"},{"key":"ref27","first-page":"1928","article-title":"Asynchronous methods for deep reinforcement learning","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Mnih","year":"2016"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1115\/DSCC2017-5209"}],"container-title":["IEEE Transactions on Vehicular Technology"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/25\/10043591\/09914655.pdf?arnumber=9914655","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T22:16:40Z","timestamp":1705961800000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9914655\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,2]]},"references-count":28,"journal-issue":{"issue":"2"},"URL":"https:\/\/doi.org\/10.1109\/tvt.2022.3212996","relation":{},"ISSN":["0018-9545","1939-9359"],"issn-type":[{"value":"0018-9545","type":"print"},{"value":"1939-9359","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,2]]}}}