{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T03:22:01Z","timestamp":1775186521505,"version":"3.50.1"},"reference-count":57,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"1","license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001475","name":"Nanyang Technological University","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001475","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001348","name":"Agency for Science, Technology and Research","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001348","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Advanced Manufacturing and Engineering (AME) Young Individual Research","award":["A2084c0156"],"award-info":[{"award-number":["A2084c0156"]}]},{"name":"MTC Individual Research","award":["M22K2c0079"],"award-info":[{"award-number":["M22K2c0079"]}]},{"name":"ANR-NRF Joint","award":["NRF2021-NRF-ANR003"],"award-info":[{"award-number":["NRF2021-NRF-ANR003"]}]},{"DOI":"10.13039\/100010449","name":"Ministry of Education","doi-asserted-by":"publisher","award":["MOE-T2EP50222-0002"],"award-info":[{"award-number":["MOE-T2EP50222-0002"]}],"id":[{"id":"10.13039\/100010449","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2024,1]]},"DOI":"10.1109\/tpami.2023.3322426","type":"journal-article","created":{"date-parts":[[2023,10,6]],"date-time":"2023-10-06T13:52:47Z","timestamp":1696600367000},"page":"267-279","source":"Crossref","is-referenced-by-count":68,"title":["Fear-Neuro-Inspired Reinforcement Learning for Safe Autonomous Driving"],"prefix":"10.1109","volume":"46","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9818-0879","authenticated-orcid":false,"given":"Xiangkun","family":"He","sequence":"first","affiliation":[{"name":"School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7336-4492","authenticated-orcid":false,"given":"Jingda","family":"Wu","sequence":"additional","affiliation":[{"name":"School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1592-7215","authenticated-orcid":false,"given":"Zhiyu","family":"Huang","sequence":"additional","affiliation":[{"name":"School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8236-7903","authenticated-orcid":false,"given":"Zhongxu","family":"Hu","sequence":"additional","affiliation":[{"name":"School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4021-4228","authenticated-orcid":false,"given":"Jun","family":"Wang","sequence":"additional","affiliation":[{"name":"Department of Computer Science, University College London, London, U.K."}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1298-8389","authenticated-orcid":false,"given":"Alberto","family":"Sangiovanni-Vincentelli","sequence":"additional","affiliation":[{"name":"Electrical Engineering and Computer Sciences (EECS) Department, University of California at Berkeley, Berkeley, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6897-4512","authenticated-orcid":false,"given":"Chen","family":"Lv","sequence":"additional","affiliation":[{"name":"School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore"}]}],"member":"263","reference":[{"key":"ref1","first-page":"5998","article-title":"Attention is all you need","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Vaswani"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48506.2021.9561349"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2023.3261387"},{"key":"ref4","first-page":"237","article-title":"Meta-learning priors for safe Bayesian optimization","volume-title":"Proc. Conf. Robot Learn.","author":"Rothfuss"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2022.3190471"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-023-05732-2"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1038\/nature16961"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-019-1724-z"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-021-04357-7"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2021.3054625"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ITSC.2018.8569448"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ITSC45102.2020.9294262"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-023-37180-x"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-019-0025-4"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1016\/j.neuron.2016.09.025"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1038\/nn.4506"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1038\/nrn.2018.22"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.2265-08.2008"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ITSC.2018.8569938"},{"key":"ref20","first-page":"1","article-title":"CARLA: An open urban driving simulator","volume-title":"Proc. Conf. Robot Learn.","author":"Dosovitskiy"},{"key":"ref21","first-page":"726","article-title":"Safety-enhanced autonomous driving using interpretable sensor fusion transformer","volume-title":"Proc. Conf. Robot Learn.","author":"Shao"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2020.3032227"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/IROS47612.2022.9981695"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2022.3144867"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2023.3271642"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TIV.2022.3165178"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00978"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01679"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19839-7_20"},{"key":"ref30","first-page":"812","article-title":"CausalAF: Causal autoregressive flow for safety-critical driving scenario generation","volume-title":"Proc. Conf. Robot Learn.","author":"Ding"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01494"},{"key":"ref32","first-page":"1","article-title":"Efficient learning of safe driving policy via human-AI copilot optimization","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Li"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3084685"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2019.8852110"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2022.3229518"},{"key":"ref36","first-page":"22","article-title":"Constrained policy optimization","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Achiam"},{"key":"ref37","first-page":"15 338","article-title":"First order constrained optimization in policy space","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Zhang"},{"key":"ref38","first-page":"1110","article-title":"Learning to walk in the real world with minimal human effort","volume-title":"Proc. Conf. Robot Learn.","author":"Ha"},{"key":"ref39","first-page":"25 636","article-title":"Reachability constrained reinforcement learning","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Yu"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/OJCSYS.2023.3256305"},{"key":"ref41","first-page":"12 151","article-title":"Safe reinforcement learning via curriculum induction","volume":"33","author":"Turchetta","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2021.3070252"},{"key":"ref43","first-page":"908","article-title":"Safe model-based reinforcement learning with stability guarantees","volume":"30","author":"Berkenkamp","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i5.20478"},{"key":"ref45","first-page":"13 859","article-title":"Safe reinforcement learning by imagining the near future","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Thomas"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/IROS51168.2021.9635984"},{"key":"ref47","first-page":"1","article-title":"Constrained policy optimization via Bayesian world models","volume-title":"Proc. Int. Conf. Learn. Representations","author":"As"},{"key":"ref48","first-page":"24 432","article-title":"Model-based safe deep reinforcement learning via a constrained proximal policy optimization algorithm","volume":"35","author":"Jayant","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1038\/npp.2014.328"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1016\/j.janxdis.2016.03.011"},{"key":"ref51","first-page":"4759","article-title":"Deep reinforcement learning in a handful of trials using probabilistic dynamics models","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Chua"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v30i1.10295"},{"key":"ref53","first-page":"12519","article-title":"When to trust your model: Model-based policy optimization","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Janner"},{"key":"ref54","article-title":"Proximal policy optimization algorithms","author":"Schulman","year":"2017"},{"key":"ref55","first-page":"1861","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Haarnoja"},{"key":"ref56","first-page":"4572","article-title":"Generative adversarial imitation learning","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Ho"},{"key":"ref57","article-title":"Next generation simulation (NGSIM) vehicle trajectories and supporting data","year":"2016"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/34\/10345401\/10273631.pdf?arnumber=10273631","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,19]],"date-time":"2023-12-19T20:15:43Z","timestamp":1703016943000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10273631\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1]]},"references-count":57,"journal-issue":{"issue":"1"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2023.3322426","relation":{"has-preprint":[{"id-type":"doi","id":"10.36227\/techrxiv.24289108.v1","asserted-by":"object"},{"id-type":"doi","id":"10.36227\/techrxiv.24289108","asserted-by":"object"}]},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"value":"0162-8828","type":"print"},{"value":"2160-9292","type":"electronic"},{"value":"1939-3539","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,1]]}}}