{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:18:06Z","timestamp":1750220286951,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":26,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,3,11]],"date-time":"2022-03-11T00:00:00Z","timestamp":1646956800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,3,11]]},"DOI":"10.1145\/3529399.3529432","type":"proceedings-article","created":{"date-parts":[[2022,6,10]],"date-time":"2022-06-10T15:43:09Z","timestamp":1654875789000},"page":"209-215","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Automatically Learning Fallback Strategies with Model-Free Reinforcement Learning in Safety-Critical Driving Scenarios"],"prefix":"10.1145","author":[{"given":"Ugo","family":"Lecerf","sequence":"first","affiliation":[{"name":"Renault Software Labs, France"}]},{"given":"Christelle","family":"Yemdji-Tchassi","sequence":"additional","affiliation":[{"name":"Renault Software Labs, France"}]},{"given":"Sebastien","family":"Aubert","sequence":"additional","affiliation":[{"name":"Renault Software Labs, France"}]},{"given":"Pietro","family":"Michiardi","sequence":"additional","affiliation":[{"name":"Data Science, EURECOM, France"}]}],"member":"320","published-online":{"date-parts":[[2022,6,10]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2983149"},{"key":"e_1_3_2_1_2_1","first-page":"1193","volume-title":"Proceedings of the 35th International Conference on Machine Learning","volume":"80","author":"Depeweg S.","unstructured":"S. Depeweg , J.-M. Hernandez-Lobato , F. Doshi-Velez , and S. Udluft . 2018. Decomposition of uncertainty in Bayesian deep learning for efficient and risk-sensitive learning . In Proceedings of the 35th International Conference on Machine Learning , vol. 80 , pp. 1184\u2013 1193 . S. Depeweg, J.-M. Hernandez-Lobato, F. Doshi-Velez, and S. Udluft. 2018. Decomposition of uncertainty in Bayesian deep learning for efficient and risk-sensitive learning. In Proceedings of the 35th International Conference on Machine Learning, vol. 80, pp. 1184\u20131193."},{"key":"e_1_3_2_1_3_1","first-page":"1","volume-title":"PP","author":"Zang S.","unstructured":"S. Zang , M. Ding , D. Smith , P. Tyler , T. Rakotoarivelo , and M. A. Kaafar . 2019. The impact of adverse weather conditions on autonomous vehicles: Examining how rain, snow, fog, and hail affect the performance of a self-driving car. IEEE Vehicular Technology Magazine, vol . PP , pp. 1\u2013 1 , 03 . S. Zang, M. Ding, D. Smith, P. Tyler, T. Rakotoarivelo, and M. A. Kaafar. 2019. The impact of adverse weather conditions on autonomous vehicles: Examining how rain, snow, fog, and hail affect the performance of a self-driving car. IEEE Vehicular Technology Magazine, vol. PP, pp. 1\u20131, 03."},{"key":"e_1_3_2_1_4_1","first-page":"1476","volume-title":"2019 IEEE Intelligent Vehicles Symposium (IV)","author":"Bouton M.","unstructured":"M. Bouton , A. Nakhaei , K. Fujimura , and M. J. Kochenderfer . 2019. Safe reinforcement learning with scene decomposition for navigating complex urban environments . In 2019 IEEE Intelligent Vehicles Symposium (IV) , pp. 1469\u2013 1476 . M. Bouton, A. Nakhaei, K. Fujimura, and M. J. Kochenderfer. 2019. Safe reinforcement learning with scene decomposition for navigating complex urban environments. In 2019 IEEE Intelligent Vehicles Symposium (IV), pp. 1469\u20131476."},{"key":"e_1_3_2_1_5_1","unstructured":"W. R. Clements B.-M. Robaglia B. V. Delft R. B. Slaoui and S. Toth. 2019. Estimating risk and uncertainty in deep reinforcement learning. ArXiv preprint arXiv:190509638.  W. R. Clements B.-M. Robaglia B. V. Delft R. B. Slaoui and S. Toth. 2019. Estimating risk and uncertainty in deep reinforcement learning. ArXiv preprint arXiv:190509638."},{"key":"e_1_3_2_1_6_1","first-page":"1569","volume-title":"2020 IEEE Intelligent Vehicles Symposium (IV)","author":"Hoel C.-J.","unstructured":"C.-J. Hoel , K. Wolff , and L. Laine . Tactical decision-making in autonomous driving by reinforcement learning with uncertainty estimation . In 2020 IEEE Intelligent Vehicles Symposium (IV) pp. 1563\u2013 1569 . C.-J. Hoel, K. Wolff, and L. Laine. Tactical decision-making in autonomous driving by reinforcement learning with uncertainty estimation. In 2020 IEEE Intelligent Vehicles Symposium (IV) pp. 1563\u20131569."},{"key":"e_1_3_2_1_7_1","first-page":"517","volume-title":"Proceedings of the 37th International Conference on Machine Learning","volume":"119","author":"Badia A. P.","unstructured":"A. P. Badia , B. Piot , S. Kapturowski , P. Sprechmann , A. Vitvitskyi , Z. D. Guo , and C. Blundell . 2020. Agent57: Outperforming the Atari human benchmark . In Proceedings of the 37th International Conference on Machine Learning , vol. 119 , pp. 507\u2013 517 . A. P. Badia, B. Piot, S. Kapturowski, P. Sprechmann, A. Vitvitskyi, Z. D. Guo, and C. Blundell. 2020. Agent57: Outperforming the Atari human benchmark. In Proceedings of the 37th International Conference on Machine Learning, vol. 119, pp. 507\u2013517."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"key":"e_1_3_2_1_9_1","unstructured":"C. Berner G. Brockman B. Chan V. Cheung P. Debiak C. Dennison D. Farhi Q. Fischer S. Hashme C. Hesse 2019. Dota 2 with large scale deep reinforcement learning. [Online]. Available: http:\/\/arxiv.org\/abs\/1912.06680  C. Berner G. Brockman B. Chan V. Cheung P. Debiak C. Dennison D. Farhi Q. Fischer S. Hashme C. Hesse 2019. Dota 2 with large scale deep reinforcement learning. [Online]. Available: http:\/\/arxiv.org\/abs\/1912.06680"},{"key":"e_1_3_2_1_10_1","unstructured":"M. Hessel J. Modayil H. van Hasselt T. Schaul G. Ostrovski W. Dabney D. Horgan B. Piot M. G. Azar and D. Silver. 2017. Rainbow: combining improvements in deep reinforcement learning. CoRR vol. abs\/1710.02298. [Online]. Available: http:\/\/arxiv.org\/abs\/1710.02298  M. Hessel J. Modayil H. van Hasselt T. Schaul G. Ostrovski W. Dabney D. Horgan B. Piot M. G. Azar and D. Silver. 2017. Rainbow: combining improvements in deep reinforcement learning. CoRR vol. abs\/1710.02298. [Online]. Available: http:\/\/arxiv.org\/abs\/1710.02298"},{"key":"e_1_3_2_1_11_1","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Schaul T.","year":"2016","unstructured":"T. Schaul , J. Quan , I. Antonoglou , and D. Silver , \u201c Prioritized experience replay ,\u201d in Proceedings of the International Conference on Learning Representations , 2016 . T. Schaul, J. Quan, I. Antonoglou, and D. Silver, \u201cPrioritized experience replay,\u201d in Proceedings of the International Conference on Learning Representations, 2016."},{"key":"e_1_3_2_1_12_1","volume-title":"Reinforcement Learning: An Introduction","author":"Sutton R. S.","year":"2018","unstructured":"R. S. Sutton and A. G. Barto . 2018 . Reinforcement Learning: An Introduction , 2 nd ed. The MIT Press . [Online]. Available: http:\/\/incompleteideas.net\/book\/the-book-2nd.html R. S. Sutton and A. G. Barto. 2018. Reinforcement Learning: An Introduction, 2nd ed. The MIT Press. [Online]. Available: http:\/\/incompleteideas.net\/book\/the-book-2nd.html","edition":"2"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"M. J. Kochenderfer C. Amato G. Chowdhary J. P. How H. J. D. Reynolds J. R. Thornton P. A. Torres-Carrasquillo N. K. \u00a8Ure and J. Vian. 2015. Decision Making Under Uncertainty: Theory and Application 1st ed. The MIT Press.  M. J. Kochenderfer C. Amato G. Chowdhary J. P. How H. J. D. Reynolds J. R. Thornton P. A. Torres-Carrasquillo N. K. \u00a8Ure and J. Vian. 2015. Decision Making Under Uncertainty: Theory and Application 1st ed. The MIT Press.","DOI":"10.7551\/mitpress\/10187.001.0001"},{"key":"e_1_3_2_1_14_1","unstructured":"C. Hoel K. R. Driggs-Campbell K. Wolff L. Laine and M. J. Kochenderfer. 2019. Combining planning and deep reinforcement learning in tactical decision making for autonomous driving. CoRR vol. abs\/1905.02680. [Online]. Available: http:\/\/arxiv.org\/abs\/1905.02680  C. Hoel K. R. Driggs-Campbell K. Wolff L. Laine and M. J. Kochenderfer. 2019. Combining planning and deep reinforcement learning in tactical decision making for autonomous driving. CoRR vol. abs\/1905.02680. [Online]. Available: http:\/\/arxiv.org\/abs\/1905.02680"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/TCIAIG.2012.2186810"},{"volume-title":"Advances in Neural Information Processing Systems","author":"McAllister R.","key":"e_1_3_2_1_16_1","unstructured":"R. McAllister and C. E. Rasmussen , \u201c Data-efficient reinforcement learning in continuous state-action gaussian-pomdps. 2017 . in Advances in Neural Information Processing Systems , vol. 30 . R. McAllister and C. E. Rasmussen, \u201cData-efficient reinforcement learning in continuous state-action gaussian-pomdps. 2017. in Advances in Neural Information Processing Systems, vol. 30."},{"key":"e_1_3_2_1_17_1","unstructured":"M. Hausknecht and P. Stone. 2015. Deep recurrent q-learning for partially observable mdps. in AAAI Fall Symposia.  M. Hausknecht and P. Stone. 2015. Deep recurrent q-learning for partially observable mdps. in AAAI Fall Symposia."},{"key":"e_1_3_2_1_18_1","unstructured":"P. Zhu X. Li and P. Poupart. 2017. On improving deep reinforcement learning for pomdps. [Online]. Available: http:\/\/arxiv.org\/abs\/1704.07978  P. Zhu X. Li and P. Poupart. 2017. On improving deep reinforcement learning for pomdps. [Online]. Available: http:\/\/arxiv.org\/abs\/1704.07978"},{"volume-title":"International Conference on Learning Representations.","author":"Badia A. P.","key":"e_1_3_2_1_19_1","unstructured":"A. P. Badia , P. Sprechmann , A. Vitvitskyi , D. Guo , B. Piot , S. Kapturowski , O. Tieleman , M. Arjovsky , A. Pritzel , A. Bolt , and C. Blundell . 2020. Never give up: Learning directed exploration strategies ,\u201d in International Conference on Learning Representations. A. P. Badia, P. Sprechmann, A. Vitvitskyi, D. Guo, B. Piot, S. Kapturowski, O. Tieleman, M. Arjovsky, A. Pritzel, A. Bolt, and C. Blundell. 2020. Never give up: Learning directed exploration strategies,\u201d in International Conference on Learning Representations."},{"volume-title":"Proceedings of European Sympiosium on Artificial Neural Networks, Computational Intelligence and Machine Learning.","author":"Eriksson H.","key":"e_1_3_2_1_20_1","unstructured":"H. Eriksson and C. Dimitrakakis , \u201c Epistemic risk-sensitive reinforcement learning. 2020 . in Proceedings of European Sympiosium on Artificial Neural Networks, Computational Intelligence and Machine Learning. H. Eriksson and C. Dimitrakakis, \u201cEpistemic risk-sensitive reinforcement learning. 2020. in Proceedings of European Sympiosium on Artificial Neural Networks, Computational Intelligence and Machine Learning."},{"key":"e_1_3_2_1_21_1","volume-title":"Proceedings of the 32nd International Conference on Machine Learning (pp. 1889-1897)","author":"P.","year":"2015","unstructured":"Schulman J, Levine S, Abbeel P, Jordan M, Moritz P. 2015 . Trust region policy optimization . In Proceedings of the 32nd International Conference on Machine Learning (pp. 1889-1897) . PMLR. Schulman J, Levine S, Abbeel P, Jordan M, Moritz P. 2015. Trust region policy optimization. In Proceedings of the 32nd International Conference on Machine Learning (pp. 1889-1897). PMLR."},{"key":"e_1_3_2_1_22_1","unstructured":"L. Kirsch S. van Steenkiste and J. Schmidhuber. 2019. Improving generalization in meta reinforcement learning using learned objectives in Proceedings of the 10 th International Conference on Learning Representations.  L. Kirsch S. van Steenkiste and J. Schmidhuber. 2019. Improving generalization in meta reinforcement learning using learned objectives in Proceedings of the 10 th International Conference on Learning Representations."},{"key":"e_1_3_2_1_23_1","first-page":"1320","volume-title":"Proceedings of the 32nd International Conference on Machine Learning","volume":"37","author":"Schaul T.","unstructured":"T. Schaul , D. Horgan , K. Gregor , and D. Silver . 2015. Universal value function approximators . in Proceedings of the 32nd International Conference on Machine Learning , vol. 37 , pp. 1312\u2013 1320 . T. Schaul, D. Horgan, K. Gregor, and D. Silver. 2015. Universal value function approximators. in Proceedings of the 32nd International Conference on Machine Learning, vol. 37, pp. 1312\u20131320."},{"key":"e_1_3_2_1_24_1","first-page":"2304","volume-title":"Proceedings of the 34th International Conference on Machine Learning","volume":"70","author":"Machado M. C.","unstructured":"M. C. Machado , M. G. Bellemare , and M. Bowling . 2017. A Laplacian framework for option discovery in reinforcement learning . In Proceedings of the 34th International Conference on Machine Learning , vol. 70 , pp. 2295\u2013 2304 . M. C. Machado, M. G. Bellemare, and M. Bowling. 2017. A Laplacian framework for option discovery in reinforcement learning. In Proceedings of the 34th International Conference on Machine Learning, vol. 70, pp. 2295\u20132304."},{"key":"e_1_3_2_1_25_1","volume-title":"Advances in Neural Information Processing Systems","volume":"30","author":"Andrychowicz M.","unstructured":"M. Andrychowicz , F. Wolski , A. Ray , J. Schneider , R. Fong , P. Welinder , B. McGrew , J. Tobin , O. Pieter Abbeel , and W. Zaremba . 2017. Hindsight experience replay . In Advances in Neural Information Processing Systems , vol. 30 . M. Andrychowicz, F. Wolski, A. Ray, J. Schneider, R. Fong, P. Welinder, B. McGrew, J. Tobin, O. Pieter Abbeel, and W. Zaremba. 2017. Hindsight experience replay. In Advances in Neural Information Processing Systems, vol. 30."},{"key":"e_1_3_2_1_26_1","volume-title":"Advances in Neural Information Processing Systems","volume":"32","author":"Eysenbach B.","unstructured":"B. Eysenbach , R. R. Salakhutdinov , and S. Levine . 2019. Search on the replay buffer: Bridging planning and reinforcement learning . in Advances in Neural Information Processing Systems , vol. 32 . B. Eysenbach, R. R. Salakhutdinov, and S. Levine. 2019. Search on the replay buffer: Bridging planning and reinforcement learning. in Advances in Neural Information Processing Systems, vol. 32."}],"event":{"name":"ICMLT 2022: 2022 7th International Conference on Machine Learning Technologies","acronym":"ICMLT 2022","location":"Rome Italy"},"container-title":["2022 7th International Conference on Machine Learning Technologies (ICMLT)"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3529399.3529432","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3529399.3529432","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:31:24Z","timestamp":1750188684000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3529399.3529432"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,3,11]]},"references-count":26,"alternative-id":["10.1145\/3529399.3529432","10.1145\/3529399"],"URL":"https:\/\/doi.org\/10.1145\/3529399.3529432","relation":{},"subject":[],"published":{"date-parts":[[2022,3,11]]},"assertion":[{"value":"2022-06-10","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}