{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T16:58:02Z","timestamp":1775667482725,"version":"3.50.1"},"reference-count":39,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,5,23]],"date-time":"2022-05-23T00:00:00Z","timestamp":1653264000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,5,23]],"date-time":"2022-05-23T00:00:00Z","timestamp":1653264000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100004410","name":"The Scientific and Technological Research Council of Turkey (TUBITAK)","doi-asserted-by":"publisher","award":["118E857"],"award-info":[{"award-number":["118E857"]}],"id":[{"id":"10.13039\/501100004410","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,5,23]]},"DOI":"10.1109\/icra46639.2022.9811537","type":"proceedings-article","created":{"date-parts":[[2022,7,12]],"date-time":"2022-07-12T19:36:40Z","timestamp":1657654600000},"page":"7299-7305","source":"Crossref","is-referenced-by-count":16,"title":["APF-RL: Safe Mapless Navigation in Unknown Environments"],"prefix":"10.1109","author":[{"given":"Kemal","family":"Bektas","sequence":"first","affiliation":[{"name":"Electrical &amp; Electronics Engineering, Bo&#x011F;azi&#x00E7;i University,Intelligent Systems Laboratory,Istanbul,Turkey"}]},{"given":"H. Isil","family":"Bozma","sequence":"additional","affiliation":[{"name":"Electrical &amp; Electronics Engineering, Bo&#x011F;azi&#x00E7;i University,Intelligent Systems Laboratory,Istanbul,Turkey"}]}],"member":"263","reference":[{"key":"ref39","year":"0"},{"key":"ref38","article-title":"Continuous control with deep reinforcement learning","volume":"abs 1509 2971","author":"lillicrap","year":"0","journal-title":"CoRR"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989182"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1002\/rob.21918"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8460861"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1177\/0278364918796267"},{"key":"ref37","article-title":"Soft actor-critic algorithms and applications","volume":"abs 1812 5905","author":"haarnoja","year":"0","journal-title":"ArXiv"},{"key":"ref36","article-title":"Asynchronous methods for deep reinforcement learning","author":"mnih","year":"0","journal-title":"ICML"},{"key":"ref35","article-title":"Proximal policy optimization algorithms","author":"schulman","year":"0","journal-title":"ArXiv"},{"key":"ref34","article-title":"Trust region policy optimization","author":"schulman","year":"0","journal-title":"ICML"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989381"},{"key":"ref11","article-title":"Agile off-road autonomous driving using end-to-end deep imitation learning","volume":"abs 1709 7174","author":"pan","year":"0","journal-title":"CoRR"},{"key":"ref12","first-page":"185","article-title":"Intentionnet: Integrating planning and deep learning for goal-directed autonomous navigation","volume":"78","author":"gao","year":"0","journal-title":"1st Annual Conference on Robot Learning CoRL 2017 ser Proceedings of Machine Learning Research"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8461096"},{"key":"ref14","article-title":"Aggressive deep driving: Model predictive control with a CNN cost model","volume":"abs 1707 5303","author":"drews","year":"0","journal-title":"CoRR"},{"key":"ref15","author":"savinov","year":"0","journal-title":"Semi-parametric topological memory for navigation"},{"key":"ref16","first-page":"1","article-title":"Driving policy transfer via modularity and abstraction","volume":"87","author":"m\u00fcller","year":"0","journal-title":"2nd Annual Conference on Robot Learning CoRL 2018"},{"key":"ref17","article-title":"Deep imitative models for flexible inference, planning, and control","volume":"abs 1810 6544","author":"rhinehart","year":"0","journal-title":"CoRR"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8794223"},{"key":"ref19","article-title":"Combining optimal control and learning for visual navigation in novel environments","volume":"abs 1903 2531","author":"bansal","year":"0","journal-title":"ArXiv"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2011.6161176"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/70.163777"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.2007.363615"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4471-2063-6_225"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-60916-4_19"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/s10514-011-9217-4"},{"key":"ref5","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","volume":"abs 1801 1290","author":"haarnoja","year":"0","journal-title":"ArXiv"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2017.8202134"},{"key":"ref7","author":"duan","year":"0","journal-title":"RL2 Fast reinforcement learning via slow reinforcement learning"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/icABCD49160.2020.9183838"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989182"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.dt.2019.04.011"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2019.2931199"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9197148"},{"key":"ref21","author":"rana","year":"0","journal-title":"Residual reactive navigation Combining classical and learned navigation strategies for deployment in unknown environments"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48506.2021.9560951"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9341783"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2014.2337571"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/100.580977"}],"event":{"name":"2022 IEEE International Conference on Robotics and Automation (ICRA)","location":"Philadelphia, PA, USA","start":{"date-parts":[[2022,5,23]]},"end":{"date-parts":[[2022,5,27]]}},"container-title":["2022 International Conference on Robotics and Automation (ICRA)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9811522\/9811357\/09811537.pdf?arnumber=9811537","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T23:08:00Z","timestamp":1667516880000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9811537\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,23]]},"references-count":39,"URL":"https:\/\/doi.org\/10.1109\/icra46639.2022.9811537","relation":{},"subject":[],"published":{"date-parts":[[2022,5,23]]}}}