{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,4]],"date-time":"2025-11-04T11:09:28Z","timestamp":1762254568887,"version":"3.28.0"},"reference-count":19,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,5,29]],"date-time":"2023-05-29T00:00:00Z","timestamp":1685318400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,5,29]],"date-time":"2023-05-29T00:00:00Z","timestamp":1685318400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,5,29]]},"DOI":"10.1109\/icra48891.2023.10161129","type":"proceedings-article","created":{"date-parts":[[2023,7,4]],"date-time":"2023-07-04T17:20:56Z","timestamp":1688491256000},"page":"9464-9470","source":"Crossref","is-referenced-by-count":6,"title":["SACPlanner: Real-World Collision Avoidance with a Soft Actor Critic Local Planner and Polar State Representations"],"prefix":"10.1109","author":[{"given":"Khaled","family":"Nakhleh","sequence":"first","affiliation":[{"name":"Nokia Bell Labs,Murray Hill,NJ"}]},{"given":"Minahil","family":"Raza","sequence":"additional","affiliation":[{"name":"Nokia Bell Labs,Murray Hill,NJ"}]},{"given":"Mack","family":"Tang","sequence":"additional","affiliation":[{"name":"Nokia Bell Labs,Murray Hill,NJ"}]},{"given":"Matthew","family":"Andrews","sequence":"additional","affiliation":[{"name":"Nokia Bell Labs,Murray Hill,NJ"}]},{"given":"Rinu","family":"Boney","sequence":"additional","affiliation":[{"name":"Nokia Bell Labs,Murray Hill,NJ"}]},{"given":"Ilija","family":"Had\u017ei\u0107","sequence":"additional","affiliation":[{"name":"Nokia Bell Labs,Murray Hill,NJ"}]},{"given":"Jeongran","family":"Lee","sequence":"additional","affiliation":[{"name":"Nokia Bell Labs,Murray Hill,NJ"}]},{"given":"Atefeh","family":"Mohajeri","sequence":"additional","affiliation":[{"name":"Nokia Bell Labs,Murray Hill,NJ"}]},{"given":"Karina","family":"Palyutina","sequence":"additional","affiliation":[{"name":"Nokia Bell Labs,Murray Hill,NJ"}]}],"member":"263","reference":[{"key":"ref13","first-page":"1861","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","author":"haarnoja","year":"0","journal-title":"International Conference on Machine Learning"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8794134"},{"key":"ref15","first-page":"1587","article-title":"Addressing function approximation error in actor-critic methods","author":"fujimoto","year":"0","journal-title":"International Conference on Machine Learning"},{"key":"ref14","article-title":"Soft actor-critic algorithms and applications","author":"haarnoja","year":"2018","journal-title":"ArXiv Preprint"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/IROS51168.2021.9636039"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2014.6942636"},{"key":"ref2","volume":"abs 1312 5602","author":"mnih","year":"2013","journal-title":"Playing atari with deep reinforcement learning"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/100.580977"},{"key":"ref17","article-title":"Curl: Contrastive unsupervised representations for reinforcement learning","author":"srinivas","year":"0","journal-title":"ArXiv Preprint"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17276"},{"journal-title":"ClearPath Robotics Product Datasheet","year":"0","key":"ref19"},{"key":"ref18","first-page":"1","article-title":"Stable-baselines3: Reliable reinforcement learning implementations","volume":"22","author":"raffin","year":"2021","journal-title":"Journal of Machine Learning Research"},{"key":"ref8","first-page":"19884","article-title":"Reinforcement learning with augmented data","volume":"33","author":"laskin","year":"2020","journal-title":"Advances in neural information processing systems"},{"journal-title":"Applying deep reinforcement learning in the navigation of mobile robots in static and dynamic environments","year":"2019","author":"g\u00fcldenring","key":"ref7"},{"key":"ref9","article-title":"Image augmentation is all you need: Regularizing deep reinforcement learning from pixels","author":"kostrikov","year":"2020","journal-title":"ArXiv Preprint"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48506.2021.9561462"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9341783"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9341540"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CASE48305.2020.9216798"}],"event":{"name":"2023 IEEE International Conference on Robotics and Automation (ICRA)","start":{"date-parts":[[2023,5,29]]},"location":"London, United Kingdom","end":{"date-parts":[[2023,6,2]]}},"container-title":["2023 IEEE International Conference on Robotics and Automation (ICRA)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10160211\/10160212\/10161129.pdf?arnumber=10161129","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,10]],"date-time":"2023-10-10T17:57:04Z","timestamp":1696960624000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10161129\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,5,29]]},"references-count":19,"URL":"https:\/\/doi.org\/10.1109\/icra48891.2023.10161129","relation":{},"subject":[],"published":{"date-parts":[[2023,5,29]]}}}