{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,3]],"date-time":"2025-12-03T18:05:20Z","timestamp":1764785120894,"version":"3.37.3"},"reference-count":27,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"name":"Helmsley Charitable Trust through the Agricultural, Biological and Cognitive Robotics Initiative"},{"name":"Marcus Endowment Fund, both at Ben Gurion University of the Negev"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2023]]},"DOI":"10.1109\/access.2023.3274675","type":"journal-article","created":{"date-parts":[[2023,5,10]],"date-time":"2023-05-10T23:51:57Z","timestamp":1683762717000},"page":"47681-47689","source":"Crossref","is-referenced-by-count":6,"title":["Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function"],"prefix":"10.1109","volume":"11","author":[{"given":"Or","family":"Simhon","sequence":"first","affiliation":[{"name":"Department of Mechanical Engineering, Ben-Gurion University of the Negev, Be&#x2019;er Sheva, Israel"}]},{"given":"Zohar","family":"Karni","sequence":"additional","affiliation":[{"name":"Department of Industrial Engineering and Management, Ben-Gurion University of the Negev, Be&#x2019;er Sheva, Israel"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7717-7259","authenticated-orcid":false,"given":"Sigal","family":"Berman","sequence":"additional","affiliation":[{"name":"Department of Industrial Engineering and Management, Ben-Gurion University of the Negev, Be&#x2019;er Sheva, Israel"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3106-1861","authenticated-orcid":false,"given":"David","family":"Zarrouk","sequence":"additional","affiliation":[{"name":"Department of Mechanical Engineering, Ben-Gurion University of the Negev, Be&#x2019;er Sheva, Israel"}]}],"member":"263","reference":[{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8461203"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2022.3161699"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/RoMoCo.2019.8787360"},{"key":"ref14","first-page":"3003","article-title":"Quadruped robot obstacle negotiation via reinforcement learning","author":"lee","year":"2006","journal-title":"Proc IEEE Int Conf Robot Autom (ICRA)"},{"key":"ref11","article-title":"Towards monocular vision based obstacle avoidance through deep reinforcement learning","author":"xie","year":"2017","journal-title":"arXiv 1706 09829 [cs]"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2022.3178791"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/BIOROB.2010.5626034"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2003.1250609"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3040896"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.23919\/SICEISCS.2018.8330161"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2013.6630551"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2018.2805165"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1007\/s10514-015-9459-7"},{"key":"ref23","article-title":"Unity: A general platform for intelligent agents","author":"juliani","year":"2018","journal-title":"arXiv 1809 02627"},{"article-title":"Implementation matters in deep RL: A case study on PPO and TRPO","year":"0","author":"engstrom","key":"ref26"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2018.07.006"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8794260"},{"key":"ref22","article-title":"Proximal policy optimization algorithms","author":"schulman","year":"2017","journal-title":"arXiv 1707 06347"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9340986"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11694"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.2002.1014237"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1177\/0278364907087548"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2021.3068639"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2012.6225261"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2009.5354561"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1177\/0278364913495721"},{"key":"ref5","volume":"1","author":"sutton","year":"1998","journal-title":"Reinforcement Learning An Introduction"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/10005208\/10121761.pdf?arnumber=10121761","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T18:29:14Z","timestamp":1686594554000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10121761\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"references-count":27,"URL":"https:\/\/doi.org\/10.1109\/access.2023.3274675","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2023]]}}}