{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,16]],"date-time":"2026-01-16T07:58:09Z","timestamp":1768550289458,"version":"3.49.0"},"reference-count":24,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,10,24]],"date-time":"2020-10-24T00:00:00Z","timestamp":1603497600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,10,24]],"date-time":"2020-10-24T00:00:00Z","timestamp":1603497600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,10,24]],"date-time":"2020-10-24T00:00:00Z","timestamp":1603497600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100013209","name":"Hellenic Foundation for Research and Innovation","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100013209","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,10,24]]},"DOI":"10.1109\/iros45743.2020.9341645","type":"proceedings-article","created":{"date-parts":[[2021,2,13]],"date-time":"2021-02-13T02:26:48Z","timestamp":1613183208000},"page":"3777-3782","source":"Crossref","is-referenced-by-count":8,"title":["Slope Handling for Quadruped Robots Using Deep Reinforcement Learning and Toe Trajectory Planning"],"prefix":"10.1109","author":[{"given":"Athanasios S.","family":"Mastrogeorgiou","sequence":"first","affiliation":[]},{"given":"Yehia S.","family":"Elbahrawy","sequence":"additional","affiliation":[]},{"given":"Andres","family":"Kecskemethy","sequence":"additional","affiliation":[]},{"given":"Evangelos G.","family":"Papadopoulos","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","year":"0"},{"key":"ref11","year":"0"},{"key":"ref12","article-title":"Openai gym","author":"brockman","year":"2016"},{"key":"ref13","year":"0"},{"key":"ref14","year":"0"},{"key":"ref15","year":"0"},{"key":"ref16","year":"0"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1145\/3099564.3099567"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-13-8285-7"},{"key":"ref19","doi-asserted-by":"crossref","DOI":"10.1007\/978-3-319-03194-1","author":"kober","year":"2014","journal-title":"Learning Motor Skills From Algorithms to Robot Experiments"},{"key":"ref4","first-page":"38","article-title":"Anymal-a highly mobile and dynamic quadrupedal robot","author":"hutter","year":"2016","journal-title":"Proc Int Conf Intell Robots Syst (IROS)"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/MRA.2015.2505910"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1126\/scirobotics.aau5872"},{"key":"ref5","article-title":"Learning to Walk via Deep Reinforcement Learning","author":"haarnoja","year":"2018"},{"key":"ref8","article-title":"Robust Recovery Controller for a Quadrupedal Robot using Deep Reinforcement Learning","author":"lee","year":"2019"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2018.XIV.010"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2018.8593885"},{"key":"ref1","article-title":"Fast online trajectory optimization for the bipedal robot cassie","author":"taylor","year":"2018","journal-title":"Robotics Science and Systems (RSS)"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2018.8593722"},{"key":"ref20","article-title":"Continuous control with deep reinforcement learning, CoRR","author":"lillicrap","year":"2015"},{"key":"ref22","author":"lapan","year":"2018","journal-title":"Deep Reinforcement Learning Hands-On Apply Modern RL Methods with Deep Q-Networks Value Iteration Policy Gradients TRPO AlphaGo Zero and More"},{"key":"ref21","article-title":"Deterministic Policy Gradient Algorithms","author":"silver","year":"2014","journal-title":"International Conference on Machine Learning ICML"},{"key":"ref24","article-title":"Adam: A Method for Stochastic Optimization","author":"kingma","year":"2015","journal-title":"International Conference on Learning Representations"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/MED.2016.7536064"}],"event":{"name":"2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)","location":"Las Vegas, NV, USA","start":{"date-parts":[[2020,10,24]]},"end":{"date-parts":[[2021,1,24]]}},"container-title":["2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9340668\/9340635\/09341645.pdf?arnumber=9341645","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,28]],"date-time":"2022-06-28T21:50:40Z","timestamp":1656453040000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9341645\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,10,24]]},"references-count":24,"URL":"https:\/\/doi.org\/10.1109\/iros45743.2020.9341645","relation":{},"subject":[],"published":{"date-parts":[[2020,10,24]]}}}