{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T15:28:46Z","timestamp":1774538926886,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":32,"publisher":"ACM","license":[{"start":{"date-parts":[[2017,7,28]],"date-time":"2017-07-28T00:00:00Z","timestamp":1501200000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2017,7,28]]},"DOI":"10.1145\/3099564.3099567","type":"proceedings-article","created":{"date-parts":[[2017,7,26]],"date-time":"2017-07-26T20:05:41Z","timestamp":1501099541000},"page":"1-13","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":121,"title":["Learning locomotion skills using DeepRL"],"prefix":"10.1145","author":[{"given":"Xue Bin","family":"Peng","sequence":"first","affiliation":[{"name":"University of British Columbia, Vancouver, Canada"}]},{"given":"Michiel","family":"van de Panne","sequence":"additional","affiliation":[{"name":"University of British Columbia, Vancouver, Canada"}]}],"member":"320","published-online":{"date-parts":[[2017,7,28]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.2312\/PE\/vriphys\/vriphysl2\/137-145"},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.12852"},{"key":"e_1_3_2_2_3_1","article-title":"Intelligence by mechanics","volume":"365","author":"Blickhan Reinhard","year":"2007","journal-title":"Philosophical Transactions of the Royal Society of London A: Mathematical, Physical and Engineering Sciences"},{"key":"e_1_3_2_2_4_1","article-title":"Robust Task-based Control Policies for Physics-based Characters","volume":"28","author":"Coros Stelian","year":"2009","journal-title":"ACM Trans. Graph. (Proc. SIGGRAPH Asia)"},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"crossref","unstructured":"Stelian Coros Philippe Beaudoin and Michiel van de Panne. 2010. Generalized Biped Walking Control. ACM Transctions on Graphics 29 4 (2010) Article 130.  Stelian Coros Philippe Beaudoin and Michiel van de Panne. 2010. Generalized Biped Walking Control. ACM Transctions on Graphics 29 4 (2010) Article 130.","DOI":"10.1145\/1778765.1781156"},{"key":"e_1_3_2_2_6_1","article-title":"Synthesis of Constrained Walking Skills","volume":"27","author":"Coros Stelian","year":"2008","journal-title":"ACM Trans. Graph. (Proc. Siggraph Asia)"},{"key":"e_1_3_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/2010324.1964954"},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/1833349.1781157"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/2508363.2508399"},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1098\/rspb.2003.2454"},{"key":"e_1_3_2_2_11_1","unstructured":"Shixiang Gu Ethan Holly Timothy Lillicrap and Sergey Levine. 2016. Deep Reinforcement Learning for Robotic Manipulation. arXiv preprint arXiv:1610.00633 (2016).  Shixiang Gu Ethan Holly Timothy Lillicrap and Sergey Levine. 2016. Deep Reinforcement Learning for Robotic Manipulation. arXiv preprint arXiv:1610.00633 (2016)."},{"key":"e_1_3_2_2_12_1","unstructured":"Matthew J. Hausknecht and Peter Stone. 2015. Deep Reinforcement Learning in Parameterized Action Space. CoRR abs\/1511.04143 (2015).  Matthew J. Hausknecht and Peter Stone. 2015. Deep Reinforcement Learning in Parameterized Action Space. CoRR abs\/1511.04143 (2015)."},{"key":"e_1_3_2_2_13_1","volume-title":"SIAM Journal on Control and Optimization","author":"Konda Vijay"},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661229.2661233"},{"key":"e_1_3_2_2_15_1","unstructured":"Sergey Levine Chelsea Finn Trevor Darrell and Pieter Abbeel. 2015. End-to-End Training of Deep Visuomotor Policies. CoRR abs\/1504.00702 (2015).  Sergey Levine Chelsea Finn Trevor Darrell and Pieter Abbeel. 2015. End-to-End Training of Deep Visuomotor Policies. CoRR abs\/1504.00702 (2015)."},{"key":"e_1_3_2_2_16_1","unstructured":"Timothy P. Lillicrap Jonathan J. Hunt Alexander Pritzel Nicolas Heess Tom Erez Yuval Tassa David Silver and Daan Wierstra. 2015. Continuous control with deep reinforcement learning. CoRR abs\/1509.02971 (2015).  Timothy P. Lillicrap Jonathan J. Hunt Alexander Pritzel Nicolas Heess Tom Erez Yuval Tassa David Silver and Daan Wierstra. 2015. Continuous control with deep reinforcement learning. CoRR abs\/1509.02971 (2015)."},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/2893476"},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/2366145.2366173"},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/IEMBS.1995.579743"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/1778765.1778808"},{"key":"e_1_3_2_2_22_1","unstructured":"Igor Mordatch Kendall Lowrey Galen Andrew Zoran Popovic and Emanuel Todorov. 2015. Interactive Control of Diverse Complex Characters with Neural Networks. In Advances in Neural Information Processing Systems 28. 3132--3140.  Igor Mordatch Kendall Lowrey Galen Andrew Zoran Popovic and Emanuel Todorov. 2015. Interactive Control of Diverse Complex Characters with Neural Networks. In Advances in Neural Information Processing Systems 28 . 3132--3140."},{"key":"e_1_3_2_2_23_1","volume-title":"Terrain-Adaptive Locomotion Skills Using Deep Reinforcement Learning. ACM Transactions on Graphics (Proc. SIGGRAPH 2016)","author":"Peng Xue Bin","year":"2016"},{"key":"e_1_3_2_2_24_1","unstructured":"John Schulman Philipp Moritz Sergey Levine Michael I. Jordan and Pieter Abbeel. 2015. High-Dimensional Continuous Control Using Generalized Advantage Estimation. CoRR abs\/1506.02438 (2015).  John Schulman Philipp Moritz Sergey Levine Michael I. Jordan and Pieter Abbeel. 2015. High-Dimensional Continuous Control Using Generalized Advantage Estimation. CoRR abs\/1506.02438 (2015)."},{"key":"e_1_3_2_2_25_1","volume-title":"Proc. International Conference on Machine Learning. 387--395","author":"Silver David","year":"2014"},{"key":"e_1_3_2_2_26_1","unstructured":"R. Sutton D. Mcallester S. Singh and Y Mansour. 2001. Policy Gradient Methods for Reinforcement Learning with Function Approximation. (2001).  R. Sutton D. Mcallester S. Singh and Y Mansour. 2001. Policy Gradient Methods for Reinforcement Learning with Function Approximation. (2001)."},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/2601097.2601121"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-27645-3_7"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1007\/BF00198959"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"crossref","unstructured":"Jack M. Wang Samuel R. Hamner Scott L. Delp Vladlen Koltun and More Specifically. 2012. Optimizing locomotion controllers using biologically-based actuators and objectives. ACM Trans. Graph (2012).  Jack M. Wang Samuel R. Hamner Scott L. Delp Vladlen Koltun and More Specifically. 2012. Optimizing locomotion controllers using biologically-based actuators and objectives. ACM Trans. Graph (2012).","DOI":"10.1145\/2185520.2185521"},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2012.11.007"},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/1276377.1276509"}],"event":{"name":"SCA '17: The ACM SIGGRAPH \/ Eurographics Symposium on Computer Animation","location":"Los Angeles California","acronym":"SCA '17","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques","EUROGRAPHICS The European Association for Computer Graphics"]},"container-title":["Proceedings of the ACM SIGGRAPH \/ Eurographics Symposium on Computer Animation"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3099564.3099567","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3099564.3099567","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T03:37:17Z","timestamp":1750217837000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3099564.3099567"}},"subtitle":["does the choice of action space matter?"],"short-title":[],"issued":{"date-parts":[[2017,7,28]]},"references-count":32,"alternative-id":["10.1145\/3099564.3099567","10.1145\/3099564"],"URL":"https:\/\/doi.org\/10.1145\/3099564.3099567","relation":{},"subject":[],"published":{"date-parts":[[2017,7,28]]},"assertion":[{"value":"2017-07-28","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}