{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T13:37:25Z","timestamp":1765546645052,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":38,"publisher":"ACM","license":[{"start":{"date-parts":[[2020,10,16]],"date-time":"2020-10-16T00:00:00Z","timestamp":1602806400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,10,16]]},"DOI":"10.1145\/3424636.3426907","type":"proceedings-article","created":{"date-parts":[[2020,11,22]],"date-time":"2020-11-22T20:43:50Z","timestamp":1606077830000},"page":"1-10","source":"Crossref","is-referenced-by-count":27,"title":["Learning to Locomote: Understanding How Environment Design Matters for Deep Reinforcement Learning"],"prefix":"10.1145","author":[{"given":"Daniele","family":"Reda","sequence":"first","affiliation":[{"name":"University of British Columbia"}]},{"given":"Tianxin","family":"Tao","sequence":"additional","affiliation":[{"name":"University of British Columbia"}]},{"given":"Michiel","family":"van de Panne","sequence":"additional","affiliation":[{"name":"University of British Columbia"}]}],"member":"320","published-online":{"date-parts":[[2020,11,22]]},"reference":[{"key":"e_1_3_2_2_1_1","unstructured":"Farzad Abdolhosseini. 2019. Learning locomotion: symmetry and torque limit considerations. https:\/\/doi.org\/10.14288\/1.0383251  Farzad Abdolhosseini. 2019. Learning locomotion: symmetry and torque limit considerations. https:\/\/doi.org\/10.14288\/1.0383251"},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359566.3360070"},{"key":"e_1_3_2_2_3_1","unstructured":"Marcin Andrychowicz Anton Raichuk Piotr Sta\u0144czyk Manu Orsini Sertan Girgin Raphael Marinier L\u00e9onard Hussenot Matthieu Geist Olivier Pietquin Marcin Michalski 2020. What Matters In On-Policy Reinforcement Learning? A Large-Scale Empirical Study. arXiv preprint arXiv:2006.05990(2020).  Marcin Andrychowicz Anton Raichuk Piotr Sta\u0144czyk Manu Orsini Sertan Girgin Raphael Marinier L\u00e9onard Hussenot Matthieu Geist Olivier Pietquin Marcin Michalski 2020. What Matters In On-Policy Reinforcement Learning? A Large-Scale Empirical Study. arXiv preprint arXiv:2006.05990(2020)."},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/1553374.1553380"},{"key":"e_1_3_2_2_5_1","unstructured":"Greg Brockman Vicki Cheung Ludwig Pettersson Jonas Schneider John Schulman Jie Tang and Wojciech Zaremba. 2016. OpenAI Gym. arXiv:arXiv:1606.01540  Greg Brockman Vicki Cheung Ludwig Pettersson Jonas Schneider John Schulman Jie Tang and Wojciech Zaremba. 2016. OpenAI Gym. arXiv:arXiv:1606.01540"},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3274247.3274506"},{"key":"e_1_3_2_2_7_1","unstructured":"Erwin Coumans and Yunfei Bai. 2016\u20132019. PyBullet a Python module for physics simulation for games robotics and machine learning. http:\/\/pybullet.org.  Erwin Coumans and Yunfei Bai. 2016\u20132019. PyBullet a Python module for physics simulation for games robotics and machine learning. http:\/\/pybullet.org."},{"volume-title":"Implementation Matters in Deep RL: A Case Study on PPO and TRPO. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=r1etN1rtPB","year":"2020","author":"Engstrom Logan","key":"e_1_3_2_2_8_1"},{"volume-title":"Reverse Curriculum Generation for Reinforcement Learning. In Conference on Robot Learning. 482\u2013495","year":"2017","author":"Florensa Carlos","key":"e_1_3_2_2_9_1"},{"key":"e_1_3_2_2_10_1","unstructured":"Scott Fujimoto Herke Van\u00a0Hoof and David Meger. 2018. Addressing function approximation error in actor-critic methods. arXiv preprint arXiv:1802.09477(2018).  Scott Fujimoto Herke Van\u00a0Hoof and David Meger. 2018. Addressing function approximation error in actor-critic methods. arXiv preprint arXiv:1802.09477(2018)."},{"volume-title":"Divide-and-Conquer Reinforcement Learning. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=rJwelMbR-","year":"2018","author":"Ghosh Dibya","key":"e_1_3_2_2_11_1"},{"volume-title":"Urban Driving with Conditional Imitation Learning. International Conference on Robotics and Automation","year":"2020","author":"Hawke Jeffrey","key":"e_1_3_2_2_12_1"},{"volume-title":"Thirty-Second AAAI Conference on Artificial Intelligence.","year":"2018","author":"Henderson Peter","key":"e_1_3_2_2_13_1"},{"key":"e_1_3_2_2_14_1","unstructured":"Matteo Hessel Hado van Hasselt Joseph Modayil and David Silver. 2019. On inductive biases in deep reinforcement learning. arXiv preprint arXiv:1907.02908(2019).  Matteo Hessel Hado van Hasselt Joseph Modayil and David Silver. 2019. On inductive biases in deep reinforcement learning. arXiv preprint arXiv:1907.02908(2019)."},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3072959.3073663"},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.5555\/3298483.3298547"},{"key":"e_1_3_2_2_17_1","unstructured":"Horia Mania Aurelia Guy and Benjamin Recht. 2018. Simple random search provides a competitive approach to reinforcement learning. arXiv preprint arXiv:1803.07055(2018).  Horia Mania Aurelia Guy and Benjamin Recht. 2018. Simple random search provides a competitive approach to reinforcement learning. arXiv preprint arXiv:1803.07055(2018)."},{"key":"e_1_3_2_2_18_1","unstructured":"Alberto\u00a0Maria Metelli Flavio Mazzolini Lorenzo Bisi Luca Sabbioni and Marcello Restelli. 2020. Control Frequency Adaptation via Action Persistence in Batch Reinforcement Learning. arxiv:2002.06836\u00a0[cs.LG]  Alberto\u00a0Maria Metelli Flavio Mazzolini Lorenzo Bisi Luca Sabbioni and Marcello Restelli. 2020. Control Frequency Adaptation via Action Persistence in Batch Reinforcement Learning. arxiv:2002.06836\u00a0[cs.LG]"},{"volume-title":"Human-level control through deep reinforcement learning. Nature 518, 7540","year":"2015","author":"Mnih Volodymyr","key":"e_1_3_2_2_19_1"},{"key":"e_1_3_2_2_20_1","unstructured":"Sanmit Narvekar Bei Peng Matteo Leonetti Jivko Sinapov Matthew\u00a0E Taylor and Peter Stone. 2020. Curriculum Learning for Reinforcement Learning Domains: A Framework and Survey. arXiv preprint arXiv:2003.04960(2020).  Sanmit Narvekar Bei Peng Matteo Leonetti Jivko Sinapov Matthew\u00a0E Taylor and Peter Stone. 2020. Curriculum Learning for Reinforcement Learning Domains: A Framework and Survey. arXiv preprint arXiv:2003.04960(2020)."},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.5555\/3306127.3331670"},{"key":"e_1_3_2_2_22_1","unstructured":"Charles Packer Katelyn Gao Jernej Kos Philipp Kr\u00e4henb\u00fchl Vladlen Koltun and Dawn Song. 2018. Assessing generalization in deep reinforcement learning. arXiv preprint arXiv:1810.12282(2018).  Charles Packer Katelyn Gao Jernej Kos Philipp Kr\u00e4henb\u00fchl Vladlen Koltun and Dawn Song. 2018. Assessing generalization in deep reinforcement learning. arXiv preprint arXiv:1810.12282(2018)."},{"volume-title":"Proc. 35th International Conference on Machine Learning (ICML","year":"2018","author":"Pardo Fabio","key":"e_1_3_2_2_23_1"},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3355089.3356501"},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3197517.3201311"},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3072959.3073602"},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3099564.3099567"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-11814-2_1"},{"key":"e_1_3_2_2_29_1","unstructured":"Aravind Rajeswaran Kendall Lowrey Emanuel\u00a0V Todorov and Sham\u00a0M Kakade. 2017. Towards generalization and simplicity in continuous control. In Advances in Neural Information Processing Systems. 6550\u20136561.  Aravind Rajeswaran Kendall Lowrey Emanuel\u00a0V Todorov and Sham\u00a0M Kakade. 2017. Towards generalization and simplicity in continuous control. In Advances in Neural Information Processing Systems. 6550\u20136561."},{"key":"e_1_3_2_2_30_1","unstructured":"John Schulman Filip Wolski Prafulla Dhariwal Alec Radford and Oleg Klimov. 2017. Proximal Policy Optimization Algorithms. arxiv:1707.06347\u00a0[cs.LG]  John Schulman Filip Wolski Prafulla Dhariwal Alec Radford and Oleg Klimov. 2017. Proximal Policy Optimization Algorithms. arxiv:1707.06347\u00a0[cs.LG]"},{"key":"e_1_3_2_2_31_1","unstructured":"David Silver Guy Lever Nicolas Heess Thomas Degris Daan Wierstra and Martin Riedmiller. 2014. Deterministic policy gradient algorithms.  David Silver Guy Lever Nicolas Heess Thomas Degris Daan Wierstra and Martin Riedmiller. 2014. Deterministic policy gradient algorithms."},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/MCG.2011.30"},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2012.6386109"},{"volume-title":"Proc. Conference on Robot Learning (CORL","year":"2019","author":"Xie Zhaoming","key":"e_1_3_2_2_34_1"},{"volume-title":"ALLSTEPS: Curriculum-driven Learning of Stepping Stone Skills. ArXiv preprint arXiv:2005.04323(2020). arxiv:2005.04323\u00a0[cs.GR]","year":"2020","author":"Xie Zhaoming","key":"e_1_3_2_2_35_1"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3197517.3201397"},{"key":"e_1_3_2_2_37_1","unstructured":"Amy Zhang Nicolas Ballas and Joelle Pineau. 2018a. A dissection of overfitting and generalization in continuous reinforcement learning. arXiv preprint arXiv:1806.07937(2018).  Amy Zhang Nicolas Ballas and Joelle Pineau. 2018a. A dissection of overfitting and generalization in continuous reinforcement learning. arXiv preprint arXiv:1806.07937(2018)."},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3197517.3201366"}],"event":{"name":"MIG '20: Motion, Interaction and Games","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"],"location":"Virtual Event SC USA","acronym":"MIG '20"},"container-title":["Motion, Interaction and Games"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3424636.3426907","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3424636.3426907","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T22:02:27Z","timestamp":1750197747000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3424636.3426907"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,10,16]]},"references-count":38,"alternative-id":["10.1145\/3424636.3426907","10.1145\/3424636"],"URL":"https:\/\/doi.org\/10.1145\/3424636.3426907","relation":{},"subject":[],"published":{"date-parts":[[2020,10,16]]}}}