{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,7]],"date-time":"2026-02-07T18:42:49Z","timestamp":1770489769554,"version":"3.49.0"},"reference-count":37,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,5,1]],"date-time":"2020-05-01T00:00:00Z","timestamp":1588291200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,5,1]],"date-time":"2020-05-01T00:00:00Z","timestamp":1588291200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,5,1]],"date-time":"2020-05-01T00:00:00Z","timestamp":1588291200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,5]]},"DOI":"10.1109\/icra40945.2020.9197108","type":"proceedings-article","created":{"date-parts":[[2020,9,15]],"date-time":"2020-09-15T21:25:46Z","timestamp":1600205146000},"page":"6498-6505","source":"Crossref","is-referenced-by-count":15,"title":["Adaptive Curriculum Generation from Demonstrations for Sim-to-Real Visuomotor Control"],"prefix":"10.1109","author":[{"given":"Lukas","family":"Hermann","sequence":"first","affiliation":[]},{"given":"Max","family":"Argus","sequence":"additional","affiliation":[]},{"given":"Andreas","family":"Eitel","sequence":"additional","affiliation":[]},{"given":"Artemij","family":"Amiranashvili","sequence":"additional","affiliation":[]},{"given":"Wolfram","family":"Burgard","sequence":"additional","affiliation":[]},{"given":"Thomas","family":"Brox","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref33","article-title":"Proximal policy optimization algorithms","author":"schulman","year":"2017","journal-title":"CoRR"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.2174\/1573399812666160613113556"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2015.7139389"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2011.VII.008"},{"key":"ref37","article-title":"Bullet physics simulation","author":"coumans","year":"0"},{"key":"ref36","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2015","journal-title":"ICLRE"},{"key":"ref35","article-title":"Qt-opt: Scalable deep reinforcement learning for vision-based robotic manipulation","author":"kalashnikov","year":"2018"},{"key":"ref34","article-title":"Dataefficient deep reinforcement learning for dexterous manipulation","author":"popov","year":"2017","journal-title":"CoRR"},{"key":"ref10","first-page":"767","article-title":"Surreal: Open-source reinforcement learning framework and robot manipulation benchmark","author":"fan","year":"2018","journal-title":"Conference on Robot Learning (CoRL)"},{"key":"ref11","first-page":"1515","article-title":"Automatic goal generation for reinforcement learning agents","volume":"80","author":"florensa","year":"0"},{"key":"ref12","article-title":"Learning by playing - solving sparse reward tasks from scratch","author":"riedmiller","year":"2018","journal-title":"CoRR"},{"key":"ref13","article-title":"Combining learned skills and reinforcement learning for robotic manipulations","author":"strudel","year":"2019"},{"key":"ref14","article-title":"Leveraging demonstrations for deep reinforcement learning on robotics problems with sparse rewards","author":"ve?er\u00edk","year":"2017"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8462891"},{"key":"ref16","first-page":"2930","article-title":"Playing hard exploration games by watching youtube","author":"aytar","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8460968"},{"key":"ref18","article-title":"Adversarial skill networks: Unsupervised robot skill learning from video","author":"mees","year":"2019"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8463162"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8460875"},{"key":"ref4","article-title":"Learning to singulate objects using a push proposal network","author":"eitel","year":"2017","journal-title":"Proc Of the International Symposium on Robotics Research (ISRR)"},{"key":"ref27","first-page":"262","article-title":"Sim-to-real robot learning from pixels with progressive nets","volume":"78","author":"rusu","year":"0","journal-title":"Conference on Robot Learning (CoRL)"},{"key":"ref3","first-page":"4238","article-title":"Learning synergies between pushing and grasping with selfsupervised deep reinforcement learning","author":"zeng","year":"2018","journal-title":"Proc of the IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8461249"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2018.XIV.002"},{"key":"ref5","first-page":"278","article-title":"Policy invariance under reward transformations: Theory and application to reward shaping","author":"ng","year":"1999","journal-title":"Proc of the International Conference on Machine Learning (ICML)"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/1553374.1553380"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2018.XIV.049"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8461044"},{"key":"ref9","first-page":"482","article-title":"Reverse curriculum generation for reinforcement learning","author":"florensa","year":"2017","journal-title":"Conference on Robot Learning (CoRL)"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8460995"},{"key":"ref20","first-page":"5048","article-title":"Hindsight experience replay","author":"andrychowicz","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2017.8202133"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2018.XIV.009"},{"key":"ref24","first-page":"1","article-title":"Simto-real transfer of robotic control with dynamics randomization","author":"peng","year":"2018","journal-title":"Proc of the IEEE International Conference on Robotics & Automation (ICRA)"},{"key":"ref23","first-page":"334","article-title":"Transferring end-to-end visuomotor control from simulation to real world for a multistage task","volume":"78","author":"james","year":"0","journal-title":"Conference on Robot Learning (CoRL)"},{"key":"ref26","article-title":"Solving rubik&#x2019;s cube with a robot hand","author":"akkaya","year":"2019"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8793789"}],"event":{"name":"2020 IEEE International Conference on Robotics and Automation (ICRA)","location":"Paris, France","start":{"date-parts":[[2020,5,31]]},"end":{"date-parts":[[2020,8,31]]}},"container-title":["2020 IEEE International Conference on Robotics and Automation (ICRA)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9187508\/9196508\/09197108.pdf?arnumber=9197108","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,28]],"date-time":"2022-06-28T00:23:27Z","timestamp":1656375807000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9197108\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,5]]},"references-count":37,"URL":"https:\/\/doi.org\/10.1109\/icra40945.2020.9197108","relation":{},"subject":[],"published":{"date-parts":[[2020,5]]}}}