{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T19:49:55Z","timestamp":1776196195973,"version":"3.50.1"},"reference-count":36,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"2","license":[{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"EPSRC iCASE Award with Thales Maritime Systems"},{"name":"EPSRC U.K. RAI Hub NCNR","award":["EPR02572X\/1"],"award-info":[{"award-number":["EPR02572X\/1"]}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100012338","name":"Alan Turing Institute","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100012338","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100000848","name":"University of Edinburgh","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100000848","id-type":"DOI","asserted-by":"publisher"}]},{"name":"EPSRC iCASE Award with Thales Maritime Systems"},{"name":"Google X AI Residency"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Robot. Autom. Lett."],"published-print":{"date-parts":[[2022,4]]},"DOI":"10.1109\/lra.2022.3150024","type":"journal-article","created":{"date-parts":[[2022,2,10]],"date-time":"2022-02-10T20:23:24Z","timestamp":1644524604000},"page":"4488-4495","source":"Crossref","is-referenced-by-count":47,"title":["Residual Learning From Demonstration: Adapting DMPs for Contact-Rich Manipulation"],"prefix":"10.1109","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0584-5163","authenticated-orcid":false,"given":"Todor","family":"Davchev","sequence":"first","affiliation":[{"name":"School of Informatics, University of Edinburgh, Edinburgh, U.K."}]},{"given":"Kevin Sebastian","family":"Luck","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering and Automation, Intelligent Robotics, Aalto University, Espoo, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7426-1498","authenticated-orcid":false,"given":"Michael","family":"Burke","sequence":"additional","affiliation":[{"name":"ECSE, Monash University, Melbourne, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1640-4619","authenticated-orcid":false,"given":"Franziska","family":"Meier","sequence":"additional","affiliation":[{"name":"Facebook AI Research, Menlo Park, CA, USA"}]},{"given":"Stefan","family":"Schaal","sequence":"additional","affiliation":[{"name":"[Google X] Intrinsic, Mountain View, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6300-5103","authenticated-orcid":false,"given":"Subramanian","family":"Ramamoorthy","sequence":"additional","affiliation":[{"name":"School of Informatics, University of Edinburgh, Edinburgh, U.K."}]}],"member":"263","reference":[{"key":"ref1","first-page":"1","article-title":"A review of robot learning for manipulation: Challenges representations, and algorithms","volume":"22","author":"Kroemer","year":"2019","journal-title":"J. Mach. Learn. Res."},{"key":"ref2","first-page":"1040","article-title":"Learning from demonstration","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"S","year":"1997"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8460986"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2020.3010633"},{"key":"ref5","article-title":"Leveraging demonstrations for deep reinforcement learning on robotics problems with sparse rewards","author":"Veerk","year":"2017"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8793506"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1162\/NECO_a_00393"},{"key":"ref8","first-page":"295","article-title":"Applying the episodic natural actor-critic architecture to motor primitive learning","volume-title":"Proc. Eur. Symp. Artif. Neural Netw.","author":"Peters","year":"2007"},{"key":"ref9","first-page":"849","article-title":"Policy search for motor primitives in robotics","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Kober","year":"2009"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2008.02.003"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9341714"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1115\/1.3149634"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2017.8202244"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1177\/0278364913482016"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CEEC.2018.8674203"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2014.6907291"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2012.2210294"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2014.2304775"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/HUMANOIDS.2015.7363457"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/HUMANOIDS.2014.7041354"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/s10514-015-9435-2"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9196765"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8794127"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2020.2988642"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2009.12.004"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/BF03256558"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.2514\/3.30149"},{"key":"ref28","article-title":"Proximal policy optimization algorithms","author":"Schulman","year":"2017"},{"key":"ref29","article-title":"TF-Agents: A library for reinforcement learning in tensorflow","author":"Guadarrama","year":"2018"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2012.6386109"},{"key":"ref31","article-title":"The SL simulation and real-time control software package","author":"S","year":"2009"},{"key":"ref32","article-title":"Robosuite: A modular simulation framework and benchmark for robot learning","author":"Zhu","year":"2020"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2016.7487156"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.1963.1105570"},{"key":"ref35","first-page":"1861","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","volume-title":"Proc. 35th Int. Conf. Mach. Learn.","volume":"80","author":"Haarnoja","year":"2018"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9197125"}],"container-title":["IEEE Robotics and Automation Letters"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7083369\/9647862\/09709544.pdf?arnumber=9709544","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,17]],"date-time":"2024-01-17T23:36:50Z","timestamp":1705534610000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9709544\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4]]},"references-count":36,"journal-issue":{"issue":"2"},"URL":"https:\/\/doi.org\/10.1109\/lra.2022.3150024","relation":{},"ISSN":["2377-3766","2377-3774"],"issn-type":[{"value":"2377-3766","type":"electronic"},{"value":"2377-3774","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4]]}}}