{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T22:39:40Z","timestamp":1773527980912,"version":"3.50.1"},"reference-count":36,"publisher":"IEEE","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2017,9]]},"DOI":"10.1109\/iros.2017.8206038","type":"proceedings-article","created":{"date-parts":[[2017,12,14]],"date-time":"2017-12-14T22:12:59Z","timestamp":1513289579000},"page":"2192-2199","source":"Crossref","is-referenced-by-count":14,"title":["Enhancing the performance of adaptive iterative learning control with reinforcement learning"],"prefix":"10.1109","author":[{"given":"Bojan","family":"Nemec","sequence":"first","affiliation":[]},{"given":"Mihael","family":"Simonic","sequence":"additional","affiliation":[]},{"given":"Nejc","family":"Likar","sequence":"additional","affiliation":[]},{"given":"Ales","family":"Ude","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref33","first-page":"281","article-title":"Path Integral Policy Improvement with Covariance Matrix Adaptation","author":"stulp","year":"2012","journal-title":"International Conference on Machine Learning (ICML)"},{"key":"ref32","author":"hlawatsch","year":"2010","journal-title":"Time-Frequency Analysis"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TOH.2016.2518670"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.2007.363772"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1016\/S0005-1098(99)00215-0"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/HUMANOIDS.2016.7803337"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2015.7353767"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TRA.2002.999653"},{"key":"ref11","author":"xu","year":"2009","journal-title":"Real-time Iterative Learning Control"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1177\/0278364908091463"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1017\/S0263574707003402"},{"key":"ref14","first-page":"19.1","article-title":"Impedance and Interaction Control","author":"hogan","year":"2005","journal-title":"Robotics and Automation Handbook"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1016\/S0921-8890(99)00014-7"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2014.6907291"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/3516.789685"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TCST.2004.824320"},{"key":"ref19","doi-asserted-by":"crossref","first-page":"43","DOI":"10.1109\/JRA.1987.1087068","article-title":"A unified approach for motion and force control of robot manipulators: The operational space formulation","volume":"3","author":"khatib","year":"1987","journal-title":"IEEE Journal of Robotics and Automation"},{"key":"ref28","first-page":"422","article-title":"Reinforcement learning and iterative learning control Similarity and difference","author":"ahn","year":"2009","journal-title":"IEEE International Conference on Mechatronics and Information Technology"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1023\/A:1015727715131"},{"key":"ref27","doi-asserted-by":"crossref","first-page":"171","DOI":"10.1007\/s10994-010-5223-6","article-title":"Policy search for motor primitives in robotics","volume":"84","author":"kober","year":"2010","journal-title":"Machine Learning"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/s10339-011-0404-1"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/MCS.2006.1636313"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/70.988969"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2010.2065430"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.automatica.2004.01.026"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TRA.2002.802210"},{"key":"ref2","first-page":"469","author":"peters","year":"2011","journal-title":"Towards Motor Skill Learning for Robotics"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1115\/1.4002384"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-79629-6_7"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-30301-5_8"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1007\/s10514-015-9435-2"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1016\/j.mechatronics.2005.08.002"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/41.121905"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2006.377582"},{"key":"ref26","first-page":"3137","article-title":"A generalized path integral control approach to reinforcement learning","volume":"11","author":"theodorou","year":"2010","journal-title":"The Journal of Machine Learning"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1177\/0278364913495721"}],"event":{"name":"2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)","location":"Vancouver, BC","start":{"date-parts":[[2017,9,24]]},"end":{"date-parts":[[2017,9,28]]}},"container-title":["2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8119304\/8202121\/08206038.pdf?arnumber=8206038","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,10,8]],"date-time":"2019-10-08T01:50:27Z","timestamp":1570499427000},"score":1,"resource":{"primary":{"URL":"http:\/\/ieeexplore.ieee.org\/document\/8206038\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2017,9]]},"references-count":36,"URL":"https:\/\/doi.org\/10.1109\/iros.2017.8206038","relation":{},"subject":[],"published":{"date-parts":[[2017,9]]}}}