{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T23:47:02Z","timestamp":1774568822109,"version":"3.50.1"},"reference-count":36,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"4","license":[{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Ministry of Science and ICT","award":["NRF-2018M3C1B9088328"],"award-info":[{"award-number":["NRF-2018M3C1B9088328"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Robot. Autom. Lett."],"published-print":{"date-parts":[[2022,10]]},"DOI":"10.1109\/lra.2022.3189446","type":"journal-article","created":{"date-parts":[[2022,7,8]],"date-time":"2022-07-08T19:24:54Z","timestamp":1657308294000},"page":"9004-9011","source":"Crossref","is-referenced-by-count":14,"title":["Retro-RL: Reinforcing Nominal Controller With Deep Reinforcement Learning for Tilting-Rotor Drones"],"prefix":"10.1109","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9515-7059","authenticated-orcid":false,"given":"I Made Aswin","family":"Nahrendra","sequence":"first","affiliation":[{"name":"School of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5998-8277","authenticated-orcid":false,"given":"Christian","family":"Tirtawardhana","sequence":"additional","affiliation":[{"name":"School of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9900-2166","authenticated-orcid":false,"given":"Byeongho","family":"Yu","sequence":"additional","affiliation":[{"name":"School of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0219-784X","authenticated-orcid":false,"given":"Eungchang Mason","family":"Lee","sequence":"additional","affiliation":[{"name":"School of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5799-2026","authenticated-orcid":false,"given":"Hyun","family":"Myung","sequence":"additional","affiliation":[{"name":"School of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, Republic of Korea"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/s12555-019-0797-7"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TCST.2017.2711602"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2018.2889686"},{"key":"ref4","first-page":"20","article-title":"Development of wall-climbing unmanned aerial vehicle system for micro-inspection of bridges","volume-title":"Proc. IEEE Workshop Future Aerial Robots.: Challenges Opportunities","author":"Myeong","year":"2019"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2021.3108489"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/MRA.2018.2866758"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2018.2885575"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/IROS51168.2021.9636558"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1177\/0278364920943654"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.14264\/uql.2018.636"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.ast.2018.02.020"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2020.3047779"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2017.2720851"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2019.2930489"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/IROS40897.2019.8967695"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2022.3181755"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/IROS51168.2021.9636372"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ISIE.2001.931953"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1016\/0952-1976(91)90045-8"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2021.XVII.042"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2019.XV.019"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2021.3061307"},{"key":"ref23","article-title":"Offline reinforcement learning: Tutorial, review, and perspectives on open problems","volume-title":"arXiv:2005.01643","author":"Levine","year":"2020"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9340823"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2022.3141602"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CDC40024.2019.9028930"},{"key":"ref27","first-page":"2021","article-title":"MRAC-RL: A framework for on-line policy adaptation under parametric model uncertainty","volume-title":"Proc. Learn. Dyn. Control Conf. (L4DC)","author":"Guha"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1214\/aoms\/1177729694"},{"key":"ref29","article-title":"F100 t-motor specifications","year":"2022"},{"key":"ref31","article-title":"Deep learning using rectified linear units (ReLU","author":"Agarap","year":"2018"},{"key":"ref32","article-title":"Adam: A method for stochastic optimization","volume-title":"Proc. Int. Conf. Learn. Representations (ICLR)","author":"Kingma","year":"2015"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2017.8202133"},{"key":"ref34","first-page":"1928","article-title":"Asynchronous methods for deep reinforcement learning","volume-title":"Proc. Int. Conf. Mach. Learn. (ICML)","author":"Mnih","year":"2016"},{"key":"ref35","article-title":"Proximal policy optimization algorithms","author":"Schulman","year":"2017"},{"key":"ref36","article-title":"Isaac gym: High performance GPU-based physics simulation for robot learning","volume-title":"Proc. NeurIPS Track Datasets Benchmarks","author":"Makoviychuk","year":"2021"},{"key":"ref37","first-page":"91","article-title":"Learning to walk in minutes using massively parallel deep reinforcement learning","volume-title":"Proc. Conf. Robot Learn. (CoRL)","author":"Rudin","year":"2021"}],"container-title":["IEEE Robotics and Automation Letters"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7083369\/9831196\/09822205.pdf?arnumber=9822205","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,1]],"date-time":"2024-02-01T04:14:27Z","timestamp":1706760867000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9822205\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,10]]},"references-count":36,"journal-issue":{"issue":"4"},"URL":"https:\/\/doi.org\/10.1109\/lra.2022.3189446","relation":{},"ISSN":["2377-3766","2377-3774"],"issn-type":[{"value":"2377-3766","type":"electronic"},{"value":"2377-3774","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,10]]}}}