{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T15:26:19Z","timestamp":1775661979705,"version":"3.50.1"},"reference-count":39,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"7","license":[{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"am","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Technology Innovation Institute"},{"name":"NSF CAREER","award":["2145277"],"award-info":[{"award-number":["2145277"]}]},{"name":"DARPA YFA","award":["D22AP00156-00"],"award-info":[{"award-number":["D22AP00156-00"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Robot. Autom. Lett."],"published-print":{"date-parts":[[2024,7]]},"DOI":"10.1109\/lra.2024.3396025","type":"journal-article","created":{"date-parts":[[2024,5,1]],"date-time":"2024-05-01T17:33:27Z","timestamp":1714584807000},"page":"6336-6343","source":"Crossref","is-referenced-by-count":34,"title":["Learning to Fly in Seconds"],"prefix":"10.1109","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3320-6502","authenticated-orcid":false,"given":"Jonas","family":"Eschmann","sequence":"first","affiliation":[{"name":"New York University, Tandon School of Engineering, Brooklyn, NY, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6692-4925","authenticated-orcid":false,"given":"Dario","family":"Albani","sequence":"additional","affiliation":[{"name":"Autonomous Robotics Research Center, Technology Innovation Institute, Abu Dhabi, UAE"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3263-5401","authenticated-orcid":false,"given":"Giuseppe","family":"Loianno","sequence":"additional","affiliation":[{"name":"New York University, Tandon School of Engineering, Brooklyn, NY, USA"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/3065386"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"key":"ref4","article-title":"Continuous control with deep reinforcement learning","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Lillicrap","year":"2016"},{"key":"ref5","article-title":"High-dimensional continuous control using generalized advantage estimation","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Schulman","year":"2016"},{"key":"ref6","article-title":"Proximal policy optimization algorithms","author":"Schulman","year":"2017"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/0893-6080(89)90020-8"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2011.5980409"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-662-44468-9_58"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-67361-5_40"},{"key":"ref11","first-page":"1147","article-title":"Flightmare: A flexible quadrotor simulator","volume-title":"Proc. Conf. Robot Learn.","author":"Song","year":"2020"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/iros51168.2021.9635857"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-26054-9_23"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2023.3336320"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2017.XIII.034"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICUAS.2018.8453449"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2018.8594249"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2021.3057046"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1007\/s10514-020-09951-8"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA46639.2022.9811564"},{"key":"ref21","article-title":"Learning to Fly via deep model-based reinforcement learning","author":"Becker-Ehmck","year":"2020"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/IROS40897.2019.8967944"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/IROS51168.2021.9636096"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2020.XVI.040"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-023-06419-4"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2017.2720851"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/IROS40897.2019.8967695"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1016\/j.conengprac.2019.104222"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/IROS51168.2021.9636053"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/IROS47612.2022.9981229"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48891.2023.10160836"},{"key":"ref32","first-page":"1587","article-title":"Addressing function approximation error in actor-critic methods","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Fujimoto","year":"2018"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2018.XIV.008"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-41188-6_3"},{"key":"ref35","article-title":"RLtools: A fast, portable deep reinforcement learning library for continuous control","author":"Eschmann","year":"2023"},{"key":"ref36","article-title":"Nonlinear quadrocopter attitude control: Technical report","author":"Brescianini","year":"2013"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.2514\/1.G001490"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1145\/3306346.3322940"},{"key":"ref39","article-title":"Partially unsupervised deep meta-reinforcement learning","author":"Eschmann","year":"2021"}],"container-title":["IEEE Robotics and Automation Letters"],"original-title":[],"link":[{"URL":"https:\/\/ieeexplore.ieee.org\/ielam\/7083369\/10534628\/10517383-aam.pdf","content-type":"application\/pdf","content-version":"am","intended-application":"syndication"},{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7083369\/10534628\/10517383.pdf?arnumber=10517383","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,1]],"date-time":"2024-06-01T04:46:51Z","timestamp":1717217211000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10517383\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7]]},"references-count":39,"journal-issue":{"issue":"7"},"URL":"https:\/\/doi.org\/10.1109\/lra.2024.3396025","relation":{},"ISSN":["2377-3766","2377-3774"],"issn-type":[{"value":"2377-3766","type":"electronic"},{"value":"2377-3774","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7]]}}}