{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T04:30:02Z","timestamp":1775190602550,"version":"3.50.1"},"reference-count":33,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T00:00:00Z","timestamp":1728864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T00:00:00Z","timestamp":1728864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,10,14]]},"DOI":"10.1109\/iros58592.2024.10802831","type":"proceedings-article","created":{"date-parts":[[2024,12,25]],"date-time":"2024-12-25T19:17:39Z","timestamp":1735154259000},"page":"11812-11818","source":"Crossref","is-referenced-by-count":4,"title":["The Power of Input: Benchmarking Zero-Shot Sim-to-Real Transfer of Reinforcement Learning Control Policies for Quadrotor Control"],"prefix":"10.1109","author":[{"given":"Alberto","family":"Dionigi","sequence":"first","affiliation":[{"name":"New York University,Tandon School of Engineering,Brooklyn,NY,USA,11201"}]},{"given":"Gabriele","family":"Costante","sequence":"additional","affiliation":[{"name":"University of Perugia,Department of Engineering,Perugia,Italy,06125"}]},{"given":"Giuseppe","family":"Loianno","sequence":"additional","affiliation":[{"name":"New York University,Tandon School of Engineering,Brooklyn,NY,USA,11201"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.arcontrol.2018.10.009"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICMECH.2009.4957154"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/s10846-016-0351-2"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0264471"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TSMC.2018.2884725"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/IROS51168.2021.9636053"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CJECE.2019.2920938"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.2004.1302409"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2004.1389776"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1117\/12.2223329"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ACC.2012.6314682"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CHICC.2008.4605334"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.2514\/6.2012-4917"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2006.282433"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1016\/j.conengprac.2013.12.017"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2016.7487274"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2018.8593739"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA46639.2022.9812077"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2023.3246839"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2017.2720851"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICAR58858.2023.10407017"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-023-06419-4"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/TIE.2022.3144586"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2022.3181755"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA46639.2022.9811564"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2015.2479878"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2018.XIV.008"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2022.3150866"},{"key":"ref29","first-page":"5556","article-title":"Controlling overestimation bias with truncated mixture of continuous distributional quantile critics","volume-title":"International Conference on Machine Learning (ICML)","author":"Kuznetsov"},{"key":"ref30","first-page":"1861","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","volume-title":"International Conference on Machine Learning (ICML)","author":"Haarnoja"},{"issue":"268","key":"ref31","first-page":"1","article-title":"Stable-baselines3: Reliable reinforcement learning implementations","volume":"22","author":"Raffin","year":"2021","journal-title":"Journal of Machine Learning Research"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2016.2633290"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1016\/j.arcontrol.2023.100900"}],"event":{"name":"2024 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)","location":"Abu Dhabi, United Arab Emirates","start":{"date-parts":[[2024,10,14]]},"end":{"date-parts":[[2024,10,18]]}},"container-title":["2024 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10801246\/10801290\/10802831.pdf?arnumber=10802831","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,26]],"date-time":"2024-12-26T07:13:12Z","timestamp":1735197192000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10802831\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,14]]},"references-count":33,"URL":"https:\/\/doi.org\/10.1109\/iros58592.2024.10802831","relation":{},"subject":[],"published":{"date-parts":[[2024,10,14]]}}}