{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T17:02:35Z","timestamp":1770742955091,"version":"3.49.0"},"reference-count":54,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2023,5,1]],"date-time":"2023-05-01T00:00:00Z","timestamp":1682899200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2023,5,1]],"date-time":"2023-05-01T00:00:00Z","timestamp":1682899200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2023,3,7]],"date-time":"2023-03-07T00:00:00Z","timestamp":1678147200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Robotics and Autonomous Systems"],"published-print":{"date-parts":[[2023,5]]},"DOI":"10.1016\/j.robot.2023.104398","type":"journal-article","created":{"date-parts":[[2023,3,7]],"date-time":"2023-03-07T16:36:22Z","timestamp":1678206982000},"page":"104398","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":30,"special_numbering":"C","title":["ArTuga: A novel multimodal fiducial marker for aerial robotics"],"prefix":"10.1016","volume":"163","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8010-7507","authenticated-orcid":false,"given":"Rafael Marques","family":"Claro","sequence":"first","affiliation":[]},{"given":"Diogo Brand\u00e3o","family":"Silva","sequence":"additional","affiliation":[]},{"given":"Andry Maykol","family":"Pinto","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.robot.2023.104398_b1","doi-asserted-by":"crossref","DOI":"10.1016\/j.comnet.2020.107148","article-title":"A compilation of UAV applications for precision agriculture","volume":"172","author":"Radoglou-Grammatikis","year":"2020","journal-title":"Comput. Netw."},{"key":"10.1016\/j.robot.2023.104398_b2","doi-asserted-by":"crossref","first-page":"445","DOI":"10.1007\/s10846-016-0348-x","article-title":"Coverage path planning for UAVs photogrammetry with energy and resolution constraints","volume":"83","author":"Franco","year":"2016","journal-title":"J. Intell. Robot. Syst."},{"key":"10.1016\/j.robot.2023.104398_b3","doi-asserted-by":"crossref","first-page":"1262","DOI":"10.3390\/rs14051262","article-title":"Development of semantic maps of vegetation cover from UAV images to support planning and management in fine-grained fire-prone landscapes","volume":"14","author":"Tren\u010danov\u00e1","year":"2022","journal-title":"Remote Sens."},{"key":"10.1016\/j.robot.2023.104398_b4","series-title":"2015 23rd Euromicro International Conference on Parallel, Distributed, and Network-Based Processing","first-page":"309","article-title":"Human and fire detection from high altitude UAV images","author":"Giitsidis","year":"2015"},{"key":"10.1016\/j.robot.2023.104398_b5","doi-asserted-by":"crossref","first-page":"32","DOI":"10.1016\/j.micpro.2018.05.014","article-title":"ROLFER: A fully autonomous aerial rescue support system","volume":"61","author":"Lygouras","year":"2018","journal-title":"Microprocess. Microsyst."},{"key":"10.1016\/j.robot.2023.104398_b6","series-title":"2019 Offshore Energy and Storage Summit","first-page":"1","article-title":"On the energy requirements of UAVs used for blade inspection in offshore wind farms","author":"Aquilina","year":"2019"},{"key":"10.1016\/j.robot.2023.104398_b7","series-title":"OCEANS 2021: San Diego \u2013 Porto","first-page":"1","article-title":"DIIUS - Distributed perception for inspection of aquatic structures","author":"Campos","year":"2021"},{"key":"10.1016\/j.robot.2023.104398_b8","series-title":"OCEANS 2021: San Diego \u2013 Porto","first-page":"1","article-title":"ATLANTIS - The atlantic testing platform for maritime robotics","author":"Pinto","year":"2021"},{"key":"10.1016\/j.robot.2023.104398_b9","doi-asserted-by":"crossref","first-page":"179252","DOI":"10.1109\/ACCESS.2019.2958680","article-title":"MEC-driven UAV-enabled routine inspection scheme in wind farm under wind influence","volume":"7","author":"Cao","year":"2019","journal-title":"IEEE Access"},{"issue":"1","key":"10.1016\/j.robot.2023.104398_b10","doi-asserted-by":"crossref","first-page":"137","DOI":"10.1002\/rob.21898","article-title":"Long-duration fully autonomous operation of rotorcraft unmanned aerial systems for remote-sensing data acquisition","volume":"37","author":"Malyuta","year":"2020","journal-title":"J. Field Robotics"},{"issue":"12","key":"10.1016\/j.robot.2023.104398_b11","doi-asserted-by":"crossref","DOI":"10.3390\/electronics8121532","article-title":"Accurate landing of unmanned aerial vehicles using ground pattern recognition","volume":"8","author":"Wubben","year":"2019","journal-title":"Electronics"},{"key":"10.1016\/j.robot.2023.104398_b12","series-title":"2020 IEEE International Conference on Autonomous Robot Systems and Competitions","first-page":"221","article-title":"Multi-domain mapping for offshore asset inspection using an autonomous surface vehicle","author":"Campos","year":"2020"},{"key":"10.1016\/j.robot.2023.104398_b13","doi-asserted-by":"crossref","DOI":"10.1007\/s42452-021-04451-5","article-title":"Multi-domain inspection of offshore wind farms using an autonomous surface vehicle","volume":"3","author":"Campos","year":"2021","journal-title":"SN Appl. Sci."},{"key":"10.1016\/j.robot.2023.104398_b14","series-title":"Image Analysis and Recognition","first-page":"379","article-title":"Dense disparity maps from RGB and sparse depth information using deep regression models","author":"Leite","year":"2020"},{"key":"10.1016\/j.robot.2023.104398_b15","article-title":"Maresye: A hybrid imaging system for underwater robotic applications","volume":"55","author":"Pinto","year":"2019","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.robot.2023.104398_b16","doi-asserted-by":"crossref","first-page":"76056","DOI":"10.1109\/ACCESS.2021.3082697","article-title":"Exploiting motion perception in depth estimation through a lightweight convolutional neural network","volume":"9","author":"Leite","year":"2021","journal-title":"IEEE Access"},{"key":"10.1016\/j.robot.2023.104398_b17","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.robot.2016.08.014","article-title":"Visual motion perception for mobile robots through dense optical flow fields","volume":"87","author":"Pinto","year":"2017","journal-title":"Robot. Auton. Syst."},{"issue":"6","key":"10.1016\/j.robot.2023.104398_b18","doi-asserted-by":"crossref","first-page":"391","DOI":"10.1016\/j.imavis.2014.04.003","article-title":"Unsupervised flow-based motion analysis for an autonomous moving system","volume":"32","author":"Pinto","year":"2014","journal-title":"Image Vis. Comput."},{"issue":"1","key":"10.1016\/j.robot.2023.104398_b19","doi-asserted-by":"crossref","first-page":"80","DOI":"10.1016\/j.image.2013.11.003","article-title":"Enhancing dynamic videos for surveillance and robotic applications: The robust bilateral and temporal filter","volume":"29","author":"Maykol Pinto","year":"2014","journal-title":"Signal Process., Image Commun."},{"issue":"21","key":"10.1016\/j.robot.2023.104398_b20","doi-asserted-by":"crossref","DOI":"10.3390\/s19214703","article-title":"An onboard vision-based system for autonomous landing of a low-cost quadrotor on a novel landing pad","volume":"19","author":"Liu","year":"2019","journal-title":"Sensors"},{"key":"10.1016\/j.robot.2023.104398_b21","doi-asserted-by":"crossref","first-page":"221","DOI":"10.1007\/s10846-010-9473-0","article-title":"Automatic take off, tracking and landing of a miniature UAV on a moving carrier vehicle","volume":"61","author":"Wenzel","year":"2011","journal-title":"J. Intell. Robot. Syst."},{"key":"10.1016\/j.robot.2023.104398_b22","doi-asserted-by":"crossref","DOI":"10.1007\/s11263-014-0725-5","article-title":"Infinitesimal plane-based pose estimation","volume":"109","author":"Collins","year":"2014","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.robot.2023.104398_b23","doi-asserted-by":"crossref","DOI":"10.1007\/s10846-016-0399-z","article-title":"Vision based autonomous landing of multirotor UAV on moving platform","volume":"85","author":"Araar","year":"2017","journal-title":"J. Intell. Robot. Syst."},{"key":"10.1016\/j.robot.2023.104398_b24","doi-asserted-by":"crossref","first-page":"447","DOI":"10.1016\/j.eswa.2017.09.033","article-title":"Survey of computer vision algorithms and applications for unmanned aerial vehicles","volume":"92","author":"Al-Kaff","year":"2018","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.robot.2023.104398_b25","series-title":"2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Vol. 2","first-page":"590","article-title":"ARTag, a fiducial marker system using digital techniques","author":"Fiala","year":"2005"},{"key":"10.1016\/j.robot.2023.104398_b26","series-title":"2011 IEEE International Conference on Robotics and Automation","first-page":"3400","article-title":"AprilTag: A robust and flexible visual fiducial system","author":"Olson","year":"2011"},{"key":"10.1016\/j.robot.2023.104398_b27","series-title":"2016 IEEE\/RSJ International Conference on Intelligent Robots and Systems","first-page":"4193","article-title":"AprilTag 2: Efficient and robust fiducial detection","author":"Wang","year":"2016"},{"issue":"6","key":"10.1016\/j.robot.2023.104398_b28","doi-asserted-by":"crossref","first-page":"2280","DOI":"10.1016\/j.patcog.2014.01.005","article-title":"Automatic generation and detection of highly reliable fiducial markers under occlusion","volume":"47","author":"Garrido-Jurado","year":"2014","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.robot.2023.104398_b29","series-title":"2016 IEEE Conference on Computer Vision and Pattern Recognition","first-page":"562","article-title":"Detection and accurate localization of circular fiducials under highly challenging conditions","author":"Calvet","year":"2016"},{"key":"10.1016\/j.robot.2023.104398_b30","doi-asserted-by":"crossref","first-page":"158","DOI":"10.1016\/j.imavis.2019.06.007","article-title":"STag: A stable fiducial marker system","volume":"89","author":"Benligiray","year":"2019","journal-title":"Image Vis. Comput."},{"issue":"5","key":"10.1016\/j.robot.2023.104398_b31","doi-asserted-by":"crossref","first-page":"667","DOI":"10.1002\/rob.21467","article-title":"Automated vision-based recovery of a rotary wing unmanned aerial vehicle onto a moving platform","volume":"30","author":"Richardson","year":"2013","journal-title":"J. Field Robotics"},{"issue":"5","key":"10.1016\/j.robot.2023.104398_b32","doi-asserted-by":"crossref","first-page":"874","DOI":"10.1002\/rob.21858","article-title":"Autonomous landing on a moving vehicle with an unmanned aerial vehicle","volume":"36","author":"Baca","year":"2019","journal-title":"J. Field Robotics"},{"issue":"1","key":"10.1016\/j.robot.2023.104398_b33","doi-asserted-by":"crossref","first-page":"204","DOI":"10.1002\/rob.21817","article-title":"Team NimbRo at MBZIRC 2017: Fast landing on a moving target and treasure hunting with a team of micro aerial vehicles","volume":"36","author":"Beul","year":"2019","journal-title":"J. Field Robotics"},{"issue":"1","key":"10.1016\/j.robot.2023.104398_b34","doi-asserted-by":"crossref","first-page":"78","DOI":"10.1002\/rob.21824","article-title":"The ETH-MAV team in the MBZ international robotics challenge","volume":"36","author":"B\u00e4hnemann","year":"2019","journal-title":"J. Field Robotics"},{"issue":"1","key":"10.1016\/j.robot.2023.104398_b35","doi-asserted-by":"crossref","first-page":"34","DOI":"10.1002\/rob.21815","article-title":"Fast vision-based autonomous detection of moving cooperative target for unmanned aerial vehicle landing","volume":"36","author":"Li","year":"2019","journal-title":"J. Field Robotics"},{"issue":"6","key":"10.1016\/j.robot.2023.104398_b36","doi-asserted-by":"crossref","DOI":"10.3390\/s18061703","article-title":"LightDenseYOLO: A fast and accurate marker tracker for autonomous UAV landing by visible light camera sensor on drone","volume":"18","author":"Nguyen","year":"2018","journal-title":"Sensors"},{"issue":"9","key":"10.1016\/j.robot.2023.104398_b37","doi-asserted-by":"crossref","DOI":"10.3390\/s17091987","article-title":"Remote marker-based tracking for UAV landing using visible-light camera sensor","volume":"17","author":"Nguyen","year":"2017","journal-title":"Sensors"},{"key":"10.1016\/j.robot.2023.104398_b38","doi-asserted-by":"crossref","first-page":"22421","DOI":"10.1109\/ACCESS.2018.2801028","article-title":"Robust tracking through the design of high quality fiducial markers: An optimization tool for ARToolKit","volume":"6","author":"Khan","year":"2018","journal-title":"IEEE Access"},{"key":"10.1016\/j.robot.2023.104398_b39","doi-asserted-by":"crossref","first-page":"45291","DOI":"10.1109\/ACCESS.2019.2909467","article-title":"Reflective fiducials for localization with 3D light detection and ranging scanners","volume":"7","author":"Davis","year":"2019","journal-title":"IEEE Access"},{"key":"10.1016\/j.robot.2023.104398_b40","doi-asserted-by":"crossref","first-page":"297","DOI":"10.1007\/s10846-009-9355-5","article-title":"Low-cost visual tracking of a landing place and hovering flight control with a microcontroller","volume":"57","author":"Wenzel","year":"2010","journal-title":"J. Intell. Robot. Syst."},{"key":"10.1016\/j.robot.2023.104398_b41","series-title":"2019 IEEE 89th Vehicular Technology Conference","first-page":"1","article-title":"High-precision UAV localization system for landing on a mobile collaborative robot based on an IR marker pattern recognition","author":"Kalinov","year":"2019"},{"issue":"6","key":"10.1016\/j.robot.2023.104398_b42","doi-asserted-by":"crossref","first-page":"1498","DOI":"10.1016\/j.cja.2013.07.049","article-title":"Use of land\u2019s cooperative object to estimate UAV\u2019s pose for autonomous landing","volume":"26","author":"Xu","year":"2013","journal-title":"Chin. J. Aeronaut."},{"issue":"6","key":"10.1016\/j.robot.2023.104398_b43","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1016\/j.patrec.2008.12.011","article-title":"Research on computer vision-based for UAV autonomous landing on a ship","volume":"30","author":"Xu","year":"2009","journal-title":"Pattern Recognit. Lett."},{"key":"10.1016\/j.robot.2023.104398_b44","series-title":"2018 International Symposium on Visual Computing","first-page":"565","article-title":"Marker based thermal-inertial localization for aerial robots in obscurant filled environments","author":"Khattak","year":"2018"},{"issue":"10","key":"10.1016\/j.robot.2023.104398_b45","doi-asserted-by":"crossref","DOI":"10.3390\/s17102173","article-title":"Uncooled thermal camera calibration and optimization of the photogrammetry process for UAV applications in agriculture","volume":"17","author":"Ribeiro-Gomes","year":"2017","journal-title":"Sensors"},{"key":"10.1016\/j.robot.2023.104398_b46","doi-asserted-by":"crossref","first-page":"53030","DOI":"10.1109\/ACCESS.2021.3070694","article-title":"Advancing autonomous surface vehicles: A 3D perception system for the recognition and assessment of docking-based structures","volume":"9","author":"Pereira","year":"2021","journal-title":"IEEE Access"},{"key":"10.1016\/j.robot.2023.104398_b47","series-title":"2015 IEEE International Conference on Autonomous Robot Systems and Competitions","first-page":"139","article-title":"Evaluation of depth sensors for robotic applications","author":"Pinto","year":"2015"},{"key":"10.1016\/j.robot.2023.104398_b48","series-title":"2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems","first-page":"5562","article-title":"Automatic extrinsic calibration of a camera and a 3D LiDAR using line and plane correspondences","author":"Zhou","year":"2018"},{"key":"10.1016\/j.robot.2023.104398_b49","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2021.114894","article-title":"A camera to LiDAR calibration approach through the optimization of atomic transformations","volume":"176","author":"Pinto de Aguiar","year":"2021","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.robot.2023.104398_b50","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2022.118000","article-title":"ATOM: A general calibration framework for multi-modal, multi-sensor systems","volume":"207","author":"Oliveira","year":"2022","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.robot.2023.104398_b51","doi-asserted-by":"crossref","first-page":"452","DOI":"10.1177\/0278364911435689","article-title":"3D LIDAR\u2013camera intrinsic and extrinsic calibration: Identifiability and analytical least-squares-based initialization","volume":"31","author":"Faraz","year":"2012","journal-title":"Int. J. Robot. Res."},{"issue":"15","key":"10.1016\/j.robot.2023.104398_b52","doi-asserted-by":"crossref","DOI":"10.3390\/s22155576","article-title":"An effective camera-to-lidar spatiotemporal calibration based on a simple calibration target","volume":"22","author":"Grammatikopoulos","year":"2022","journal-title":"Sensors"},{"key":"10.1016\/j.robot.2023.104398_b53","series-title":"2D and 3D Object Measurement for Control and Quality Assurance in the Industry","author":"Gramkow","year":"1999"},{"key":"10.1016\/j.robot.2023.104398_b54","series-title":"The 3rd International Conference on Information Sciences and Interaction Sciences","first-page":"403","article-title":"A camera calibration technique based on opencv","author":"Wang","year":"2010"}],"container-title":["Robotics and Autonomous Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0921889023000374?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0921889023000374?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,10,13]],"date-time":"2025-10-13T04:37:09Z","timestamp":1760330229000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0921889023000374"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,5]]},"references-count":54,"alternative-id":["S0921889023000374"],"URL":"https:\/\/doi.org\/10.1016\/j.robot.2023.104398","relation":{},"ISSN":["0921-8890"],"issn-type":[{"value":"0921-8890","type":"print"}],"subject":[],"published":{"date-parts":[[2023,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"ArTuga: A novel multimodal fiducial marker for aerial robotics","name":"articletitle","label":"Article Title"},{"value":"Robotics and Autonomous Systems","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.robot.2023.104398","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2023 The Authors. Published by Elsevier B.V.","name":"copyright","label":"Copyright"}],"article-number":"104398"}}