{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T13:42:29Z","timestamp":1765546949543,"version":"3.37.3"},"reference-count":58,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,5,23]],"date-time":"2022-05-23T00:00:00Z","timestamp":1653264000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,5,23]],"date-time":"2022-05-23T00:00:00Z","timestamp":1653264000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["1943205,2024741"],"award-info":[{"award-number":["1943205,2024741"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,5,23]]},"DOI":"10.1109\/icra46639.2022.9811695","type":"proceedings-article","created":{"date-parts":[[2022,7,12]],"date-time":"2022-07-12T19:36:40Z","timestamp":1657654600000},"page":"1113-1121","source":"Crossref","is-referenced-by-count":21,"title":["High Definition, Inexpensive, Underwater Mapping"],"prefix":"10.1109","author":[{"given":"Bharat","family":"Joshi","sequence":"first","affiliation":[{"name":"University of South Carolina,Computer Science and Engineering Department,Columbia,SC,USA,29208"}]},{"given":"Marios","family":"Xanthidis","sequence":"additional","affiliation":[{"name":"University of South Carolina,Computer Science and Engineering Department,Columbia,SC,USA,29208"}]},{"given":"Sharmin","family":"Rahman","sequence":"additional","affiliation":[{"name":"University of South Carolina,Computer Science and Engineering Department,Columbia,SC,USA,29208"}]},{"given":"Ioannis","family":"Rekleitis","sequence":"additional","affiliation":[{"name":"University of South Carolina,Computer Science and Engineering Department,Columbia,SC,USA,29208"}]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/38.909011"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1177\/0278364919883346"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.3390\/rs12060997"},{"key":"ref32","first-page":"35","article-title":"Wreck and ruin-mapping the shipwreck resource of qatar","author":"bain","year":"2014","journal-title":"Proceedings of the Seminar for Arabian Studies"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.4031\/002533202787913387"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2018.2808364"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.2004.1308080"},{"key":"ref36","first-page":"1","article-title":"Large-scale habitat mapping using vision-based auvs: Experiences, chal-lenges & vehicle design","author":"dunbabin","year":"2007","journal-title":"OCEANS 2007-Europe IEEE"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2006.282097"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-27702-8_4"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1016\/j.culher.2018.02.017"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1016\/j.ifacol.2015.06.032"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.5194\/isprsarchives-XL-5-W5-7-2015"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1177\/0278364906072512"},{"journal-title":"The silent world","year":"1953","author":"cousteau","key":"ref1"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.5194\/isprs-archives-XLII-2-W10-135-2019"},{"key":"ref22","first-page":"9","article-title":"Third-generation underwater landscape mosaics for coral reef mapping and moni-toring","author":"gintert","year":"0","journal-title":"Proceedings of the 12th international coral reef symposium Cairns Australia"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.5194\/isprs-archives-XLI-B5-821-2016"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.7717\/peerj.1960"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.5194\/isprsarchives-XL-5-W5-231-2015"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1016\/j.ifacol.2015.06.034"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/OCEANS.2012.6404786"},{"journal-title":"Gpmf-parser","year":"2021","key":"ref50"},{"journal-title":"Kalibr-Allan","year":"2021","author":"geneva","key":"ref51"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/IROS51168.2021.9636124"},{"key":"ref57","first-page":"892","article-title":"N avigation in the presence of obstacles for an agile autonomous underwater vehicle","author":"xanthidis","year":"0","journal-title":"IEEE International Conference on Robotics and Automation"},{"key":"ref56","doi-asserted-by":"crossref","DOI":"10.1109\/TRO.2020.3033695","article-title":"Teaser: Fast and cer-tifiable point cloud registration","author":"yang","year":"2021","journal-title":"IEEE Transactions on Robotics"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.2009.5152473"},{"key":"ref54","doi-asserted-by":"crossref","first-page":"376","DOI":"10.1109\/34.88573","article-title":"Least-squares estimation of transfor-mation parameters between two point patterns","volume":"13","author":"umeyama","year":"1991","journal-title":"IEEE Trans Pattern Anal Mach In-tell"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2005.74"},{"key":"ref52","article-title":"The turn vi benchmark for evaluating visual-inertial odometry","author":"schubert","year":"0","journal-title":"International Conference on Intelligent Robots and Systems (IROS)"},{"key":"ref10","first-page":"1861","article-title":"An Underwater SLAM System using Sonar, Visual, In-ertial, and Depth Sensor","author":"rahman","year":"0","journal-title":"IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)"},{"key":"ref11","doi-asserted-by":"crossref","DOI":"10.1109\/TRO.2018.2853729","article-title":"VINS-Mono: A robust and versatile monocular visual-inertial state estimator","volume":"34","author":"qin","year":"2018","journal-title":"IEEE Transactions on Robotics"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1061\/40476(299)19"},{"key":"ref12","doi-asserted-by":"crossref","DOI":"10.1109\/TRO.2021.3075644","article-title":"ORB-SLAM3: An accurate open-source library for visual, visual-inertial, and multimap slam","author":"campos","year":"2021","journal-title":"IEEE Transactions on Robotics"},{"key":"ref13","article-title":"Open VINS: A research platform for visual-inertial estimation","author":"geneva","year":"0","journal-title":"Proc of the IEEE International Conference on Robotics and Automation"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.445"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/OCEANS.2018.8604819"},{"journal-title":"KELDAN VIDEO 8X LIGHT (18 000 LUMEN CRI85)","year":"2021","key":"ref16"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989672"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.5194\/isprs-annals-IV-2-W5-575-2019"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.5194\/isprs-archives-XLII-2-W10-143-2019"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1002\/aqc.2505"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1179\/0093469014Z.00000000077"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.5244\/C.22.101"},{"key":"ref5","first-page":"1874","article-title":"Augmenting coral reef monitoring with an enhanced detection system utilizing a deep semi -supervised learning approach","author":"modasshir","year":"0","journal-title":"IEEE International Conference on Robotics and Automation"},{"key":"ref8","first-page":"7221","article-title":"Exper-imental Comparison of Open Source Visual- Inertial-Based State Estimation Algorithms in the Underwater Domain","author":"joshi","year":"0","journal-title":"IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/IROS40897.2019.8967557"},{"journal-title":"HERO9 Black - Video Settings and Resolutions","year":"2021","key":"ref49"},{"key":"ref9","article-title":"Ex-perimental comparison of open source vision based state estimation algorithms","author":"li","year":"0","journal-title":"International Sympo-sium of Experimental Robotics (ISER)"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1177\/0278364914554813"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8460545"},{"key":"ref48","first-page":"3400","article-title":"Apriltag: A robust and flexible visual fidu-cial system","author":"olson","year":"0","journal-title":"2011 IEEE International Conference on Robotics and Automation"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/IROS40897.2019.8967697"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.5194\/isprs-archives-XLII-1-329-2018"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1061\/41003(327)20"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1002\/rob.21640"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.5194\/isprs-archives-XLII-2-W15-857-2019"}],"event":{"name":"2022 IEEE International Conference on Robotics and Automation (ICRA)","start":{"date-parts":[[2022,5,23]]},"location":"Philadelphia, PA, USA","end":{"date-parts":[[2022,5,27]]}},"container-title":["2022 International Conference on Robotics and Automation (ICRA)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9811522\/9811357\/09811695.pdf?arnumber=9811695","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,29]],"date-time":"2024-09-29T00:06:21Z","timestamp":1727568381000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9811695\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,23]]},"references-count":58,"URL":"https:\/\/doi.org\/10.1109\/icra46639.2022.9811695","relation":{},"subject":[],"published":{"date-parts":[[2022,5,23]]}}}