{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T10:43:43Z","timestamp":1742985823364,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":19,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819965038"},{"type":"electronic","value":"9789819965045"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-981-99-6504-5_23","type":"book-chapter","created":{"date-parts":[[2023,10,12]],"date-time":"2023-10-12T20:41:21Z","timestamp":1697143281000},"page":"261-272","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Visual-Inertial Fusion Mapping and\u00a0Localization System Using AprilTag in\u00a0GNSS-Denied Scenes"],"prefix":"10.1007","author":[{"given":"Zhihong","family":"Wu","sequence":"first","affiliation":[]},{"given":"Hao","family":"An","sequence":"additional","affiliation":[]},{"given":"Boyu","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Huaide","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ke","family":"Lu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,13]]},"reference":[{"key":"23_CR1","doi-asserted-by":"publisher","first-page":"1327","DOI":"10.1109\/COMST.2016.2632427","volume":"19","author":"A Yassin","year":"2017","unstructured":"Yassin, A., et al.: Recent advances in indoor localization: a survey on theoretical approaches and applications. IEEE Commun. Surv. Tutor. 19, 1327\u20131346 (2017)","journal-title":"IEEE Commun. Surv. Tutor."},{"key":"23_CR2","doi-asserted-by":"publisher","first-page":"3133","DOI":"10.1109\/JIOT.2020.2965115","volume":"7","author":"D Feng","year":"2020","unstructured":"Feng, D., Wang, C., He, C., Zhuang, Y., Xia, X.-G.: Kalman-filter-based integration of IMU and UWB for high-accuracy indoor positioning and navigation. IEEE Internet Things J. 7, 3133\u20133146 (2020)","journal-title":"IEEE Internet Things J."},{"key":"23_CR3","doi-asserted-by":"crossref","unstructured":"Shan, T., Englot, B.: LeGO-LOAM: lightweight and ground-optimized lidar odometry and mapping on variable terrain. In: 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, pp. 4758\u20134765. IEEE Press (2018)","DOI":"10.1109\/IROS.2018.8594299"},{"key":"23_CR4","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"240","DOI":"10.1007\/978-3-031-13835-5_22","volume-title":"Intelligent Robotics and Applications","author":"X Zhou","year":"2022","unstructured":"Zhou, X., Huang, R.: A state-of-the-art review on SLAM. In: Liu, H., et al. (eds.) ICIRA 2022. LNCS, pp. 240\u2013251. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-13835-5_22"},{"key":"23_CR5","doi-asserted-by":"publisher","first-page":"1255","DOI":"10.1109\/TRO.2017.2705103","volume":"33","author":"R Mur-Artal","year":"2017","unstructured":"Mur-Artal, R., Tard\u00f3s, J.D.: ORB-SLAM2: an open-source SLAM system for monocular, stereo, and RGB-D cameras. IEEE Trans. Rob. 33, 1255\u20131262 (2017)","journal-title":"IEEE Trans. Rob."},{"key":"23_CR6","doi-asserted-by":"crossref","unstructured":"Qin, T., Chen, T., Chen, Y., Su, Q.: AVP-SLAM: semantic visual mapping and localization for autonomous vehicles in the parking lot. In: 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5939\u20135945 (2020)","DOI":"10.1109\/IROS45743.2020.9340939"},{"key":"23_CR7","doi-asserted-by":"publisher","first-page":"1619","DOI":"10.1109\/TASE.2019.2900980","volume":"16","author":"H-J Liang","year":"2019","unstructured":"Liang, H.-J., Sanket, N.J., Ferm\u00fcller, C., Aloimonos, Y.: SalientDSO: bringing attention to direct sparse odometry. IEEE Trans. Autom. Sci. Eng. 16, 1619\u20131626 (2019). https:\/\/doi.org\/10.1109\/TASE.2019.2900980","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"23_CR8","doi-asserted-by":"crossref","unstructured":"Olson, E.: AprilTag: a robust and flexible visual fiducial system. In: 2011 IEEE International Conference on Robotics and Automation, pp. 3400\u20133407 (2011)","DOI":"10.1109\/ICRA.2011.5979561"},{"key":"23_CR9","doi-asserted-by":"crossref","unstructured":"Wang, J., Olson, E.: AprilTag 2: efficient and robust fiducial detection. In: 2016 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4193\u20134198 (2016)","DOI":"10.1109\/IROS.2016.7759617"},{"key":"23_CR10","doi-asserted-by":"publisher","unstructured":"Fiala, M.: ARTag, a fiducial marker system using digital techniques. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2005), vol. 2, pp. 590\u2013596 (2005). https:\/\/doi.org\/10.1109\/CVPR.2005.74","DOI":"10.1109\/CVPR.2005.74"},{"key":"23_CR11","doi-asserted-by":"publisher","first-page":"2280","DOI":"10.1016\/j.patcog.2014.01.005","volume":"47","author":"S Garrido-Jurado","year":"2014","unstructured":"Garrido-Jurado, S., Mu\u00f1oz-Salinas, R., Madrid-Cuevas, F.J., Mar\u00edn-Jim\u00e9nez, M.J.: Automatic generation and detection of highly reliable fiducial markers under occlusion. Pattern Recogn. 47, 2280\u20132292 (2014)","journal-title":"Pattern Recogn."},{"key":"23_CR12","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1007\/s10846-020-01307-9","volume":"101","author":"M Kalaitzakis","year":"2021","unstructured":"Kalaitzakis, M., Cain, B., Carroll, S., Ambrosi, A., Whitehead, C., Vitzilaios, N.: Fiducial markers for pose estimation. J. Intell. Robot. Syst. 101, 71 (2021)","journal-title":"J. Intell. Robot. Syst."},{"key":"23_CR13","doi-asserted-by":"crossref","unstructured":"Hoang, V.T., Tang, Q.N., Truong, X.T., Nguyen, D.Q.: An indoor localization method for mobile robot using ceiling mounted AprilTag. JST 17 (2022)","DOI":"10.56651\/lqdtu.jst.v17.n05.531"},{"key":"23_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2021.102241","volume":"73","author":"G Popovi\u0107","year":"2022","unstructured":"Popovi\u0107, G., Cvi\u0161i\u0107, I., \u00c9corchard, G., Markovi\u0107, I., P\u0159eu\u010dil, L., Petrovi\u0107, I.: Human localization in robotized warehouses based on stereo odometry and ground-marker fusion. Robot. Comput.-Integr. Manuf. 73, 102241 (2022)","journal-title":"Robot. Comput.-Integr. Manuf."},{"key":"23_CR15","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2022.107295","volume":"201","author":"W Zhang","year":"2022","unstructured":"Zhang, W., Gong, L., Huang, S., Wu, S., Liu, C.: Factor graph-based high-precision visual positioning for agricultural robots with fiducial markers. Comput. Electron. Agric. 201, 107295 (2022)","journal-title":"Comput. Electron. Agric."},{"key":"23_CR16","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2021.104112","volume":"135","author":"N Kayhani","year":"2022","unstructured":"Kayhani, N., Zhao, W., McCabe, B., Schoellig, A.P.: Tag-based visual-inertial localization of unmanned aerial vehicles in indoor construction environments using an on-manifold ex-tended Kalman filter. Autom. Constr. 135, 104112 (2022)","journal-title":"Autom. Constr."},{"key":"23_CR17","unstructured":"Sol\u00e1, J.: Quaternion kinematics for the error-state Kalman filter. http:\/\/arxiv.org\/abs\/1711.02508 (2017)"},{"key":"23_CR18","doi-asserted-by":"publisher","first-page":"216","DOI":"10.1177\/0278364911430419","volume":"31","author":"M Kaess","year":"2012","unstructured":"Kaess, M., Johannsson, H., Roberts, R., Ila, V., Leonard, J.J., Dellaert, F.: iSAM2: incremental smoothing and mapping using the Bayes tree. Int. J. Robot. Res. 31, 216\u2013235 (2012)","journal-title":"Int. J. Robot. Res."},{"key":"23_CR19","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1109\/TRO.2011.2170332","volume":"28","author":"T Lupton","year":"2012","unstructured":"Lupton, T., Sukkarieh, S.: Visual-inertial-aided navigation for high-dynamic motion in built environments without initial conditions. IEEE Trans. Rob. 28, 61\u201376 (2012). https:\/\/doi.org\/10.1109\/TRO.2011.2170332","journal-title":"IEEE Trans. Rob."}],"container-title":["Lecture Notes in Computer Science","Intelligent Robotics and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-6504-5_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,12]],"date-time":"2023-10-12T20:43:13Z","timestamp":1697143393000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-6504-5_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9789819965038","9789819965045"],"references-count":19,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-6504-5_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"13 October 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIRA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Robotics and Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hangzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 July 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 July 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icira2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icira2023.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"630","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"431","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"68% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}