{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T18:17:44Z","timestamp":1772907464108,"version":"3.50.1"},"reference-count":26,"publisher":"MDPI AG","issue":"6","license":[{"start":{"date-parts":[[2024,3,14]],"date-time":"2024-03-14T00:00:00Z","timestamp":1710374400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100002341","name":"Academy of Finland projects \u201cUltrafast Data Production with Broadband Photodetectors for Active Hyperspectral Space Imaging\u201d","doi-asserted-by":"publisher","award":["336145"],"award-info":[{"award-number":["336145"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland projects \u201cUltrafast Data Production with Broadband Photodetectors for Active Hyperspectral Space Imaging\u201d","doi-asserted-by":"publisher","award":["353363"],"award-info":[{"award-number":["353363"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland projects \u201cUltrafast Data Production with Broadband Photodetectors for Active Hyperspectral Space Imaging\u201d","doi-asserted-by":"publisher","award":["337656"],"award-info":[{"award-number":["337656"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland projects \u201cUltrafast Data Production with Broadband Photodetectors for Active Hyperspectral Space Imaging\u201d","doi-asserted-by":"publisher","award":["314312"],"award-info":[{"award-number":["314312"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland projects \u201cUltrafast Data Production with Broadband Photodetectors for Active Hyperspectral Space Imaging\u201d","doi-asserted-by":"publisher","award":["343678"],"award-info":[{"award-number":["343678"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland projects \u201cUltrafast Data Production with Broadband Photodetectors for Active Hyperspectral Space Imaging\u201d","doi-asserted-by":"publisher","award":["181811KYSB20160040"],"award-info":[{"award-number":["181811KYSB20160040"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland projects \u201cUltrafast Data Production with Broadband Photodetectors for Active Hyperspectral Space Imaging\u201d","doi-asserted-by":"publisher","award":["XDA22030202"],"award-info":[{"award-number":["XDA22030202"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland projects \u201cUltrafast Data Production with Broadband Photodetectors for Active Hyperspectral Space Imaging\u201d","doi-asserted-by":"publisher","award":["1003-YQR23046"],"award-info":[{"award-number":["1003-YQR23046"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cBright- Biotic Damage Mapping with Ultrawide Spectral Range LiDARs for Sustainable Forest Growth\u201d","doi-asserted-by":"publisher","award":["336145"],"award-info":[{"award-number":["336145"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cBright- Biotic Damage Mapping with Ultrawide Spectral Range LiDARs for Sustainable Forest Growth\u201d","doi-asserted-by":"publisher","award":["353363"],"award-info":[{"award-number":["353363"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cBright- Biotic Damage Mapping with Ultrawide Spectral Range LiDARs for Sustainable Forest Growth\u201d","doi-asserted-by":"publisher","award":["337656"],"award-info":[{"award-number":["337656"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cBright- Biotic Damage Mapping with Ultrawide Spectral Range LiDARs for Sustainable Forest Growth\u201d","doi-asserted-by":"publisher","award":["314312"],"award-info":[{"award-number":["314312"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cBright- Biotic Damage Mapping with Ultrawide Spectral Range LiDARs for Sustainable Forest Growth\u201d","doi-asserted-by":"publisher","award":["343678"],"award-info":[{"award-number":["343678"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cBright- Biotic Damage Mapping with Ultrawide Spectral Range LiDARs for Sustainable Forest Growth\u201d","doi-asserted-by":"publisher","award":["181811KYSB20160040"],"award-info":[{"award-number":["181811KYSB20160040"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cBright- Biotic Damage Mapping with Ultrawide Spectral Range LiDARs for Sustainable Forest Growth\u201d","doi-asserted-by":"publisher","award":["XDA22030202"],"award-info":[{"award-number":["XDA22030202"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cBright- Biotic Damage Mapping with Ultrawide Spectral Range LiDARs for Sustainable Forest Growth\u201d","doi-asserted-by":"publisher","award":["1003-YQR23046"],"award-info":[{"award-number":["1003-YQR23046"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cForest-Human\u2013Machine Interplay\u2014Building Resilience, Redefining Value Networks and Enabling Meaningful Experiences (UNITE)\u201d","doi-asserted-by":"publisher","award":["336145"],"award-info":[{"award-number":["336145"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cForest-Human\u2013Machine Interplay\u2014Building Resilience, Redefining Value Networks and Enabling Meaningful Experiences (UNITE)\u201d","doi-asserted-by":"publisher","award":["353363"],"award-info":[{"award-number":["353363"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cForest-Human\u2013Machine Interplay\u2014Building Resilience, Redefining Value Networks and Enabling Meaningful Experiences (UNITE)\u201d","doi-asserted-by":"publisher","award":["337656"],"award-info":[{"award-number":["337656"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cForest-Human\u2013Machine Interplay\u2014Building Resilience, Redefining Value Networks and Enabling Meaningful Experiences (UNITE)\u201d","doi-asserted-by":"publisher","award":["314312"],"award-info":[{"award-number":["314312"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cForest-Human\u2013Machine Interplay\u2014Building Resilience, Redefining Value Networks and Enabling Meaningful Experiences (UNITE)\u201d","doi-asserted-by":"publisher","award":["343678"],"award-info":[{"award-number":["343678"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cForest-Human\u2013Machine Interplay\u2014Building Resilience, Redefining Value Networks and Enabling Meaningful Experiences (UNITE)\u201d","doi-asserted-by":"publisher","award":["181811KYSB20160040"],"award-info":[{"award-number":["181811KYSB20160040"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cForest-Human\u2013Machine Interplay\u2014Building Resilience, Redefining Value Networks and Enabling Meaningful Experiences (UNITE)\u201d","doi-asserted-by":"publisher","award":["XDA22030202"],"award-info":[{"award-number":["XDA22030202"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"\u201cForest-Human\u2013Machine Interplay\u2014Building Resilience, Redefining Value Networks and Enabling Meaningful Experiences (UNITE)\u201d","doi-asserted-by":"publisher","award":["1003-YQR23046"],"award-info":[{"award-number":["1003-YQR23046"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100009047","name":"Strategic Research Council project \u201cCompetence-Based Growth Through Integrated Disruptive Technologies of 3D Digitalization, Robotics, Geospatial Information and Image Processing\/Computing\u2014Point Cloud Ecosystem\u201d","doi-asserted-by":"publisher","award":["336145"],"award-info":[{"award-number":["336145"]}],"id":[{"id":"10.13039\/501100009047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100009047","name":"Strategic Research Council project \u201cCompetence-Based Growth Through Integrated Disruptive Technologies of 3D Digitalization, Robotics, Geospatial Information and Image Processing\/Computing\u2014Point Cloud Ecosystem\u201d","doi-asserted-by":"publisher","award":["353363"],"award-info":[{"award-number":["353363"]}],"id":[{"id":"10.13039\/501100009047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100009047","name":"Strategic Research Council project \u201cCompetence-Based Growth Through Integrated Disruptive Technologies of 3D Digitalization, Robotics, Geospatial Information and Image Processing\/Computing\u2014Point Cloud Ecosystem\u201d","doi-asserted-by":"publisher","award":["337656"],"award-info":[{"award-number":["337656"]}],"id":[{"id":"10.13039\/501100009047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100009047","name":"Strategic Research Council project \u201cCompetence-Based Growth Through Integrated Disruptive Technologies of 3D Digitalization, Robotics, Geospatial Information and Image Processing\/Computing\u2014Point Cloud Ecosystem\u201d","doi-asserted-by":"publisher","award":["314312"],"award-info":[{"award-number":["314312"]}],"id":[{"id":"10.13039\/501100009047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100009047","name":"Strategic Research Council project \u201cCompetence-Based Growth Through Integrated Disruptive Technologies of 3D Digitalization, Robotics, Geospatial Information and Image Processing\/Computing\u2014Point Cloud Ecosystem\u201d","doi-asserted-by":"publisher","award":["343678"],"award-info":[{"award-number":["343678"]}],"id":[{"id":"10.13039\/501100009047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100009047","name":"Strategic Research Council project \u201cCompetence-Based Growth Through Integrated Disruptive Technologies of 3D Digitalization, Robotics, Geospatial Information and Image Processing\/Computing\u2014Point Cloud Ecosystem\u201d","doi-asserted-by":"publisher","award":["181811KYSB20160040"],"award-info":[{"award-number":["181811KYSB20160040"]}],"id":[{"id":"10.13039\/501100009047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100009047","name":"Strategic Research Council project \u201cCompetence-Based Growth Through Integrated Disruptive Technologies of 3D Digitalization, Robotics, Geospatial Information and Image Processing\/Computing\u2014Point Cloud Ecosystem\u201d","doi-asserted-by":"publisher","award":["XDA22030202"],"award-info":[{"award-number":["XDA22030202"]}],"id":[{"id":"10.13039\/501100009047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100009047","name":"Strategic Research Council project \u201cCompetence-Based Growth Through Integrated Disruptive Technologies of 3D Digitalization, Robotics, Geospatial Information and Image Processing\/Computing\u2014Point Cloud Ecosystem\u201d","doi-asserted-by":"publisher","award":["1003-YQR23046"],"award-info":[{"award-number":["1003-YQR23046"]}],"id":[{"id":"10.13039\/501100009047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["336145"],"award-info":[{"award-number":["336145"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["353363"],"award-info":[{"award-number":["353363"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["337656"],"award-info":[{"award-number":["337656"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["314312"],"award-info":[{"award-number":["314312"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["343678"],"award-info":[{"award-number":["343678"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["181811KYSB20160040"],"award-info":[{"award-number":["181811KYSB20160040"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["XDA22030202"],"award-info":[{"award-number":["XDA22030202"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["1003-YQR23046"],"award-info":[{"award-number":["1003-YQR23046"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Chinese Academy of Science","award":["336145"],"award-info":[{"award-number":["336145"]}]},{"name":"Chinese Academy of Science","award":["353363"],"award-info":[{"award-number":["353363"]}]},{"name":"Chinese Academy of Science","award":["337656"],"award-info":[{"award-number":["337656"]}]},{"name":"Chinese Academy of Science","award":["314312"],"award-info":[{"award-number":["314312"]}]},{"name":"Chinese Academy of Science","award":["343678"],"award-info":[{"award-number":["343678"]}]},{"name":"Chinese Academy of Science","award":["181811KYSB20160040"],"award-info":[{"award-number":["181811KYSB20160040"]}]},{"name":"Chinese Academy of Science","award":["XDA22030202"],"award-info":[{"award-number":["XDA22030202"]}]},{"name":"Chinese Academy of Science","award":["1003-YQR23046"],"award-info":[{"award-number":["1003-YQR23046"]}]},{"name":"Nanjing University of Aeronautics and Astronautics Startup Funding for Yong Talents","award":["336145"],"award-info":[{"award-number":["336145"]}]},{"name":"Nanjing University of Aeronautics and Astronautics Startup Funding for Yong Talents","award":["353363"],"award-info":[{"award-number":["353363"]}]},{"name":"Nanjing University of Aeronautics and Astronautics Startup Funding for Yong Talents","award":["337656"],"award-info":[{"award-number":["337656"]}]},{"name":"Nanjing University of Aeronautics and Astronautics Startup Funding for Yong Talents","award":["314312"],"award-info":[{"award-number":["314312"]}]},{"name":"Nanjing University of Aeronautics and Astronautics Startup Funding for Yong Talents","award":["343678"],"award-info":[{"award-number":["343678"]}]},{"name":"Nanjing University of Aeronautics and Astronautics Startup Funding for Yong Talents","award":["181811KYSB20160040"],"award-info":[{"award-number":["181811KYSB20160040"]}]},{"name":"Nanjing University of Aeronautics and Astronautics Startup Funding for Yong Talents","award":["XDA22030202"],"award-info":[{"award-number":["XDA22030202"]}]},{"name":"Nanjing University of Aeronautics and Astronautics Startup Funding for Yong Talents","award":["1003-YQR23046"],"award-info":[{"award-number":["1003-YQR23046"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Indoor positioning plays a crucial role in various domains. It is employed in various applications, such as navigation, asset tracking, and location-based services (LBS), in Global Navigation Satellite System (GNSS) denied or degraded areas. The visual-based positioning technique is a promising solution for high-accuracy indoor positioning. However, most visual positioning research uses the side-view perspective, which is susceptible to interferences and may cause concerns about privacy and public security. Therefore, this paper innovatively proposes an up-view visual-based indoor positioning algorithm. It uses the up-view images to realize indoor positioning. Firstly, we utilize a well-trained YOLO V7 model to realize landmark detection and gross extraction. Then, we use edge detection operators to realize the precision landmark extraction, obtaining the landmark pixel size. The target position is calculated based on the landmark detection and extraction results and the pre-labeled landmark sequence via the Similar Triangle Principle. Additionally, we also propose an inertial navigation system (INS)-based landmark matching method to match the landmark within an up-view image with a landmark in the pre-labeled landmark sequence. This is necessary for kinematic indoor positioning. Finally, we conduct static and kinematic experiments to verify the feasibility and performance of the up-view-based indoor positioning method. The results demonstrate that the up-view visual-based positioning is prospective and worthy of research.<\/jats:p>","DOI":"10.3390\/rs16061024","type":"journal-article","created":{"date-parts":[[2024,3,14]],"date-time":"2024-03-14T05:43:08Z","timestamp":1710394988000},"page":"1024","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["An Up-View Visual-Based Indoor Positioning Method via Deep Learning"],"prefix":"10.3390","volume":"16","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-6836-8901","authenticated-orcid":false,"given":"Chen","family":"Chen","sequence":"first","affiliation":[{"name":"Department of Remote Sensing and Photogrammetry, Finnish Geospatial Research Institute, FI-02150 Espoo, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0148-3609","authenticated-orcid":false,"given":"Yuwei","family":"Chen","sequence":"additional","affiliation":[{"name":"Department of Remote Sensing and Photogrammetry, Finnish Geospatial Research Institute, FI-02150 Espoo, Finland"}]},{"given":"Jianliang","family":"Zhu","sequence":"additional","affiliation":[{"name":"School of Automation, Nanjing University of Science and Technology, Nanjing 210094, China"}]},{"given":"Changhui","family":"Jiang","sequence":"additional","affiliation":[{"name":"Flight Control Institute, Nanjing University of Aeronautics and Astronautics, Nanjing 210095, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4366-4547","authenticated-orcid":false,"given":"Jianxin","family":"Jia","sequence":"additional","affiliation":[{"name":"Department of Remote Sensing and Photogrammetry, Finnish Geospatial Research Institute, FI-02150 Espoo, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0228-6749","authenticated-orcid":false,"given":"Yuming","family":"Bo","sequence":"additional","affiliation":[{"name":"School of Automation, Nanjing University of Science and Technology, Nanjing 210094, China"}]},{"given":"Xuanzhi","family":"Liu","sequence":"additional","affiliation":[{"name":"Department of Remote Sensing and Photogrammetry, Finnish Geospatial Research Institute, FI-02150 Espoo, Finland"}]},{"given":"Haojie","family":"Dai","sequence":"additional","affiliation":[{"name":"School of Automation, Nanjing University of Science and Technology, Nanjing 210094, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0985-4443","authenticated-orcid":false,"given":"Eetu","family":"Puttonen","sequence":"additional","affiliation":[{"name":"Department of Remote Sensing and Photogrammetry, Finnish Geospatial Research Institute, FI-02150 Espoo, Finland"}]},{"given":"Juha","family":"Hyypp\u00e4","sequence":"additional","affiliation":[{"name":"Department of Remote Sensing and Photogrammetry, Finnish Geospatial Research Institute, FI-02150 Espoo, Finland"}]}],"member":"1968","published-online":{"date-parts":[[2024,3,14]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Werner, M., Kessel, M., and Marouane, C. (2011, January 21\u201323). Indoor positioning using smartphone camera. Proceedings of the 2011 International Conference on Indoor Positioning and Indoor Navigation, Guimaraes, Portugal.","DOI":"10.1109\/IPIN.2011.6071954"},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Rudi\u0107, B., Klaffenb\u00f6ck, M.A., Pichler-Scheder, M., Efrosinin, D., and Kastl, C. (2020, January 23). Geometry-aided ble-based smartphone positioning for indoor location-based services. Proceedings of the 2020 IEEE MTT-S International Conference on Microwaves for Intelligent Mobility (ICMIM), Linz, Austria.","DOI":"10.1109\/ICMIM48759.2020.9299009"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"1347","DOI":"10.1109\/COMST.2016.2637663","article-title":"A survey of selected indoor positioning methods for smartphones","volume":"19","author":"Davidson","year":"2016","journal-title":"IEEE Commun. Surv. Tutor."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Quezada-Gaibor, D., Torres-Sospedra, J., Nurmi, J., Koucheryavy, Y., and Huerta, J. (2021). Cloud platforms for context-adaptive positioning and localisation in GNSS-denied scenarios\u2014A systematic review. Sensors, 22.","DOI":"10.3390\/s22010110"},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"166","DOI":"10.1016\/j.inffus.2022.08.013","article-title":"Survey of landmark-based indoor positioning technologies","volume":"89","author":"Jang","year":"2023","journal-title":"Inf. Fusion"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"7","DOI":"10.1186\/s43020-021-00041-3","article-title":"Indoor navigation: State of the art and future trends","volume":"2","author":"Li","year":"2021","journal-title":"Satell. Navig."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"466","DOI":"10.1109\/COMST.2015.2464084","article-title":"Wi-Fi fingerprint-based indoor positioning: Recent advances and comparisons","volume":"18","author":"He","year":"2015","journal-title":"IEEE Commun. Surv. Tutor."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"79","DOI":"10.1007\/s10291-019-0870-y","article-title":"Performance analysis of indoor pseudolite positioning based on the unscented Kalman filter","volume":"23","author":"Li","year":"2019","journal-title":"GPS Solut."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"3971","DOI":"10.1109\/JSEN.2020.2964380","article-title":"Light-based indoor positioning systems: A review","volume":"20","author":"Maheepala","year":"2020","journal-title":"IEEE Sens. J."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"122428","DOI":"10.1109\/ACCESS.2019.2937464","article-title":"A novel clustering algorithm for Wi-Fi indoor positioning","volume":"7","author":"Ren","year":"2019","journal-title":"IEEE Access"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Cheng, Y., and Zhou, T. (2019, January 23\u201325). UWB indoor positioning algorithm based on TDOA technology. Proceedings of the 2019 10th International Conference on Information Technology in Medicine and Education (ITME), Qingdao, China.","DOI":"10.1109\/ITME.2019.00177"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Phutcharoen, K., Chamchoy, M., and Supanakoon, P. (2020, January 11\u201314). Accuracy study of indoor positioning with bluetooth low energy beacons. Proceedings of the 2020 Joint International Conference on Digital Arts, Media and Technology with ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI DAMT & NCON), Pattaya, Thailand.","DOI":"10.1109\/ECTIDAMTNCON48261.2020.9090691"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Yao, L., Wu, Y.W.A., Yao, L., and Liao, Z.Z. (2017, January 18\u201321). An integrated IMU and UWB sensor based indoor positioning system. Proceedings of the 2017 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Sapporo, Japan.","DOI":"10.1109\/IPIN.2017.8115911"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Cioffi, G., and Scaramuzza, D. (2020\u201324, January 24). Tightly-coupled fusion of global positional measurements in optimization-based visual-inertial odometry. Proceedings of the 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA.","DOI":"10.1109\/IROS45743.2020.9341697"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"6907","DOI":"10.1109\/TITS.2021.3063477","article-title":"A comparative analysis of LiDAR SLAM-based indoor navigation for autonomous vehicles","volume":"23","author":"Zou","year":"2021","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Li, M., Chen, R., Liao, X., Guo, B., Zhang, W., and Guo, G. (2020). A precise indoor visual positioning approach using a built image feature database and single user image from smartphone cameras. Remote Sens., 12.","DOI":"10.3390\/rs12050869"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"289","DOI":"10.1109\/TRO.2017.2788045","article-title":"Robust visual localization across seasons","volume":"34","author":"Naseer","year":"2018","journal-title":"IEEE Trans. Robot."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"103666","DOI":"10.1016\/j.robot.2020.103666","article-title":"A review on absolute visual localization for UAV","volume":"135","author":"Couturier","year":"2021","journal-title":"Robot. Auton. Syst."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"32","DOI":"10.1007\/s00138-020-01161-7","article-title":"Visual localization and servoing for drone use in indoor remote laboratory environment","volume":"32","author":"Khattar","year":"2021","journal-title":"Mach. Vis. Appl."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Lei, X., Zhang, F., Zhou, J., and Shang, W. (2022, January 7\u201310). Visual Localization Strategy for Indoor Mobile Robots in the Complex Environment. Proceedings of the 2022 IEEE International Conference on Mechatronics and Automation (ICMA), Guilin, China.","DOI":"10.1109\/ICMA54519.2022.9856360"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Zhang, H., and Ye, C. (August, January 31). A visual positioning system for indoor blind navigation. Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France.","DOI":"10.1109\/ICRA40945.2020.9196782"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Himawan, R.W., Baylon, P.B.A., Sembiring, J., and Jenie, Y.I. (2023, January 26\u201327). Development of an Indoor Visual-Based Monocular Positioning System for Multirotor UAV. Proceedings of the 2023 IEEE International Conference on Aerospace Electronics and Remote Sensing Technology (ICARES), Bali, Indonesia.","DOI":"10.1109\/ICARES60489.2023.10329792"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"20055","DOI":"10.1109\/JSEN.2023.3297232","article-title":"Low-latency Visual-based High-Quality 3D Reconstruction using Point Cloud Optimization","volume":"23","author":"Chi","year":"2023","journal-title":"IEEE Sens. J."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Bai, X., Zhang, B., Wen, W., Hsu, L.T., and Li, H. (2020, January 20\u201323). Perception-aided visual-inertial integrated positioning in dynamic urban areas. Proceedings of the 2020 IEEE\/ION Position, Location and Navigation Symposium (PLANS), Portland, OR, USA.","DOI":"10.1109\/PLANS46316.2020.9109963"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016, January 27\u201330). You only look once: Unified, real-time object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Cao, L., Zheng, X., and Fang, L. (2023). The Semantic Segmentation of Standing Tree Images Based on the Yolo V7 Deep Learning Algorithm. Electronics, 12.","DOI":"10.3390\/electronics12040929"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/6\/1024\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T14:13:20Z","timestamp":1760105600000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/6\/1024"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,14]]},"references-count":26,"journal-issue":{"issue":"6","published-online":{"date-parts":[[2024,3]]}},"alternative-id":["rs16061024"],"URL":"https:\/\/doi.org\/10.3390\/rs16061024","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,3,14]]}}}