{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,14]],"date-time":"2026-01-14T17:03:08Z","timestamp":1768410188332,"version":"3.49.0"},"reference-count":24,"publisher":"MDPI AG","issue":"2","license":[{"start":{"date-parts":[[2023,2,20]],"date-time":"2023-02-20T00:00:00Z","timestamp":1676851200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Data"],"abstract":"<jats:p>Computer vision is a new approach to navigation aiding that assists visually impaired people to travel independently. A deep learning-based solution implemented on a portable device that uses a monocular camera to capture public objects could be a low-cost and handy navigation aid. By recognizing public objects in the street and estimating their distance from the user, visually impaired people are able to avoid obstacles in the outdoor environment and walk safely. In this paper, we created a dataset of public objects in an uncontrolled environment for navigation aiding. The dataset contains three classes of objects which commonly exist on pavements in the city. It was verified that the dataset was of high quality for object detection and distance estimation, and was ultimately utilized as a navigation aid solution.<\/jats:p>","DOI":"10.3390\/data8020042","type":"journal-article","created":{"date-parts":[[2023,2,20]],"date-time":"2023-02-20T04:28:57Z","timestamp":1676867337000},"page":"42","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Dataset of Public Objects in Uncontrolled Environment for Navigation Aiding"],"prefix":"10.3390","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1980-8590","authenticated-orcid":false,"given":"Teng-Lai","family":"Wong","sequence":"first","affiliation":[{"name":"Faculty of Applied Sciences, Macao Polytechnic University, Macao SAR, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8684-4866","authenticated-orcid":false,"given":"Ka-Seng","family":"Chou","sequence":"additional","affiliation":[{"name":"Faculty of Applied Sciences, Macao Polytechnic University, Macao SAR, China"},{"name":"Department of Computer Science and Engineering, Alma Mater Studiorum, University of Bologna, 47521 Bologna, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4566-3575","authenticated-orcid":false,"given":"Kei-Long","family":"Wong","sequence":"additional","affiliation":[{"name":"Faculty of Applied Sciences, Macao Polytechnic University, Macao SAR, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8104-7887","authenticated-orcid":false,"given":"Su-Kit","family":"Tang","sequence":"additional","affiliation":[{"name":"Faculty of Applied Sciences, Macao Polytechnic University, Macao SAR, China"}]}],"member":"1968","published-online":{"date-parts":[[2023,2,20]]},"reference":[{"key":"ref_1","first-page":"2317","article-title":"Global Prevalence of Blindness and Distance and Near Vision Impairment in 2020: Progress towards the Vision 2020 Targets and What the Future Holds","volume":"61","author":"Bourne","year":"2020","journal-title":"Investig. Ophthalmol. Vis. Sci."},{"key":"ref_2","first-page":"2088","article-title":"An assistive model of obstacle detection based on deep learning: YOLOv3 for visually impaired people","volume":"11","author":"Rachburee","year":"2021","journal-title":"Int. J. Electr. Comput. Eng."},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Pehlivan, S., Unay, M., and Akan, A. (2019, January 3\u20135). Designing an obstacle detection and alerting system for visually impaired people on sidewalks. Proceedings of the 2019 Medical Technologies Congress (TIPTEKNO), Izmir, Turkey.","DOI":"10.1109\/TIPTEKNO.2019.8895181"},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Yadav, S., Joshi, R., Dutta, M., Kiac, M., and Sikora, P. (2020, January 7\u20139). Fusion of object recognition and obstacle detection approach for assisting visually challenged person. Proceedings of the 2020 43rd International Conference On Telecommunications And Signal Processing (TSP), Milan, Italy.","DOI":"10.1109\/TSP49548.2020.9163434"},{"key":"ref_5","first-page":"219","article-title":"Smartphone application program of obstacle detection for visually impaired people","volume":"10","author":"Shimakawa","year":"2019","journal-title":"ICIC Express Lett. Part B Appl. Int. J. Res. Surv."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Hussain, S., Durrani, D., Khan, A., Atta, R., and Ahmed, L. (November, January 29). In-door obstacle detection and avoidance system for visually impaired people. Proceedings of the 2020 IEEE Global Humanitarian Technology Conference (GHTC), Seattle, WA, USA.","DOI":"10.1109\/GHTC46280.2020.9342942"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"13428","DOI":"10.1109\/ACCESS.2022.3146320","article-title":"Smart assistive system for visually impaired people obstruction avoidance through object detection and classification","volume":"10","author":"Masud","year":"2022","journal-title":"IEEE Access"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Tse, R., Mirri, S., Tang, S.-K., Pau, G., and Salomoni, P. (2020, January 14\u201316). Building an Italian-Chinese Parallel Corpus for Machine Translation from the Web. Proceedings of the 6th EAI International Conference on Smart Objects and Technologies for Social Good, Antwerp, Belgium.","DOI":"10.1145\/3411170.3411258"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"122","DOI":"10.1016\/j.procs.2021.12.218","article-title":"Crawling Parallel Data for Bilingual Corpus Using Hybrid Crawling Architecture","volume":"198","author":"Cheok","year":"2022","journal-title":"Procedia Comput. Sci."},{"key":"ref_10","first-page":"100134","article-title":"Deep Learning in Computer Vision: A Critical Review of Emerging Techniques and Application Scenarios","volume":"6","author":"Chai","year":"2021","journal-title":"Mach. Learn. Appl."},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., and Malik, J. (2014, January 23\u201328). Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA.","DOI":"10.1109\/CVPR.2014.81"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Girshick, R. (2015, January 7\u201313). Fast R-CNN. Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile.","DOI":"10.1109\/ICCV.2015.169"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","article-title":"Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks","volume":"39","author":"Ren","year":"2017","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Li, M., Zhang, Z., Lei, L., Wang, X., and Guo, X. (2020). Agricultural Greenhouses Detection in High-Resolution Satellite Images Based on Convolutional Neural Networks: Comparison of Faster R-CNN, YOLO v3 and SSD. Sensors, 20.","DOI":"10.3390\/s20174938"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"1577614","DOI":"10.1155\/2021\/1577614","article-title":"A Real-Time Vehicle Counting, Speed Estimation, and Classification System Based on Virtual Detection Zone and YOLO","volume":"2021","author":"Lin","year":"2021","journal-title":"Math. Probl. Eng."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016, January 27\u201330). You Only Look Once: Unified, Real-Time Object Detection. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Redmon, J., and Farhadi, A. (2017, January 21\u201326). YOLO9000: Better, Faster, Stronger. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.690"},{"key":"ref_18","unstructured":"Redmon, J., and Farhadi, A. (2018). YOLOv3: An Incremental Improvement. arXiv."},{"key":"ref_19","unstructured":"Bochkovskiy, A., Wang, C.-Y., and Liao, H.-Y.M. (2020). YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Hiemann, A., Kautz, T., Zottmann, T., and Hlawitschka, M. (2021). Enhancement of Speed and Accuracy Trade-off for Sports Ball Detection in Videos-Finding Fast Moving, Small Objects in Real Time. Sensors, 21.","DOI":"10.3390\/s21093214"},{"key":"ref_21","unstructured":"(2022, March 28). Programming Comments-Darknet FAQ. Available online: https:\/\/www.ccoderun.ca\/programming\/darknet_faq\/."},{"key":"ref_22","unstructured":"(2021, October 24). labelImg: LabelImg Is a Graphical Image Annotation Tool and Label Object Bounding Boxes in Images. Available online: https:\/\/github.com\/tzutalin\/labelImg."},{"key":"ref_23","unstructured":"Mohammadian, M. (2004, January 12\u201314). Monocular Vision as a Range Sensor. Proceedings of the International Conference on Computational Intelligence for Modeling, Control and Automation, Gold Coast, Australia."},{"key":"ref_24","unstructured":"Stein, G.P., Mano, O., and Shashua, A. (2003, January 9\u201311). Vision-Based ACC with a Single Camera: Bounds on Range and Range Rate Accuracy. Proceedings of the IEEE IV2003 Intelligent Vehicles Symposium, Columbus, OH, USA. Proceedings (Cat. No.03TH8683)."}],"container-title":["Data"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2306-5729\/8\/2\/42\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T18:37:04Z","timestamp":1760121424000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2306-5729\/8\/2\/42"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,2,20]]},"references-count":24,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2023,2]]}},"alternative-id":["data8020042"],"URL":"https:\/\/doi.org\/10.3390\/data8020042","relation":{},"ISSN":["2306-5729"],"issn-type":[{"value":"2306-5729","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,2,20]]}}}