{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T14:52:46Z","timestamp":1770994366841,"version":"3.50.1"},"reference-count":28,"publisher":"MDPI AG","issue":"1","license":[{"start":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T00:00:00Z","timestamp":1736985600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Ministry of Education, Culture, Sports, Science, and Technology of Japan (MEXT)"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Informatics"],"abstract":"<jats:p>Advancements in automation and artificial intelligence have significantly impacted accessibility for individuals with visual impairments, particularly in the realm of bus public transportation. Effective bus detection and bus point-of-view (POV) classification are crucial for enhancing the independence of visually impaired individuals. This study introduces the Improved-YOLOv10, a novel model designed to tackle challenges in bus identification and pov classification by integrating Coordinate Attention (CA) and Adaptive Kernel Convolution (AKConv) into the YOLOv10 framework. The Improved YOLOv10 advances the YOLOv10 architecture through the incorporation of CA, which enhances long-range dependency modeling and spatial awareness, and AKConv, which dynamically adjusts convolutional kernels for superior feature extraction. These enhancements aim to improve both detection accuracy and efficiency, essential for real-time applications in assistive technologies. Evaluation results demonstrate that the Improved-YOLOv10 offers significant improvements in detection performance, including better Accuracy, Precision and Recall compared to YOLOv10. The model also exhibits reduced computational complexity and storage requirements, highlighting its efficiency. While the classification results show some trade-offs, with slightly decreased overall F1 score, the complexity of Giga Floating Point Operations (GFLOPs), Parameters, and Weight\/MB in the Improved-YOLOv10 remains advantageous for classification tasks. The model\u2019s architectural improvements contribute to its robustness and efficiency, making it a suitable choice for real-time applications and assistive technologies.<\/jats:p>","DOI":"10.3390\/informatics12010007","type":"journal-article","created":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T06:46:10Z","timestamp":1737009970000},"page":"7","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["Improved YOLOv10 for Visually Impaired: Balancing Model Accuracy and Efficiency in the Case of Public Transportation"],"prefix":"10.3390","volume":"12","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1730-9963","authenticated-orcid":false,"given":"Rio","family":"Arifando","sequence":"first","affiliation":[{"name":"Graduate School of Life Science and Systems Engineering, Kyushu Institute of Technology, 2\u20134 Hibikino, Wakamatsu-ku, Kitakyushu 808-0196, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7684-1936","authenticated-orcid":false,"given":"Shinji","family":"Eto","sequence":"additional","affiliation":[{"name":"Graduate School of Life Science and Systems Engineering, Kyushu Institute of Technology, 2\u20134 Hibikino, Wakamatsu-ku, Kitakyushu 808-0196, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9399-4160","authenticated-orcid":false,"given":"Tibyani","family":"Tibyani","sequence":"additional","affiliation":[{"name":"Department of Information Systems, Faculty of Computer Science, Brawijaya University, Malang 65145, Indonesia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8349-7141","authenticated-orcid":false,"given":"Chikamune","family":"Wada","sequence":"additional","affiliation":[{"name":"Graduate School of Life Science and Systems Engineering, Kyushu Institute of Technology, 2\u20134 Hibikino, Wakamatsu-ku, Kitakyushu 808-0196, Japan"}]}],"member":"1968","published-online":{"date-parts":[[2025,1,16]]},"reference":[{"key":"ref_1","unstructured":"World Health Organization (2019). World Report on Vision, World Health Organization."},{"key":"ref_2","unstructured":"Venter, C., Bogopane, H., Rickert, T., Camba, J., Venkatesh, A., Mulikita, N., Maunder, D., Savill, T., and Stone, J. (2002, January 15\u201318). Improving Accessibility for People with Disabilities in Urban Areas: Am\u00e9liorer l\u2019accessibilit\u00e9 du milieu urbain aux personnes handicap\u00e9es. Proceedings of the International Conference on Urban Transport and the Environment for the 21st Century, Transport Research Laboratory, Pretoria, South Africa."},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Hersh, M., and Johnson, M. (2008). Assistive Technology for Visually Impaired and Blind People, Springer.","DOI":"10.1007\/978-1-84628-867-8"},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Anu, V.M., Sarikha, D., Keerthy, G.S., and Jabez, J. (2015, January 19\u201320). An RFID based system for bus location tracking and display. Proceedings of the International Confernce on Innovation Information in Computing Technologies, Chennai, India.","DOI":"10.1109\/ICIICT.2015.7396080"},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"012146","DOI":"10.1088\/1757-899X\/114\/1\/012146","article-title":"RFID-based automatic bus ticketing: Features and trends","volume":"114","author":"Oudah","year":"2016","journal-title":"IOP Conf. Ser. Mater. Sci. Eng."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Vinod, D., and Mohan, A. (2018, January 28\u201331). A Successful Approach to Bus Tracking Using RFID and Low Power Wireless Networks. Proceedings of the TENCON 2018\u20142018 IEEE Region 10 Conference, Jeju, Republic of Korea.","DOI":"10.1109\/TENCON.2018.8650342"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"7797","DOI":"10.3390\/s130607797","article-title":"Performance Evaluation of UHF RFID Technologies for Real-Time Bus Recognition in the Taipei Bus Station","volume":"13","author":"Own","year":"2013","journal-title":"Sensors"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"43","DOI":"10.31871\/IJNTR.4.9.16","article-title":"Classification of Viewpoints Related to Bus-Waiting for the Assistance of Blind People","volume":"4","author":"Tangsuksant","year":"2018","journal-title":"Int. J. New Technol. Res."},{"key":"ref_9","unstructured":"Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J., and Ding, G. (2024). YOLOv10: Real-Time End-to-End Object Detection. arXiv."},{"key":"ref_10","unstructured":"Tan, M., and Le, Q.V. (2019). EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv."},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., and Weinberger, K.Q. (2016). Densely Connected Convolutional Networks. arXiv.","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref_12","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., and Adam, H. (2017). MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S.E., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2014). Going Deeper with Convolutions. arXiv.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Chollet, F. (2016). Xception: Deep Learning with Depthwise Separable Convolutions. arXiv.","DOI":"10.1109\/CVPR.2017.195"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Albanie, S., Sun, G., and Wu, E. (2019). Squeeze-and-Excitation Networks. arXiv.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J., and Kweon, I.S. (2018). CBAM: Convolutional Block Attention Module. arXiv.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., and Feng, J. (2021). Coordinate Attention for Efficient Mobile Network Design. arXiv.","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"ref_18","unstructured":"Gennari, M., Fawcett, R., and Prisacariu, V.A. (2019). DSConv: Efficient Convolution Operator. arXiv."},{"key":"ref_19","unstructured":"Zhang, X., Song, Y., Song, T., Yang, D., Ye, Y., Zhou, J., and Zhang, L. (2023). AKConv: Convolutional Kernel with Arbitrary Sampled Shapes and Arbitrary Number of Parameters. arXiv."},{"key":"ref_20","unstructured":"Dai, J., Li, Y., He, K., and Sun, J. (2016). R-FCN: Object Detection via Region-based Fully Convolutional Networks. arXiv."},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., and Girshick, R.B. (2017). Mask R-CNN. arXiv.","DOI":"10.1109\/ICCV.2017.322"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Cai, Z., and Vasconcelos, N. (2017). Cascade R-CNN: Delving into High Quality Object Detection. arXiv.","DOI":"10.1109\/CVPR.2018.00644"},{"key":"ref_23","unstructured":"Bochkovskiy, A., Wang, C., and Liao, H.M. (2020). YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., and Le, Q.V. (2019). EfficientDet: Scalable and Efficient Object Detection. arXiv.","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Duan, K., Bai, S., Xie, L., Qi, H., Huang, Q., and Tian, Q. (2019). CenterNet: Keypoint Triplets for Object Detection. arXiv.","DOI":"10.1109\/ICCV.2019.00667"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Liao, H.Y.M., Yeh, I.H., Wu, Y.H., Chen, P.Y., and Hsieh, J.W. (2019). CSPNet: A New Backbone that can Enhance Learning Capability of CNN. arXiv.","DOI":"10.1109\/CVPRW50498.2020.00203"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Liu, S., Qi, L., Qin, H., Shi, J., and Jia, J. (2018). Path Aggregation Network for Instance Segmentation. arXiv.","DOI":"10.1109\/CVPR.2018.00913"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., and Hu, Q. (2019). ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. arXiv.","DOI":"10.1109\/CVPR42600.2020.01155"}],"container-title":["Informatics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2227-9709\/12\/1\/7\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,8]],"date-time":"2025-10-08T10:29:46Z","timestamp":1759919386000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2227-9709\/12\/1\/7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,16]]},"references-count":28,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,3]]}},"alternative-id":["informatics12010007"],"URL":"https:\/\/doi.org\/10.3390\/informatics12010007","relation":{},"ISSN":["2227-9709"],"issn-type":[{"value":"2227-9709","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,16]]}}}