{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T22:29:26Z","timestamp":1776119366769,"version":"3.50.1"},"reference-count":51,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2023,2,2]],"date-time":"2023-02-02T00:00:00Z","timestamp":1675296000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001871","name":"FCT\u2014Fundac\u0327\u00e3o para a Ci\u00eancia e Tecnologia","doi-asserted-by":"publisher","award":["UIDB\/00319\/2020"],"award-info":[{"award-number":["UIDB\/00319\/2020"]}],"id":[{"id":"10.13039\/501100001871","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001871","name":"FCT\u2014Fundac\u0327\u00e3o para a Ci\u00eancia e Tecnologia","doi-asserted-by":"publisher","award":["UIDB\/05549\/2020"],"award-info":[{"award-number":["UIDB\/05549\/2020"]}],"id":[{"id":"10.13039\/501100001871","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001871","name":"FCT\u2014Fundac\u0327\u00e3o para a Ci\u00eancia e Tecnologia","doi-asserted-by":"publisher","award":["UIDP\/05549\/2020"],"award-info":[{"award-number":["UIDP\/05549\/2020"]}],"id":[{"id":"10.13039\/501100001871","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Applied Sciences"],"abstract":"<jats:p>The ways in which people dress, as well as the styles that they prefer for different contexts and occasions, are part of their identity. Every day, blind people face limitations in identifying and inspecting their garments, and dressing can be a difficult and stressful task. Taking advantage of the great technological advancements, it becomes of the utmost importance to minimize, as much as possible, the limitations of a blind person when choosing garments. Hence, this work aimed at categorizing and detecting the presence of stains on garments, using artificial intelligence algorithms. In our approach, transfer learning was used for category classification, where a benchmark was performed between convolutional neural networks (CNNs), with the best model achieving an F1 score of 91%. Stain detection was performed through the fine tuning of a deep learning object detector, i.e., the mask R (region-based)-CNN. This approach is also analyzed and discussed, as it allowed us to achieve better results than those available in the literature.<\/jats:p>","DOI":"10.3390\/app13031925","type":"journal-article","created":{"date-parts":[[2023,2,2]],"date-time":"2023-02-02T01:53:54Z","timestamp":1675302834000},"page":"1925","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Blind People: Clothing Category Classification and Stain Detection Using Transfer Learning"],"prefix":"10.3390","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7778-0912","authenticated-orcid":false,"given":"Daniel","family":"Rocha","sequence":"first","affiliation":[{"name":"Algoritmi Research Centre\/LASI, University of Minho, 4800-058 Guimar\u00e3es, Portugal"},{"name":"2Ai, School of Technology, Polytechnic Institute of C\u00e1vado and Ave, 4750-810 Barcelos, Portugal"},{"name":"INL\u2014International Nanotechnology Laboratory, 4715-330 Braga, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4438-6713","authenticated-orcid":false,"given":"Filomena","family":"Soares","sequence":"additional","affiliation":[{"name":"Algoritmi Research Centre\/LASI, University of Minho, 4800-058 Guimar\u00e3es, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8394-7088","authenticated-orcid":false,"given":"Eva","family":"Oliveira","sequence":"additional","affiliation":[{"name":"2Ai, School of Technology, Polytechnic Institute of C\u00e1vado and Ave, 4750-810 Barcelos, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4658-5844","authenticated-orcid":false,"given":"V\u00edtor","family":"Carvalho","sequence":"additional","affiliation":[{"name":"Algoritmi Research Centre\/LASI, University of Minho, 4800-058 Guimar\u00e3es, Portugal"},{"name":"2Ai, School of Technology, Polytechnic Institute of C\u00e1vado and Ave, 4750-810 Barcelos, Portugal"}]}],"member":"1968","published-online":{"date-parts":[[2023,2,2]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Wade, N.J., and Swanston, M. (2013). Visual Perception, Psychology Press. [3rd ed.].","DOI":"10.4324\/9780203082263"},{"key":"ref_2","unstructured":"GBD 2019 Blindness and Vision Impairment Collaborators (2021). Causes of blindness and vision impairment in 2020 and trends over 30 years, and prevalence of avoidable blindness in relation to VISION 2020: The Right to Sight: An analysis for the Global Burden of Disease Study. Lancet. Glob. Health, 9, e144\u2013e160."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"149","DOI":"10.1007\/s12193-016-0235-6","article-title":"An insight into assistive technology for the visually impaired and blind people: State-of-the-art and future trends","volume":"11","author":"Bhowmick","year":"2017","journal-title":"J. Multimodal User Interfaces"},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Messaoudi, M.D., Menelas, B.-A.J., and Mcheick, H. (2022). Review of Navigation Assistive Tools and Technologies for the Visually Impaired. Sensors, 22.","DOI":"10.3390\/s22207888"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Elmannai, W., and Elleithy, K. (2017). Sensor-based assistive devices for visually-impaired people: Current status, challenges, and future directions. Sensors, 17.","DOI":"10.3390\/s17030565"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"20","DOI":"10.1186\/s40691-014-0020-7","article-title":"Dress, body and self: Research in the social psychology of dress","volume":"1","author":"Johnson","year":"2014","journal-title":"Fash. Text."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Rocha, D., Carvalho, V., Oliveira, E., Goncalves, J., and Azevedo, F. (2017, January 2\u20134). MyEyes-automatic combination system of clothing parts to blind people: First insights. Proceedings of the 2017 IEEE 5th International Conference on Serious Games and Applications for Health (SeGAH), Perth, Australia.","DOI":"10.1109\/SeGAH.2017.7939298"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Rocha, D., Carvalho, V., and Oliveira, E. (2017, January 10\u201314). MyEyes\u2014Automatic Combination System of Clothing Parts to Blind People: Prototype Validation. Proceedings of the SENSORDEVICES\u2019 2017 Conference, Rome, Italy.","DOI":"10.1109\/SeGAH.2017.7939298"},{"key":"ref_9","first-page":"26","article-title":"Development of an Automatic Combination System of Clothing Parts for Blind People: MyEyes","volume":"219","author":"Rocha","year":"2017","journal-title":"Sens. Transducers"},{"key":"ref_10","unstructured":"Tavares, J.M.R.S., and Natal Jorge, R.M. (2019). Extracting Clothing Features for Blind People Using Image Processing and Machine Learning Techniques: First Insights BT\u2014VipIMAGE 2019, Springer International Publishing."},{"key":"ref_11","unstructured":"Ahram, T., and Taiar, R. (2021, January 27\u201329). Understand the Importance of Garments\u2019 Identification and Combination to Blind People. Proceedings of the Human Interaction, Emerging Technologies and Future Systems V, Paris, France."},{"key":"ref_12","unstructured":"Rocha, D., Carvalho, V., Soares, F., and Oliveira, E. (2020). Design, Learning, and Innovation, Springer International Publishing."},{"key":"ref_13","unstructured":"Garcia, N.M., Pires, I.M., and Goleva, R. (2020). IoT Technologies for HealthCare, Proceedings of the 8th EAI International Conference, HealthyIoT 2021, Virtual Event, 24\u201326 November 2021, Springer International Publishing."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"7068349","DOI":"10.1155\/2018\/7068349","article-title":"Deep Learning for Computer Vision: A Brief Review","volume":"2018","author":"Voulodimos","year":"2018","journal-title":"Comput. Intell. Neurosci."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Bhatt, D., Patel, C., Talsania, H., Patel, J., Vaghela, R., Pandya, S., Modi, K., and Ghayvat, H. (2021). CNN Variants for Computer Vision: History, Architecture, Application, Challenges and Future Scope. Electronics, 10.","DOI":"10.3390\/electronics10202470"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Patel, C., Bhatt, D., Sharma, U., Patel, R., Pandya, S., Modi, K., Cholli, N., Patel, A., Bhatt, U., and Khan, M.A. (2022). DBGC: Dimension-Based Generic Convolution Block for Object Recognition. Sensors, 22.","DOI":"10.3390\/s22051780"},{"key":"ref_17","unstructured":"Pereira, F., Burges, C.J.C., Bottou, L., and Weinberger, K.Q. (2012, January 3\u20138). ImageNet Classification with Deep Convolutional Neural Networks. Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA."},{"key":"ref_18","unstructured":"(2020, July 13). ImageNet Large Scale Visual Recognition Competition (ILSVRC). Available online: http:\/\/www.image-net.org\/challenges\/LSVRC\/."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. (2009, January 20\u201325). ImageNet: A large-scale hierarchical image database. Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref_20","unstructured":"Simonyan, K., and Zisserman, A. (2015, January 7\u20139). Very deep convolutional networks for large-scale image recognition. Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015\u2014Conference Track Proceedings, San Diego, CA, USA."},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2014). Going Deeper with Convolutions. arXiv.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_22","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., and Keutzer, K. (2016). SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and. arXiv."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z. (2015). Rethinking the Inception Architecture for Computer Vision. arXiv.","DOI":"10.1109\/CVPR.2016.308"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2015). Deep Residual Learning for Image Recognition. arXiv.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Ma, N., Zhang, X., Zheng, H.-T., and Sun, J. (2018). ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. arXiv.","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., and Chen, L.-C. (2018). MobileNetV2: Inverted Residuals and Linear Bottlenecks. arXiv.","DOI":"10.1109\/CVPR.2018.00474"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Howard, A., Sandler, M., Chu, G., Chen, L.-C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., and Vasudevan, V. (2019). Searching for MobileNetV3. arXiv.","DOI":"10.1109\/ICCV.2019.00140"},{"key":"ref_28","unstructured":"Tan, M., and Le, Q.V. (2019). EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Radosavovic, I., Kosaraju, R.P., Girshick, R., He, K., and Doll\u00e1r, P. (2020). Designing Network Design Spaces. arXiv.","DOI":"10.1109\/CVPR42600.2020.01044"},{"key":"ref_30","first-page":"72","article-title":"Fashion meets computer vision: A survey","volume":"54","author":"Cheng","year":"2021","journal-title":"ACM Comput. Surv."},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Chen, Q., Huang, J., Feris, R., Brown, L.M., Dong, J., and Yan, S. (2015, January 7\u201312). Deep domain adaptation for describing people based on fine-grained clothing attributes. Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7299169"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Qiu, S., Wang, X., and Tang, X. (2016, January 27\u201330). DeepFashion: Powering Robust Clothes Recognition and Retrieval with Rich Annotations. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.124"},{"key":"ref_33","unstructured":"Hara, K., Jagadeesh, V., and Piramuthu, R. (2014). Fashion Apparel Detection: The Role of Deep Convolutional Neural Network and Pose-dependent Priors. arXiv."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Corbi\u00e8re, C., Ben-Younes, H., Ram\u00e9, A., and Ollion, C. (2017, January 22\u201329). Leveraging Weakly Annotated Data for Fashion Image Retrieval and Label Prediction. Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy.","DOI":"10.1109\/ICCVW.2017.266"},{"key":"ref_35","unstructured":"Wang, W., Xu, Y., Shen, J., and Zhu, S.-C. Proceedings of the Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18\u201323 June 2018."},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Li, P., Li, Y., Jiang, X., and Zhen, X. (2019). Two-Stream Multi-Task Network for Fashion Recognition. arXiv.","DOI":"10.1109\/ICIP.2019.8803394"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Cho, H., Ahn, C., Yoo, K.M., Seol, J., and Lee, S. (2019, January 27\u201328). Leveraging Class Hierarchy in Fashion Classification. Proceedings of the 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea.","DOI":"10.1109\/ICCVW.2019.00398"},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Lu, Y., Kumar, A., Zhai, S., Cheng, Y., Javidi, T., and Feris, R. (2016). Fully-adaptive Feature Sharing in Multi-Task Networks with Applications in Person Attribute Classification. arXiv.","DOI":"10.1109\/CVPR.2017.126"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"328","DOI":"10.1016\/j.eswa.2018.09.022","article-title":"Hierarchical convolutional neural networks for fashion image classification","volume":"116","author":"Seo","year":"2019","journal-title":"Expert Syst. Appl."},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Fengzi, L., Kant, S., Araki, S., Bangera, S., and Shukla, S. (2020). Neural Networks for Fashion Image Classification and Visual Search. arXiv.","DOI":"10.2139\/ssrn.3602664"},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"115195","DOI":"10.1016\/j.eswa.2021.115195","article-title":"Condition-CNN: A hierarchical multi-label fashion image classification model","volume":"182","author":"Kolisnik","year":"2021","journal-title":"Expert Syst. Appl."},{"key":"ref_42","first-page":"9948808","article-title":"Fabric Defect Detection in Textile Manufacturing: A Survey of the State of the Art","volume":"2021","author":"Li","year":"2021","journal-title":"Secur. Commun. Netw."},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2015). You Only Look Once: Unified, Real-Time Object Detection. arXiv.","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref_44","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., and Berg, A.C. (2015). Computer Vision\u2014ECCV 2016, Springer."},{"key":"ref_45","unstructured":"Ren, S., He, K., Girshick, R., and Sun, J. (2015). Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. arXiv."},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., and Girshick, R. (2017). Mask R-CNN. arXiv.","DOI":"10.1109\/ICCV.2017.322"},{"key":"ref_47","doi-asserted-by":"crossref","unstructured":"Liu, Z., Yan, S., Luo, P., Wang, X., and Tang, X. (2016). Fashion Landmark Detection in the Wild. arXiv.","DOI":"10.1007\/978-3-319-46475-6_15"},{"key":"ref_48","unstructured":"Xiao, H., Rasul, K., and Vollgraf, R. (2017). Fashion-MNIST: A Novel Image Dataset for Benchmarking Machine Learning Algorithms. arXiv."},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Ge, Y., Zhang, R., Wu, L., Wang, X., Tang, X., and Luo, P. (2019, January 15\u201320). A Versatile Benchmark for Detection, Pose Estimation, Segmentation and Re-Identification of Clothing Images. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00548"},{"key":"ref_50","unstructured":"(2021, December 28). Fashion Product Images Dataset|Kaggle. Available online: https:\/\/www.kaggle.com\/paramaggarwal\/fashion-product-images-dataset."},{"key":"ref_51","unstructured":"Ioffe, S., and Szegedy, C. (2015). Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv."}],"container-title":["Applied Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2076-3417\/13\/3\/1925\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T18:21:52Z","timestamp":1760120512000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2076-3417\/13\/3\/1925"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,2,2]]},"references-count":51,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2023,2]]}},"alternative-id":["app13031925"],"URL":"https:\/\/doi.org\/10.3390\/app13031925","relation":{},"ISSN":["2076-3417"],"issn-type":[{"value":"2076-3417","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,2,2]]}}}