{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T21:17:21Z","timestamp":1776115041043,"version":"3.50.1"},"reference-count":23,"publisher":"MDPI AG","issue":"12","license":[{"start":{"date-parts":[[2024,12,23]],"date-time":"2024-12-23T00:00:00Z","timestamp":1734912000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Computation"],"abstract":"<jats:p>This study addresses the accessibility challenges faced by individuals with visual impairments due to limited access to graphic information, which significantly impacts their educational and social integration. Traditional methods for producing tactile graphics are labor-intensive and require specialized expertise, limiting their availability. Recent advancements in generative models, such as GANs, diffusion models, and VAEs, offer potential solutions to automate the creation of tactile images. In this work, we propose a novel generative model conditioned on text prompts, integrating a Bidirectional and Auto-Regressive Transformer (BART) and Vector Quantized Variational Auto-Encoder (VQ-VAE). This model transforms textual descriptions into tactile graphics, addressing key requirements for legibility and accessibility. The model\u2019s performance was evaluated using cross-entropy, perplexity, mean square error, and CLIP Score metrics, demonstrating its ability to generate high-quality, customizable tactile images. Testing with educational and rehabilitation institutions confirmed the practicality and efficiency of the system, which significantly reduces production time and requires minimal operator expertise. The proposed approach enhances the production of inclusive educational materials, enabling improved access to quality education and fostering greater independence for individuals with visual impairments. Future research will focus on expanding the training dataset and refining the model for complex scenarios.<\/jats:p>","DOI":"10.3390\/computation12120251","type":"journal-article","created":{"date-parts":[[2024,12,23]],"date-time":"2024-12-23T10:06:52Z","timestamp":1734948412000},"page":"251","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["Enhancing Accessibility: Automated Tactile Graphics Generation for Individuals with Visual Impairments"],"prefix":"10.3390","volume":"12","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1636-1447","authenticated-orcid":false,"given":"Yehor","family":"Dzhurynskyi","sequence":"first","affiliation":[{"name":"SQUAD Ukraine LLC, 3 Korolenkivskya Str., 01033 Kyiv, Ukraine"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6650-2703","authenticated-orcid":false,"given":"Volodymyr","family":"Mayik","sequence":"additional","affiliation":[{"name":"Department of Printing Technologies and Packaging (PPT), Institute of Printing Art and Media Technologies, Lviv Polytechnic National University, 28a, Stepan Bandera Str., 79013 Lviv, Ukraine"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8552-0942","authenticated-orcid":false,"given":"Lyudmyla","family":"Mayik","sequence":"additional","affiliation":[{"name":"Department of Multimedia Technologies (MT), Institute of Printing Art and Media Technologies, Lviv Polytechnic National University, 28a, Stepan Bandera Str., 79013 Lviv, Ukraine"}]}],"member":"1968","published-online":{"date-parts":[[2024,12,23]]},"reference":[{"key":"ref_1","first-page":"71","article-title":"World blindness and visual impairment: Despite many successes, the problem is growing","volume":"8","author":"Ackland","year":"2018","journal-title":"Community Eye Health J."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"215","DOI":"10.1177\/0145482X211016918","article-title":"Graphic Reading Performance of Students with Visual Impairments and Its Implication for Instruction and Assessment","volume":"115","author":"Zebehazy","year":"2021","journal-title":"J. Vis. Impair. Blind."},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Mukhiddinov, M., and Kim, S.-Y. (2021). A Systematic Literature Review on the Automatic Creation of Tactile Graphics for the Blind and Visually Impaired. Processes, 9.","DOI":"10.3390\/pr9101726"},{"key":"ref_4","unstructured":"GBD 2019 Blindness and Vision Impairment Collaborators on behalf of the Vision Loss Expert Group of the Global Burden of Disease Study (2021). Trends in prevalence of blindness and distance and near vision impairment over 30 years: An analysis for the Global Burden of Disease Study. Lancet Glob. Health, 9, e130\u2013e143."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Mayik, V., Dudok, T., Mayik, L., Lotoshynska, N., Izonin, I., and Kusmierczyk, J. (2022). An Approach Towards Vacuum Forming Process Using PostScript for Making Braille. Advances in Computer Science for Engineering and Manufacturing, Springer International Publishing.","DOI":"10.1007\/978-3-031-03877-8_4"},{"key":"ref_6","first-page":"161","article-title":"Towards Automatic Generation of Tactile Graphics","volume":"96","author":"Way","year":"1996","journal-title":"Rehabil. Eng. Assist. Technol. Soc. N. Am."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"81","DOI":"10.1109\/86.559353","article-title":"Automatic visual to tactile translation-Part I: Human factors, access methods, and image manipulation","volume":"5","author":"Way","year":"1997","journal-title":"IEEE Trans. Rehabil. Eng."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"95","DOI":"10.1109\/86.559354","article-title":"Automatic visual to tactile translation. II. Evaluation of the TACTile image creation system","volume":"5","author":"Way","year":"1997","journal-title":"IEEE Trans. Rehabil. Eng."},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Ferro, T., and Pawluk, D. (2013, January 23). Automatic image conversion to tactile graphic. Proceedings of the 15th International ACM SIGACCESS Conference on Computers and Accessibility, Bellevue, WA, USA.","DOI":"10.1145\/2513383.2513406"},{"key":"ref_10","unstructured":"(2024, November 18). Braille Authority of North America & Canadian Braille Authority: Guidelines and Standards for Tactile Graphics. Available online: https:\/\/www.brailleauthority.org\/guidelines-and-standards-tactile-graphics."},{"key":"ref_11","unstructured":"(2024, November 18). Polish Association of the Blind: Instructions for Creating and Adapting Illustrations and Typhlographic Materials for Blind Students. Available online: https:\/\/pzn.org.pl\/wp-content\/uploads\/2016\/07\/instrukcja_tworzenia_i_adaptowania_ilustracji_i_materialow_tyflograficznych_dla_niewidomych.pdf."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Pak\u0117nait\u0117, K., Nedelev, P., Kamperou, E., Proulx, M., and Hall, P. (2022). Communicating Photograph Content Through Tactile Images to People with Visual Impairments. Front. Comput. Sci., 3.","DOI":"10.3389\/fcomp.2021.787735"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Pakenaite, K., Kamperou, E., Proulx, M.J., Sharma, A., and Hall, P. (2024, January 11). Pic2Tac: Creating Accessible Tactile Images using Semantic Information from Photographs. Proceedings of the Eighteenth International Conference on Tangible, Embedded, and Embodied Interaction, Cork, Ireland.","DOI":"10.1145\/3623509.3633377"},{"key":"ref_14","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y. (2014). Generative adversarial networks 2014. arXiv."},{"key":"ref_15","unstructured":"Kingma, D.P., and Welling, M. (2014, January 8). Auto-Encoding Variational Bayes. Proceedings of the International Conference on Learning Representations, London, UK."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. (2022, January 19). High-Resolution Image Synthesis with Latent Diffusion Models. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref_17","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., and Polosukhin, I. (2017, January 4). Attention Is All You Need. Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA."},{"key":"ref_18","unstructured":"Yingchen, Y., Fangneng, Z., Rongliang, W., Jianxiong, P., Kaiwen, C., Shijian, L., Feiying, M., Xuansong, X., and Chunyan, M. (2021, January 20). Diverse Image Inpainting with Bidirectional and Autoregressive Transformers. Proceedings of the 29th ACM International Conference on Multimedia, New York, NY, USA."},{"key":"ref_19","unstructured":"Van den Oord, A., Vinyals, O., and Kavukcuoglu, K. (2017, January 4). Neural Discrete Representation Learning. Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA."},{"key":"ref_20","unstructured":"Rogers, A., Boyd-Graber, J., and Okazaki, N. (2023). A Formal Perspective on Byte-Pair Encoding. Findings of the Association for Computational Linguistics: ACL 2023, Association for Computational Linguistics. [1st ed.]."},{"key":"ref_21","unstructured":"(2024, November 18). LanguageTool API NLP UK. Available online: https:\/\/github.com\/brown-uk\/nlp_uk."},{"key":"ref_22","unstructured":"(2024, November 18). American Printing House: Tactile Graphic Image Library. Available online: https:\/\/imagelibrary.aph.org\/portals\/aphb\/#page\/welcome."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Hessel, J., Holtzman, A., Forbes, M., Le Bras, R., and Choi, Y. (2021, January 7). CLIPScore: A Reference-free Evaluation Metric for Image Captioning. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic.","DOI":"10.18653\/v1\/2021.emnlp-main.595"}],"container-title":["Computation"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2079-3197\/12\/12\/251\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T16:58:42Z","timestamp":1760115522000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2079-3197\/12\/12\/251"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,23]]},"references-count":23,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2024,12]]}},"alternative-id":["computation12120251"],"URL":"https:\/\/doi.org\/10.3390\/computation12120251","relation":{},"ISSN":["2079-3197"],"issn-type":[{"value":"2079-3197","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,23]]}}}