{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T13:43:48Z","timestamp":1774446228346,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":87,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,3,27]],"date-time":"2023-03-27T00:00:00Z","timestamp":1679875200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,3,27]]},"DOI":"10.1145\/3581641.3584099","type":"proceedings-article","created":{"date-parts":[[2023,3,27]],"date-time":"2023-03-27T16:16:52Z","timestamp":1679933812000},"page":"46-64","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":25,"title":["SeeChart: Enabling Accessible Visualizations Through Interactive Natural Language Interface For People with Visual Impairments"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9079-9757","authenticated-orcid":false,"given":"Md Zubair Ibne","family":"Alam","sequence":"first","affiliation":[{"name":"Department Electrical Engineering and Computer Science, York University, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4032-1445","authenticated-orcid":false,"given":"Shehnaz","family":"Islam","sequence":"additional","affiliation":[{"name":"York University, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9789-6645","authenticated-orcid":false,"given":"Enamul","family":"Hoque","sequence":"additional","affiliation":[{"name":"York University, Canada"}]}],"member":"320","published-online":{"date-parts":[[2023,3,27]]},"reference":[{"key":"e_1_3_2_2_1_1","unstructured":"2022. Chartblocks. https:\/\/www.chartblocks.com\/."},{"key":"e_1_3_2_2_2_1","unstructured":"2022. Graphiq. https:\/\/www.graphiq.com\/."},{"key":"e_1_3_2_2_3_1","unstructured":"2022. NV Access. https:\/\/www.nvaccess.org\/."},{"key":"e_1_3_2_2_4_1","unstructured":"2022. Plotly. https:\/\/plot.ly\/."},{"key":"e_1_3_2_2_5_1","volume-title":"Retrieved","year":"2022","unstructured":"2022. Sample Chart from Statista. Retrieved March 20, 2022 from https:\/\/www.statista.com\/statistics\/420703\/canadian-snowmobile-registrations-by-province\/."},{"key":"e_1_3_2_2_6_1","unstructured":"2022. SAS. https:\/\/support.sas.com\/accessibility\/."},{"key":"e_1_3_2_2_7_1","unstructured":"2022. Web Speech API. https:\/\/developer.mozilla.org\/en-US\/docs\/Web\/API\/Web_Speech_API\/Using_the_Web_Speech_API."},{"key":"e_1_3_2_2_8_1","volume-title":"World blindness and visual impairment: despite many successes, the problem is growing. Community eye health 30, 100","author":"Ackland Peter","year":"2017","unstructured":"Peter Ackland, Serge Resnikoff, and Rupert Bourne. 2017. World blindness and visual impairment: despite many successes, the problem is growing. Community eye health 30, 100 (2017), 71."},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFVIS.2005.1532136"},{"key":"e_1_3_2_2_10_1","unstructured":"Amazon. 2022. Turn text into lifelike speech using deep learning. https:\/\/aws.amazon.com\/polly\/."},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3174168"},{"key":"e_1_3_2_2_12_1","first-page":"519","article-title":"Beyond memorability: Visualization recognition and recall","volume":"22","author":"Borkin A","year":"2015","unstructured":"Michelle\u00a0A Borkin, Zoya Bylinskii, Nam\u00a0Wook Kim, Constance\u00a0May Bainbridge, Chelsea\u00a0S Yeh, Daniel Borkin, Hanspeter Pfister, and Aude Oliva. 2015. Beyond memorability: Visualization recognition and recall. IEEE TVCG 22, 1 (2015), 519\u2013528.","journal-title":"IEEE TVCG"},{"key":"e_1_3_2_2_13_1","volume-title":"What makes a visualization memorable?IEEE TVCG 19, 12","author":"Borkin A","year":"2013","unstructured":"Michelle\u00a0A Borkin, Azalea\u00a0A Vo, Zoya Bylinskii, Phillip Isola, Shashank Sunkavalli, Aude Oliva, and Hanspeter Pfister. 2013. What makes a visualization memorable?IEEE TVCG 19, 12 (2013), 2306\u20132315."},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2011.185"},{"key":"e_1_3_2_2_15_1","volume-title":"International Conference on Auditory Display.","author":"Brown M","year":"2003","unstructured":"Lorna\u00a0M Brown, Stephen\u00a0A Brewster, SA Ramloll, R Burton, and Beate Riedel. 2003. Design guidelines for audio presentation of graphs and tables. International Conference on Auditory Display."},{"key":"e_1_3_2_2_16_1","unstructured":"Fighting\u00a0Blindness Canada. 2022. TOGETHER LET\u2019S MOVE RESEARCH FORWARD. https:\/\/www.fightingblindness.ca\/."},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/2395123.2395126"},{"key":"e_1_3_2_2_18_1","volume-title":"Information visualization","author":"Carpendale Sheelagh","unstructured":"Sheelagh Carpendale. 2008. Evaluating information visualizations. In Information visualization. Springer, 19\u201345."},{"key":"e_1_3_2_2_19_1","unstructured":"Diagram Center. 2022. Specific Guidelines \u2013 Graphs. http:\/\/diagramcenter.org\/specific-guidelines-e.html."},{"key":"e_1_3_2_2_20_1","unstructured":"CFPB. 2022. Data visualization guidelines. https:\/\/cfpb.github.io\/design-system\/guidelines\/data-visualization-guidelines."},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"crossref","unstructured":"J. Choi Sanghun Jung Deok\u00a0Gun Park J. Choo and N. Elmqvist. 2019. Visualizing for the Non\u2010Visual: Enabling the Visually Impaired to Use Visualization. Computer Graphics Forum 38 (2019).","DOI":"10.1111\/cgf.13686"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/1753846.1753999"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2021.3114829"},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-71249-9_9"},{"key":"e_1_3_2_2_25_1","volume-title":"The web: Access and inclusion for disabled people","author":"Disability\u00a0Rights Commission","unstructured":"Disability\u00a0Rights Commission. 2004. The web: Access and inclusion for disabled people; a formal investigation. The Stationery Office."},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1177\/1473871618806555"},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1162\/COLI_a_00091"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/1805986.1806009"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.14522"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2010.10.003"},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/2533682.2533683"},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","unstructured":"Leo Ferres Avi Parush Zhihong Li Yandu Oppacher and Gitte Lindgaard. 2006. Representing and querying line graphs in natural language: The iGraph system. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 4073 LNCS. https:\/\/doi.org\/10.1007\/11795018_25","DOI":"10.1007\/11795018_25"},{"key":"e_1_3_2_2_33_1","unstructured":"Canadian National\u00a0Institute for\u00a0the Blind. 2022. Canadian National Institute for the Blind. https:\/\/www.cnib.ca\/en?region=on."},{"key":"e_1_3_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/2700648.2809840"},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/2642918.2647411"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2017.2659744"},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.1037\/e577632012-009"},{"key":"e_1_3_2_2_38_1","unstructured":"Highcharts. 2022. Interactive JavaScript charts library. https:\/\/www.highcharts.com\/."},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3517465"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2019.2934431"},{"key":"e_1_3_2_2_41_1","first-page":"309","article-title":"Applying pragmatics principles for interaction with visual analytics","volume":"24","author":"Hoque Enamul","year":"2017","unstructured":"Enamul Hoque, Vidya Setlur, Melanie Tory, and Isaac Dykeman. 2017. Applying pragmatics principles for interaction with visual analytics. IEEE TVCG 24, 1 (2017), 309\u2013318.","journal-title":"IEEE TVCG"},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/2470654.2481374"},{"key":"e_1_3_2_2_43_1","volume-title":"Communicating Visualizations without Visuals: Investigation of Visualization Alternative Text for People with Visual Impairments","author":"Jung Crescentia","year":"2021","unstructured":"Crescentia Jung, Shubham Mehta, Atharva Kulkarni, Yuhang Zhao, and Yea-Seul Kim. 2021. Communicating Visualizations without Visuals: Investigation of Visualization Alternative Text for People with Visual Impairments. IEEE TVCG (2021)."},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3025453.3025957"},{"key":"e_1_3_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.2203.06486"},{"key":"e_1_3_2_2_46_1","first-page":"2009","article-title":"Open coding","volume":"23","author":"Khandkar Shahedul\u00a0Huq","year":"2009","unstructured":"Shahedul\u00a0Huq Khandkar. 2009. Open coding. University of Calgary 23(2009), 2009.","journal-title":"University of Calgary"},{"key":"e_1_3_2_2_47_1","volume-title":"12th International Conference on Auditory Display (ICAD2006)","author":"Kildal Johan","year":"2006","unstructured":"Johan Kildal and Stephen\u00a0A Brewster. 2006. Providing a size-independent overview of non-visual tables. In 12th International Conference on Auditory Display (ICAD2006). 8\u201315."},{"key":"e_1_3_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376467"},{"key":"e_1_3_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.14298"},{"key":"e_1_3_2_2_50_1","unstructured":"Xiaoyi Liu Diego Klabjan and Patrick\u00a0N. Bless. 2019. Data Extraction from Charts via Single Deep Neural Network. ArXiv abs\/1906.11906(2019)."},{"key":"e_1_3_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/VISUAL.2019.8933762"},{"key":"e_1_3_2_2_52_1","volume-title":"Accessible Visualization via Natural Language Descriptions: A Four-Level Model of Semantic Content","author":"Lundgard Alan","year":"2022","unstructured":"Alan Lundgard and Arvind Satyanarayan. 2022. Accessible Visualization via Natural Language Descriptions: A Four-Level Model of Semantic Content. IEEE Trans. Visualization & Comp. Graphics (Proc. IEEE VIS) (2022). http:\/\/vis.csail.mit.edu\/pubs\/vis-text-model"},{"key":"e_1_3_2_2_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV48630.2021.00196"},{"key":"e_1_3_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cola.2022.101107"},{"key":"e_1_3_2_2_55_1","doi-asserted-by":"publisher","DOI":"10.5555\/972749.972754"},{"key":"e_1_3_2_2_56_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10209-007-0098-4"},{"key":"e_1_3_2_2_57_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.inlg-1.20"},{"key":"e_1_3_2_2_58_1","unstructured":"Canadian\u00a0Council of\u00a0the Blind. 2022. Canadian Council of the Blind Is the Voice of the Blind in Canada. https:\/\/ccbnational.net\/."},{"key":"e_1_3_2_2_59_1","unstructured":"PennState. 2022. Charts & Accessibility. https:\/\/accessibility.psu.edu\/images\/charts\/."},{"key":"e_1_3_2_2_60_1","doi-asserted-by":"crossref","unstructured":"Jorge Poco and Jeffrey Heer. 2017. Reverse-Engineering Visualizations: Recovering Visual Encodings from Chart Images. In Computer Graphics Forum (Proc. EuroVis) Vol.\u00a036. 353\u2013363.","DOI":"10.1111\/cgf.13193"},{"key":"e_1_3_2_2_61_1","first-page":"637","article-title":"Extracting and retargeting color mappings from bitmap images of visualizations","volume":"24","author":"Poco Jorge","year":"2018","unstructured":"Jorge Poco, Angela Mayhua, and Jeffrey Heer. 2018. Extracting and retargeting color mappings from bitmap images of visualizations. IEEE TVCG (Proc. InfoVis) 24, 1 (2018), 637\u2013646.","journal-title":"IEEE TVCG (Proc. InfoVis)"},{"key":"e_1_3_2_2_62_1","unstructured":"Quill. 2020. Narrative science. https:\/\/narrativescience.com\/quill\/."},{"key":"e_1_3_2_2_63_1","volume-title":"People and Computers XV\u2014Interaction without Frontiers","author":"Ramloll Rameshsharma","unstructured":"Rameshsharma Ramloll, Stephen Brewster, Wai Yu, and Beate Riedel. 2001. Using non-speech sounds to improve access to 2D tabular numerical information for visually impaired users. In People and Computers XV\u2014Interaction without Frontiers. Springer, 515\u2013529."},{"key":"e_1_3_2_2_64_1","doi-asserted-by":"publisher","DOI":"10.5555\/1610163.1610180"},{"key":"e_1_3_2_2_65_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-29381-9_16"},{"key":"e_1_3_2_2_66_1","first-page":"341","article-title":"Vega-lite: A grammar of interactive graphics","volume":"23","author":"Satyanarayan Arvind","year":"2017","unstructured":"Arvind Satyanarayan, Dominik Moritz, Kanit Wongsuphasawat, and Jeffrey Heer. 2017. Vega-lite: A grammar of interactive graphics. IEEE TVCG 23, 1 (2017), 341\u2013350.","journal-title":"IEEE TVCG"},{"key":"e_1_3_2_2_67_1","doi-asserted-by":"publisher","DOI":"10.1145\/2047196.2047247"},{"key":"e_1_3_2_2_68_1","unstructured":"System\u00a0Usability Scale. 2022. SUS. https:\/\/www.usability.gov\/how-to-and-tools\/methods\/system-usability-scale.html."},{"key":"e_1_3_2_2_69_1","unstructured":"Freedom Scientific. 2022. JAWS-Job Access With Speech. https:\/\/www.freedomscientific.com\/products\/software\/jaws\/."},{"key":"e_1_3_2_2_70_1","volume-title":"Rixie Tiffany Ko\u00a0Leong","author":"Ahmed Masry Megh Thakkar Xiang Lin","year":"2020","unstructured":"Xiang Lin Ahmed Masry Megh Thakkar Enamul Hoque Shafiq\u00a0Joty Shankar\u00a0Kanthara, Rixie Tiffany Ko\u00a0Leong. 2020. Chart-to-Text: A Large-Scale Benchmark for Chart Summarization. arXiv preprint arXiv:2203.06486(2020)."},{"key":"e_1_3_2_2_71_1","doi-asserted-by":"publisher","DOI":"10.1145\/3441852.3471202"},{"key":"e_1_3_2_2_72_1","doi-asserted-by":"publisher","DOI":"10.1109\/CCNC.2018.8319239"},{"key":"e_1_3_2_2_73_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3517431"},{"key":"e_1_3_2_2_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3517678"},{"key":"e_1_3_2_2_75_1","doi-asserted-by":"publisher","DOI":"10.1145\/3399715.3399829"},{"key":"e_1_3_2_2_76_1","first-page":"672","article-title":"Augmenting visualizations with interactive data facts to facilitate interpretation and communication","volume":"25","author":"Srinivasan Arjun","year":"2018","unstructured":"Arjun Srinivasan, Steven\u00a0M Drucker, Alex Endert, and John Stasko. 2018. Augmenting visualizations with interactive data facts to facilitate interpretation and communication. IEEE TVCG 25, 1 (2018), 672\u2013681.","journal-title":"IEEE TVCG"},{"key":"e_1_3_2_2_77_1","first-page":"511","article-title":"Orko: Facilitating Multimodal Interaction for Visual Exploration and Analysis of Networks","volume":"24","author":"Srinivasan Arjun","year":"2018","unstructured":"Arjun Srinivasan and John Stasko. 2018. Orko: Facilitating Multimodal Interaction for Visual Exploration and Analysis of Networks. IEEE TVCG 24, 1 (2018), 511\u2013521.","journal-title":"IEEE TVCG"},{"key":"e_1_3_2_2_78_1","unstructured":"Statista. 2022. Statista. https:\/\/statista.com\/."},{"key":"e_1_3_2_2_79_1","volume-title":"Interactive sonification of spreadsheets","author":"Stockman Tony","unstructured":"Tony Stockman, Christopher Frauenberger, and Greg Hind. 2005. Interactive sonification of spreadsheets. Georgia Institute of Technology."},{"key":"e_1_3_2_2_80_1","unstructured":"W3C. 2022. STANDARDS. https:\/\/www.w3.org\/standards\/."},{"key":"e_1_3_2_2_81_1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.14523"},{"key":"e_1_3_2_2_82_1","first-page":"895","article-title":"DataShot: Automatic Generation of Fact Sheets from Tabular Data","volume":"26","author":"Wang Yun","year":"2019","unstructured":"Yun Wang, Zhida Sun, Haidong Zhang, Weiwei Cui, Ke Xu, Xiaojuan Ma, and Dongmei Zhang. 2019. DataShot: Automatic Generation of Fact Sheets from Tabular Data. IEEE TVCG 26, 1 (2019), 895\u2013905.","journal-title":"IEEE TVCG"},{"key":"e_1_3_2_2_83_1","unstructured":"WCAG. 2022. Complex Images. https:\/\/www.w3.org\/WAI\/tutorials\/images\/complex\/."},{"key":"e_1_3_2_2_84_1","unstructured":"wordcounter. 2022. How Fast Does the Average Person Speak?https:\/\/wordcounter.net\/blog\/2016\/06\/02\/101702_how-fast-average-person-speaks.html."},{"key":"e_1_3_2_2_85_1","unstructured":"Wordsmith. 2020. Wordsmith by automated insights inc.https:\/\/automatedinsights.com\/wordsmith."},{"key":"e_1_3_2_2_86_1","doi-asserted-by":"publisher","DOI":"10.1145\/1352782.1352786"},{"key":"e_1_3_2_2_87_1","volume-title":"Rich Screen Reader Experiences for Accessible Data Visualization. (05","author":"Zong Jonathan","year":"2022","unstructured":"Jonathan Zong, Crystal Lee, Alan Lundgard, JiWoong Jang, Daniel Hajas, and Arvind Satyanarayan. 2022. Rich Screen Reader Experiences for Accessible Data Visualization. (05 2022)."}],"event":{"name":"IUI '23: 28th International Conference on Intelligent User Interfaces","location":"Sydney NSW Australia","acronym":"IUI '23","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence","SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Proceedings of the 28th International Conference on Intelligent User Interfaces"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581641.3584099","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3581641.3584099","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:36:21Z","timestamp":1750178181000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581641.3584099"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,3,27]]},"references-count":87,"alternative-id":["10.1145\/3581641.3584099","10.1145\/3581641"],"URL":"https:\/\/doi.org\/10.1145\/3581641.3584099","relation":{},"subject":[],"published":{"date-parts":[[2023,3,27]]},"assertion":[{"value":"2023-03-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}