{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T20:03:34Z","timestamp":1776110614163,"version":"3.50.1"},"reference-count":122,"publisher":"Elsevier BV","issue":"4","license":[{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,8,28]],"date-time":"2025-08-28T00:00:00Z","timestamp":1756339200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100005950","name":"Hong Kong University of Science and Technology","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100005950","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62172398"],"award-info":[{"award-number":["62172398"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Visual Informatics"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1016\/j.visinf.2025.100271","type":"journal-article","created":{"date-parts":[[2025,8,29]],"date-time":"2025-08-29T23:29:27Z","timestamp":1756510167000},"page":"100271","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":2,"title":["A survey of visual insight mining: Connecting data and insights via visualization"],"prefix":"10.1016","volume":"9","author":[{"given":"Yijie","family":"Lian","sequence":"first","affiliation":[]},{"given":"Jianing","family":"Hao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5600-8824","authenticated-orcid":false,"given":"Wei","family":"Zeng","sequence":"additional","affiliation":[]},{"given":"Qiong","family":"Luo","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.visinf.2025.100271_b1","series-title":"GPT-4 technical report","author":"Achiam","year":"2023"},{"key":"10.1016\/j.visinf.2025.100271_b2","series-title":"Color Theory and Its Application in Art and Design","author":"Agoston","year":"2013"},{"key":"10.1016\/j.visinf.2025.100271_b3","series-title":"GPT-4: A review on advancements and opportunities in natural language processing","author":"Baktash","year":"2023"},{"key":"10.1016\/j.visinf.2025.100271_b4","doi-asserted-by":"crossref","unstructured":"Battle,\u00a0L., Chang,\u00a0R., Stonebraker,\u00a0M., 2016. Dynamic prefetching of data tiles for interactive visualization. In: Proc. ACM SIGMOD. pp. 1363\u20131375.","DOI":"10.1145\/2882903.2882919"},{"issue":"9","key":"10.1016\/j.visinf.2025.100271_b5","doi-asserted-by":"crossref","first-page":"6075","DOI":"10.1109\/TVCG.2023.3326698","article-title":"What Do We Mean When We Say \u201cInsight\u201d? A formal synthesis of existing theory","volume":"30","author":"Battle","year":"2023","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"2","key":"10.1016\/j.visinf.2025.100271_b6","doi-asserted-by":"crossref","first-page":"131","DOI":"10.1037\/1082-989X.2.2.131","article-title":"Principles and procedures of exploratory data analysis","volume":"2","author":"Behrens","year":"1997","journal-title":"Psychol. Methods"},{"key":"10.1016\/j.visinf.2025.100271_b7","series-title":"Semiology of Graphics","author":"Bertin","year":"1983"},{"issue":"6","key":"10.1016\/j.visinf.2025.100271_b8","doi-asserted-by":"crossref","first-page":"1857","DOI":"10.1073\/pnas.1807180116","article-title":"Data visualization literacy: Definitions, conceptual frameworks, exercises, and assessments","volume":"116","author":"B\u00f6rner","year":"2019","journal-title":"Proc. Natl. Acad. Sci."},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b9","doi-asserted-by":"crossref","first-page":"111","DOI":"10.1109\/TVCG.2016.2598468","article-title":"Characterizing guidance in visual analytics","volume":"23","author":"Ceneda","year":"2016","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"3","key":"10.1016\/j.visinf.2025.100271_b10","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3641289","article-title":"A survey on evaluation of large language models","volume":"15","author":"Chang","year":"2024","journal-title":"ACM Trans. Intell. Syst. Technol."},{"key":"10.1016\/j.visinf.2025.100271_b11","doi-asserted-by":"crossref","unstructured":"Chen,\u00a0T., Liu,\u00a0S., Chang,\u00a0S., Cheng,\u00a0Y., Amini,\u00a0L., Wang,\u00a0Z., 2020. Adversarial robustness: From self-supervised pre-training to fine-tuning. In: Proc. CVPR. pp. 699\u2013708.","DOI":"10.1109\/CVPR42600.2020.00078"},{"key":"10.1016\/j.visinf.2025.100271_b12","doi-asserted-by":"crossref","unstructured":"Cheng,\u00a0Z.-Q., Dai,\u00a0Q., Hauptmann,\u00a0A.G., 2023. ChartReader: A unified framework for chart derendering and comprehension without heuristic rules. In: Proc. ICCV. pp. 22202\u201322213.","DOI":"10.1109\/ICCV51070.2023.02029"},{"key":"10.1016\/j.visinf.2025.100271_b13","series-title":"Visualizing Data","author":"Cleveland","year":"1993"},{"key":"10.1016\/j.visinf.2025.100271_b14","series-title":"Multivariate Data Analysis","author":"Cooley","year":"1971"},{"issue":"3","key":"10.1016\/j.visinf.2025.100271_b15","doi-asserted-by":"crossref","first-page":"378","DOI":"10.1109\/TVCG.2003.1207445","article-title":"From visual data exploration to visual data mining: A survey","volume":"9","author":"De\u00a0Oliveira","year":"2003","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"12","key":"10.1016\/j.visinf.2025.100271_b16","doi-asserted-by":"crossref","first-page":"1937","DOI":"10.14778\/3137765.3137813","article-title":"Foresight: recommending visual insights","volume":"10","author":"Demiralp","year":"2017","journal-title":"Proc. VLDB Endow."},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b17","first-page":"690","article-title":"DashBot: Insight-driven dashboard generation based on deep reinforcement learning","volume":"29","author":"Deng","year":"2022","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b18","doi-asserted-by":"crossref","unstructured":"Dibia,\u00a0V., 2023. LIDA: A Tool for Automatic Generation of Grammar-Agnostic Visualizations and Infographics using Large Language Models. In: Proc. ACL. pp. 113\u2013126.","DOI":"10.18653\/v1\/2023.acl-demo.11"},{"issue":"5","key":"10.1016\/j.visinf.2025.100271_b19","doi-asserted-by":"crossref","first-page":"33","DOI":"10.1109\/MCG.2019.2924636","article-title":"Data2Vis: Automatic generation of data visualizations using sequence-to-sequence recurrent neural networks","volume":"39","author":"Dibia","year":"2019","journal-title":"IEEE Comput. Graph. Appl."},{"issue":"4","key":"10.1016\/j.visinf.2025.100271_b20","doi-asserted-by":"crossref","first-page":"5","DOI":"10.1167\/2.4.5","article-title":"Ecological statistics of gestalt laws for the perceptual organization of contours","volume":"2","author":"Elder","year":"2002","journal-title":"J. Vis."},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b21","first-page":"197","article-title":"Dead or alive: Continuous data profiling for interactive data science","volume":"30","author":"Epperson","year":"2023","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"11","key":"10.1016\/j.visinf.2025.100271_b22","doi-asserted-by":"crossref","first-page":"27","DOI":"10.1145\/240455.240464","article-title":"The KDD process for extracting useful knowledge from volumes of data","volume":"39","author":"Fayyad","year":"1996","journal-title":"CACM"},{"key":"10.1016\/j.visinf.2025.100271_b23","doi-asserted-by":"crossref","unstructured":"Finnie-Ansley,\u00a0J., Denny,\u00a0P., Becker,\u00a0B.A., Luxton-Reilly,\u00a0A., Prather,\u00a0J., 2022. The robots are coming: Exploring the implications of openai codex on introductory programming. In: Proceedings of the Australasian Computing Education Conference. pp. 10\u201319.","DOI":"10.1145\/3511861.3511863"},{"issue":"8","key":"10.1016\/j.visinf.2025.100271_b24","doi-asserted-by":"crossref","first-page":"2142","DOI":"10.1109\/TKDE.2016.2557324","article-title":"Interactive visualization of large data sets","volume":"28","author":"Godfrey","year":"2016","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"10.1016\/j.visinf.2025.100271_b25","series-title":"Deep Learning","author":"Goodfellow","year":"2016"},{"issue":"3","key":"10.1016\/j.visinf.2025.100271_b26","doi-asserted-by":"crossref","first-page":"90","DOI":"10.1109\/MCG.2016.59","article-title":"Data-driven healthcare: challenges and opportunities for interactive visualization","volume":"36","author":"Gotz","year":"2016","journal-title":"IEEE Comput. Graph. Appl."},{"key":"10.1016\/j.visinf.2025.100271_b27","doi-asserted-by":"crossref","unstructured":"Guan,\u00a0T., Liu,\u00a0F., Wu,\u00a0X., Xian,\u00a0R., Li,\u00a0Z., Liu,\u00a0X., Wang,\u00a0X., Chen,\u00a0L., Huang,\u00a0F., Yacoob,\u00a0Y., et al., 2024. HallusionBench: An Advanced Diagnostic Suite for Entangled Language Hallucination and Visual Illusion in Large Vision-Language Models. In: Proc. ICCV. pp. 14375\u201314385.","DOI":"10.1109\/CVPR52733.2024.01363"},{"issue":"2","key":"10.1016\/j.visinf.2025.100271_b28","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3643894","article-title":"Talk2Data: A natural language interface for exploratory visual analysis via question decomposition","volume":"14","author":"Guo","year":"2024","journal-title":"ACM Trans. Interact. Intell. Syst."},{"key":"10.1016\/j.visinf.2025.100271_b29","series-title":"ChartLlama: A multimodal LLM for chart understanding and generation","author":"Han","year":"2023"},{"key":"10.1016\/j.visinf.2025.100271_b30","series-title":"VisTR: Visualizations as representations for time-series table reasoning","author":"Hao","year":"2024"},{"key":"10.1016\/j.visinf.2025.100271_b31","first-page":"1","article-title":"FinFlier: Automating graphical overlays for financial visualizations with knowledge-grounding large language model","author":"Hao","year":"2024","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"3","key":"10.1016\/j.visinf.2025.100271_b32","doi-asserted-by":"crossref","first-page":"555","DOI":"10.1111\/cgf.14573","article-title":"Chart question answering: State of the art and future directions","volume":"41","author":"Hoque","year":"2022","journal-title":"Comput. Graph. Forum"},{"key":"10.1016\/j.visinf.2025.100271_b33","doi-asserted-by":"crossref","unstructured":"Hu,\u00a0K., Bakker,\u00a0M.A., Li,\u00a0S., Kraska,\u00a0T., Hidalgo,\u00a0C., 2019. VizML: A Machine Learning Approach to Visualization Recommendation. In: Proc. ACM CHI. pp. 1\u201312.","DOI":"10.1145\/3290605.3300358"},{"key":"10.1016\/j.visinf.2025.100271_b34","doi-asserted-by":"crossref","unstructured":"Hu,\u00a0L., Wang,\u00a0D., Pan,\u00a0Y., Yu,\u00a0J., Shao,\u00a0Y., Feng,\u00a0C., Nie,\u00a0L., 2024. NovaChart: A Large-scale Dataset towards Chart Understanding and Generation of Multimodal Large Language Models. In: Proc. ACMMM. pp. 3917\u20133925.","DOI":"10.1145\/3664647.3680790"},{"key":"10.1016\/j.visinf.2025.100271_b35","doi-asserted-by":"crossref","unstructured":"Idreos,\u00a0S., Papaemmanouil,\u00a0O., Chaudhuri,\u00a0S., 2015. Overview of data exploration techniques. In: Proc. ACM SIGMOD. pp. 277\u2013281.","DOI":"10.1145\/2723372.2731084"},{"issue":"6245","key":"10.1016\/j.visinf.2025.100271_b36","doi-asserted-by":"crossref","first-page":"255","DOI":"10.1126\/science.aaa8415","article-title":"Machine learning: Trends, perspectives, and prospects","volume":"349","author":"Jordan","year":"2015","journal-title":"Science"},{"key":"10.1016\/j.visinf.2025.100271_b37","doi-asserted-by":"crossref","unstructured":"Jung,\u00a0D., Kim,\u00a0W., Song,\u00a0H., Hwang,\u00a0J.-i., Lee,\u00a0B., Kim,\u00a0B., Seo,\u00a0J., 2017. ChartSense: Interactive data extraction from chart images. In: Proc. ACM CHI. pp. 6706\u20136717.","DOI":"10.1145\/3025453.3025957"},{"key":"10.1016\/j.visinf.2025.100271_b38","doi-asserted-by":"crossref","unstructured":"Kafle,\u00a0K., Price,\u00a0B., Cohen,\u00a0S., Kanan,\u00a0C., 2018. DVQA: Understanding Data Visualizations via Question Answering. In: Proc. CVPR. pp. 5648\u20135656.","DOI":"10.1109\/CVPR.2018.00592"},{"key":"10.1016\/j.visinf.2025.100271_b39","doi-asserted-by":"crossref","unstructured":"Kafle,\u00a0K., Shrestha,\u00a0R., Cohen,\u00a0S., Price,\u00a0B., Kanan,\u00a0C., 2020. Answering questions about data visualizations using efficient bimodal fusion. In: Proc. WACV. pp. 1498\u20131507.","DOI":"10.1109\/WACV45572.2020.9093494"},{"key":"10.1016\/j.visinf.2025.100271_b40","series-title":"FigureQA: An annotated figure dataset for visual reasoning","author":"Kahou","year":"2017"},{"key":"10.1016\/j.visinf.2025.100271_b41","doi-asserted-by":"crossref","unstructured":"Kantharaj,\u00a0S., Leong,\u00a0R.T., Lin,\u00a0X., Masry,\u00a0A., Thakkar,\u00a0M., Hoque,\u00a0E., Joty,\u00a0S., 2022. Chart-to-Text: A Large-Scale Benchmark for Chart Summarization. In: Proc. ACL. pp. 4005\u20134023.","DOI":"10.18653\/v1\/2022.acl-long.277"},{"key":"10.1016\/j.visinf.2025.100271_b42","doi-asserted-by":"crossref","unstructured":"Keim,\u00a0D., Andrienko,\u00a0G., Fekete,\u00a0J.-D., G\u00f6rg,\u00a0C., Kohlhammer,\u00a0J., Melan\u00e7on,\u00a0G., 2008. Visual analytics: Definition, process, and challenges. In: Information Visualization: Human-Centered Issues and Perspectives. pp. 154\u2013175.","DOI":"10.1007\/978-3-540-70956-5_7"},{"key":"10.1016\/j.visinf.2025.100271_b43","doi-asserted-by":"crossref","unstructured":"Kim,\u00a0D.H., Hoque,\u00a0E., Agrawala,\u00a0M., 2020. Answering Questions about Charts and Generating Visual Explanations. In: Proc. ACM CHI. pp. 1\u201313.","DOI":"10.1145\/3313831.3376467"},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b44","doi-asserted-by":"crossref","first-page":"51","DOI":"10.1007\/BF00365595","article-title":"The singularities of the visual mapping","volume":"24","author":"Koenderink","year":"1976","journal-title":"Biol. Cybernet."},{"issue":"7553","key":"10.1016\/j.visinf.2025.100271_b45","doi-asserted-by":"crossref","first-page":"436","DOI":"10.1038\/nature14539","article-title":"Deep learning","volume":"521","author":"LeCun","year":"2015","journal-title":"Nature"},{"key":"10.1016\/j.visinf.2025.100271_b46","unstructured":"Lee,\u00a0K., Joshi,\u00a0M., Turc,\u00a0I.R., Hu,\u00a0H., Liu,\u00a0F., Eisenschlos,\u00a0J.M., Khandelwal,\u00a0U., Shaw,\u00a0P., Chang,\u00a0M.-W., Toutanova,\u00a0K., 2023. Pix2Struct: Screenshot Parsing as Pretraining for Visual Language Understanding. In: Proc. ICML. pp. 18893\u201318912."},{"issue":"4","key":"10.1016\/j.visinf.2025.100271_b47","doi-asserted-by":"crossref","DOI":"10.1007\/s00778-025-00912-0","article-title":"Prompt4Vis: Prompting large language models with example mining and schema filtering for tabular data visualization","volume":"34","author":"Li","year":"2025","journal-title":"VLDB J."},{"key":"10.1016\/j.visinf.2025.100271_b48","series-title":"Visualization literacy of multimodal large language models: A comparative study","author":"Li","year":"2024"},{"key":"10.1016\/j.visinf.2025.100271_b49","series-title":"Visualization generation with large language models: An evaluation","author":"Li","year":"2024"},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b50","doi-asserted-by":"crossref","first-page":"195","DOI":"10.1109\/TVCG.2021.3114863","article-title":"KG4Vis: A knowledge graph-based approach for visualization recommendation","volume":"28","author":"Li","year":"2021","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b51","series-title":"Using large language models to generate engaging captions for data visualizations","author":"Liew","year":"2022"},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b52","first-page":"944","article-title":"InkSight: Leveraging sketch interaction for documenting chart findings in computational notebooks","volume":"30","author":"Lin","year":"2023","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"8","key":"10.1016\/j.visinf.2025.100271_b53","doi-asserted-by":"crossref","first-page":"5276","DOI":"10.1109\/TVCG.2023.3290241","article-title":"AutoTitle: An interactive title generator for visualizations","volume":"30","author":"Liu","year":"2024","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b54","doi-asserted-by":"crossref","unstructured":"Liu,\u00a0C., Han,\u00a0Y., Jiang,\u00a0R., Yuan,\u00a0X., 2021. ADVISor: Automatic Visualization Answer for Natural-Language Question on Tabular Data. In: Proc. IEEE PacificVis. pp. 11\u201320.","DOI":"10.1109\/PacificVis52677.2021.00010"},{"key":"10.1016\/j.visinf.2025.100271_b55","doi-asserted-by":"crossref","unstructured":"Liu,\u00a0F., Piccinno,\u00a0F., Krichene,\u00a0S., Pang,\u00a0C., Lee,\u00a0K., Joshi,\u00a0M., Altun,\u00a0Y., Collier,\u00a0N., Eisenschlos,\u00a0J., 2023a. MatCha: Enhancing Visual Language Pretraining with Math Reasoning and Chart Derendering. In: Proc. ACL. pp. 12756\u201312770.","DOI":"10.18653\/v1\/2023.acl-long.714"},{"key":"10.1016\/j.visinf.2025.100271_b56","doi-asserted-by":"crossref","unstructured":"Liu,\u00a0S.-C., Wang,\u00a0S., Chang,\u00a0T., Lin,\u00a0W., Hsiung,\u00a0C.-W., Hsieh,\u00a0Y.-C., Cheng,\u00a0Y.-P., Luo,\u00a0S.-H., Zhang,\u00a0J., 2023b. JarviX: A LLM No code Platform for Tabular Data Analysis and Optimization. In: Proc. EMNLP. pp. 622\u2013630.","DOI":"10.18653\/v1\/2023.emnlp-industry.59"},{"key":"10.1016\/j.visinf.2025.100271_b57","doi-asserted-by":"crossref","unstructured":"Liu,\u00a0F., Wang,\u00a0X., Yao,\u00a0W., Chen,\u00a0J., Song,\u00a0K., Cho,\u00a0S., Yacoob,\u00a0Y., Yu,\u00a0D., 2024b. MMC: Advancing Multimodal Chart Understanding with Large-scale Instruction Tuning. In: Proc. NAACL. pp. 1287\u20131310.","DOI":"10.18653\/v1\/2024.naacl-long.70"},{"key":"10.1016\/j.visinf.2025.100271_b58","doi-asserted-by":"crossref","unstructured":"Liu,\u00a0C., Xie,\u00a0L., Han,\u00a0Y., Wei,\u00a0D., Yuan,\u00a0X., 2020. AutoCaption: An approach to generate natural language description from visualization automatically. In: Proc. IEEE PacificVis. pp. 191\u2013195.","DOI":"10.1109\/PacificVis48177.2020.1043"},{"key":"10.1016\/j.visinf.2025.100271_b59","doi-asserted-by":"crossref","unstructured":"Luo,\u00a0Y., Qin,\u00a0X., Tang,\u00a0N., Li,\u00a0G., 2018. DeepEye: Towards Automatic Data Visualization. In: Proc. IEEE ICDE. pp. 101\u2013112.","DOI":"10.1109\/ICDE.2018.00019"},{"key":"10.1016\/j.visinf.2025.100271_b60","doi-asserted-by":"crossref","unstructured":"Ma,\u00a0P., Ding,\u00a0R., Wang,\u00a0S., Han,\u00a0S., Zhang,\u00a0D., 2023. InsightPilot: An LLM-Empowered Automated Data Exploration System. In: Proc. EMNLP. pp. 346\u2013352.","DOI":"10.18653\/v1\/2023.emnlp-demo.31"},{"issue":"12","key":"10.1016\/j.visinf.2025.100271_b61","doi-asserted-by":"crossref","first-page":"6345","DOI":"10.1007\/s00371-022-02733-6","article-title":"Uncertainty-aware visual analytics: scope, opportunities, and challenges","volume":"39","author":"Maack","year":"2023","journal-title":"Vis. Comput."},{"issue":"2","key":"10.1016\/j.visinf.2025.100271_b62","doi-asserted-by":"crossref","first-page":"110","DOI":"10.1145\/22949.22950","article-title":"Automating the design of graphical presentations of relational information","volume":"5","author":"Mackinlay","year":"1986","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.visinf.2025.100271_b63","doi-asserted-by":"crossref","first-page":"45181","DOI":"10.1109\/ACCESS.2023.3274199","article-title":"Chat2VIS: Generating data visualizations via natural language using ChatGPT, codex and GPT-3 large language models","volume":"11","author":"Maddigan","year":"2023","journal-title":"IEEE Access"},{"key":"10.1016\/j.visinf.2025.100271_b64","series-title":"GPTEval: A survey on assessments of ChatGPT and GPT-4","author":"Mao","year":"2023"},{"key":"10.1016\/j.visinf.2025.100271_b65","unstructured":"Masry,\u00a0A., Hoque,\u00a0E., 2021. Integrating image data extraction and table parsing methods for chart question answering. In: Proc. CVPR ChartQA Workshop. pp. 1\u20135."},{"key":"10.1016\/j.visinf.2025.100271_b66","doi-asserted-by":"crossref","unstructured":"Masry,\u00a0A., Kavehzadeh,\u00a0P., Do,\u00a0X.L., Hoque,\u00a0E., Joty,\u00a0S., 2023. UniChart: A Universal Vision-language Pretrained Model for Chart Comprehension and Reasoning. In: Proc. EMNLP. pp. 14662\u201314684.","DOI":"10.18653\/v1\/2023.emnlp-main.906"},{"key":"10.1016\/j.visinf.2025.100271_b67","doi-asserted-by":"crossref","unstructured":"Masry,\u00a0A., Long,\u00a0D.X., Tan,\u00a0J.Q., Joty,\u00a0S., Hoque,\u00a0E., 2022. ChartQA: A Benchmark for Question Answering about Charts with Visual and Logical Reasoning. In: Proc. ACL Findings. pp. 2263\u20132279.","DOI":"10.18653\/v1\/2022.findings-acl.177"},{"key":"10.1016\/j.visinf.2025.100271_b68","unstructured":"Masry,\u00a0A., Thakkar,\u00a0M., Bajaj,\u00a0A., Kartha,\u00a0A., Hoque,\u00a0E., Joty,\u00a0S., 2025. ChartGemma: Visual Instruction-tuning for Chart Reasoning in the Wild. In: Proc. COLING. pp. 625\u2013643."},{"key":"10.1016\/j.visinf.2025.100271_b69","doi-asserted-by":"crossref","first-page":"623","DOI":"10.1007\/s10648-015-9342-2","article-title":"How visual displays affect cognitive processing","volume":"29","author":"McCrudden","year":"2017","journal-title":"Educ. Psychol. Rev."},{"key":"10.1016\/j.visinf.2025.100271_b70","doi-asserted-by":"crossref","unstructured":"Meng,\u00a0F., Shao,\u00a0W., Lu,\u00a0Q., Gao,\u00a0P., Zhang,\u00a0K., Qiao,\u00a0Y., Luo,\u00a0P., 2024. ChartAssistant: A Universal Chart Multimodal Language Model via Chart-to-Table Pre-training and Multitask Instruction Tuning. In: Proc. ACL Findings. pp. 7775\u20137803.","DOI":"10.18653\/v1\/2024.findings-acl.463"},{"key":"10.1016\/j.visinf.2025.100271_b71","doi-asserted-by":"crossref","unstructured":"Methani,\u00a0N., Ganguly,\u00a0P., Khapra,\u00a0M.M., Kumar,\u00a0P., 2020. PlotQA: Reasoning over Scientific Plots. In: Proc. WACV. pp. 1527\u20131536.","DOI":"10.1109\/WACV45572.2020.9093523"},{"key":"10.1016\/j.visinf.2025.100271_b72","series-title":"Insight types specification","author":"Microsoft","year":"2016"},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b73","doi-asserted-by":"crossref","first-page":"438","DOI":"10.1109\/TVCG.2018.2865240","article-title":"Formalizing visualization design knowledge as constraints: Actionable and extensible models in draco","volume":"25","author":"Moritz","year":"2018","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b74","series-title":"Visualization Analysis and Design","author":"Munzner","year":"2014"},{"issue":"2","key":"10.1016\/j.visinf.2025.100271_b75","doi-asserted-by":"crossref","first-page":"369","DOI":"10.1109\/TVCG.2020.3030378","article-title":"NL4DV: A toolkit for generating analytic specifications for data visualization from natural language queries","volume":"27","author":"Narechania","year":"2020","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"3","key":"10.1016\/j.visinf.2025.100271_b76","doi-asserted-by":"crossref","first-page":"353","DOI":"10.1111\/cgf.13193","article-title":"Reverse-engineering visualizations: Recovering visual encodings from chart images","volume":"36","author":"Poco","year":"2017","journal-title":"Comput. Graph. Forum"},{"issue":"3","key":"10.1016\/j.visinf.2025.100271_b77","doi-asserted-by":"crossref","first-page":"1203","DOI":"10.1111\/j.1467-8659.2009.01694.x","article-title":"The perception of correlation in scatterplots","volume":"29","author":"Rensink","year":"2010","journal-title":"Comput. Graph. Forum"},{"key":"10.1016\/j.visinf.2025.100271_b78","doi-asserted-by":"crossref","unstructured":"Rule,\u00a0A., Tabard,\u00a0A., Hollan,\u00a0J.D., 2018. Exploration and Explanation in Computational Notebooks. In: Proc. ACM CHI. pp. 1\u201312.","DOI":"10.1145\/3173574.3173606"},{"issue":"3","key":"10.1016\/j.visinf.2025.100271_b79","article-title":"Visual mapping to enhance learning and critical thinking skills","volume":"36","author":"Santiago","year":"2011","journal-title":"Optom. Educ."},{"key":"10.1016\/j.visinf.2025.100271_b80","doi-asserted-by":"crossref","unstructured":"Savva,\u00a0M., Kong,\u00a0N., Chhajta,\u00a0A., Fei-Fei,\u00a0L., Agrawala,\u00a0M., Heer,\u00a0J., 2011. ReVision: Automated classification, analysis and redesign of chart images. In: Proc. ACM UIST. pp. 393\u2013402.","DOI":"10.1145\/2047196.2047247"},{"key":"10.1016\/j.visinf.2025.100271_b81","series-title":"DashChat: Interactive authoring of industrial dashboard design prototypes through conversation with LLM-powered agents","author":"Shen","year":"2025"},{"key":"10.1016\/j.visinf.2025.100271_b82","unstructured":"Shen,\u00a0L., Shen,\u00a0E., Tai,\u00a0Z., Song,\u00a0Y., Wang,\u00a0J., Agus,\u00a0M., Garth,\u00a0C., Kerren,\u00a0A., 2021. TaskVis: Task-oriented Visualization Recommendation. In: Proc. EuroVis. pp. 91\u201395."},{"issue":"2","key":"10.1016\/j.visinf.2025.100271_b83","doi-asserted-by":"crossref","first-page":"453","DOI":"10.1109\/TVCG.2020.3030403","article-title":"Calliope: Automatic visual data story generation from a spreadsheet","volume":"27","author":"Shi","year":"2020","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b84","author":"Sutton","year":"1998"},{"issue":"03","key":"10.1016\/j.visinf.2025.100271_b85","doi-asserted-by":"crossref","first-page":"1731","DOI":"10.1109\/TVCG.2024.3368621","article-title":"Chartgpt: Leveraging LLMs to generate charts from abstract natural language","volume":"31","author":"Tian","year":"2025","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b86","series-title":"The Visual Display of Quantitative Information","author":"Tufte","year":"1983"},{"key":"10.1016\/j.visinf.2025.100271_b87","series-title":"Exploratory Data Analysis","author":"Tukey","year":"1977"},{"issue":"13","key":"10.1016\/j.visinf.2025.100271_b88","doi-asserted-by":"crossref","first-page":"2182","DOI":"10.14778\/2831360.2831371","article-title":"SeeDB: Efficient data-driven visualization recommendations to support visual analytics","volume":"8","author":"Vartak","year":"2015","journal-title":"Proc. VLDB Endow."},{"key":"10.1016\/j.visinf.2025.100271_b89","series-title":"Hybrid llm\/rule-based approaches to business insights generation from structured data","author":"Vertsel","year":"2024"},{"issue":"6","key":"10.1016\/j.visinf.2025.100271_b90","doi-asserted-by":"crossref","first-page":"3197","DOI":"10.1007\/s11845-023-03377-8","article-title":"GPT-4: a new era of artificial intelligence in medicine","volume":"192","author":"Waisberg","year":"2023","journal-title":"Ir. J. Med. Sci. (1971-)"},{"issue":"12","key":"10.1016\/j.visinf.2025.100271_b91","doi-asserted-by":"crossref","first-page":"5134","DOI":"10.1109\/TVCG.2021.3106142","article-title":"A survey on ML4VIS: Applying machine learning advances to data visualization","volume":"28","author":"Wang","year":"2021","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b92","doi-asserted-by":"crossref","unstructured":"Wang,\u00a0A.Y., Epperson,\u00a0W., DeLine,\u00a0R.A., Drucker,\u00a0S.M., 2022. Diff in the Loop: Supporting Data Comparison in Exploratory Data Analysis. In: Proc. ACM CHI. pp. 1\u201310.","DOI":"10.1145\/3491102.3502123"},{"issue":"01","key":"10.1016\/j.visinf.2025.100271_b93","doi-asserted-by":"crossref","first-page":"710","DOI":"10.1109\/TVCG.2024.3456350","article-title":"DracoGPT: Extracting visualization design preferences from large language models","volume":"31","author":"Wang","year":"2025","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"2","key":"10.1016\/j.visinf.2025.100271_b94","doi-asserted-by":"crossref","first-page":"106","DOI":"10.1016\/j.visinf.2024.06.002","article-title":"AVA: An automated and ai-driven intelligent visual analytics framework","volume":"8","author":"Wang","year":"2024","journal-title":"Vis. Informatics"},{"key":"10.1016\/j.visinf.2025.100271_b95","series-title":"ChartInsighter: An approach for mitigating hallucination in time-series chart summary generation with a benchmark dataset","author":"Wang","year":"2025"},{"issue":"6","key":"10.1016\/j.visinf.2025.100271_b96","doi-asserted-by":"crossref","DOI":"10.1007\/s11704-023-2691-y","article-title":"VIS+ AI: integrating visualization with artificial intelligence for efficient data analysis","volume":"17","author":"Wang","year":"2023","journal-title":"Front. Comput. Sci."},{"key":"10.1016\/j.visinf.2025.100271_b97","doi-asserted-by":"crossref","unstructured":"Wang,\u00a0L., Zhang,\u00a0S., Wang,\u00a0Y., Lim,\u00a0E.-P., Wang,\u00a0Y., 2023b. LLM4Vis: Explainable Visualization Recommendation using ChatGPT. In: Proc. EMNLP. pp. 675\u2013692.","DOI":"10.18653\/v1\/2023.emnlp-industry.64"},{"key":"10.1016\/j.visinf.2025.100271_b98","first-page":"24824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume":"35","author":"Wei","year":"2022","journal-title":"Proc. NeurIPS"},{"key":"10.1016\/j.visinf.2025.100271_b99","series-title":"Mchartqa: A universal benchmark for multimodal chart question answer based on vision-language alignment and reasoning","author":"Wei","year":"2024"},{"issue":"06","key":"10.1016\/j.visinf.2025.100271_b100","doi-asserted-by":"crossref","first-page":"3719","DOI":"10.1109\/TVCG.2025.3567131","article-title":"InsightLens: Augmenting LLM-powered data analysis with interactive insight management and navigation","volume":"31","author":"Weng","year":"2025","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b101","series-title":"ggplot2: elegant graphics for data analysis","author":"Wickham","year":"2009"},{"key":"10.1016\/j.visinf.2025.100271_b102","doi-asserted-by":"crossref","unstructured":"Wilkinson,\u00a0L., 2011. The grammar of graphics. In: Handbook of Computational Statistics: Concepts and Methods. pp. 375\u2013414.","DOI":"10.1007\/978-3-642-21551-3_13"},{"key":"10.1016\/j.visinf.2025.100271_b103","series-title":"Goals, process, and challenges of exploratory data analysis: An interview study","author":"Wongsuphasawat","year":"2019"},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b104","doi-asserted-by":"crossref","first-page":"649","DOI":"10.1109\/TVCG.2015.2467191","article-title":"Voyager: Exploratory analysis via faceted browsing of visualization recommendations","volume":"22","author":"Wongsuphasawat","year":"2015","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b105","doi-asserted-by":"crossref","unstructured":"Wongsuphasawat,\u00a0K., Moritz,\u00a0D., Anand,\u00a0A., Mackinlay,\u00a0J., Howe,\u00a0B., Heer,\u00a0J., 2016. Towards a general-purpose query language for visualization recommendation. In: Proceedings of the Workshop on Human-in-the-Loop Data Analytics. pp. 1\u20136.","DOI":"10.1145\/2939502.2939506"},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b106","first-page":"1026","article-title":"In defence of visual analytics systems: Replies to critics","volume":"29","author":"Wu","year":"2022","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b107","doi-asserted-by":"crossref","unstructured":"Wu,\u00a0J., Gan,\u00a0W., Chen,\u00a0Z., Wan,\u00a0S., Yu,\u00a0P.S., 2023a. Multimodal large language models: A survey. In: Proc. IEEE BigData. pp. 2247\u20132256.","DOI":"10.1109\/BigData59044.2023.10386743"},{"issue":"5","key":"10.1016\/j.visinf.2025.100271_b108","doi-asserted-by":"crossref","first-page":"1122","DOI":"10.1109\/JAS.2023.123618","article-title":"A brief overview of ChatGPT: The history, status quo and potential future development","volume":"10","author":"Wu","year":"2023","journal-title":"IEEE\/CAA J. Autom. Sin."},{"key":"10.1016\/j.visinf.2025.100271_b109","doi-asserted-by":"crossref","unstructured":"Wu,\u00a0T., Terry,\u00a0M., Cai,\u00a0C.J., 2022b. AI Chains: Transparent and Controllable Human-AI Interaction by Chaining Large Language Model Prompts. In: Proc. ACM CHI. pp. 1\u201322.","DOI":"10.1145\/3491102.3517582"},{"issue":"12","key":"10.1016\/j.visinf.2025.100271_b110","doi-asserted-by":"crossref","first-page":"5049","DOI":"10.1109\/TVCG.2021.3099002","article-title":"AI4VIS: Survey on artificial intelligence approaches for data visualization","volume":"28","author":"Wu","year":"2021","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b111","series-title":"StructChart: Perception, structuring, reasoning for visual chart understanding","author":"Xia","year":"2023"},{"key":"10.1016\/j.visinf.2025.100271_b112","series-title":"Chartx & chartvlm: A versatile benchmark and foundation model for complicated chart reasoning","author":"Xia","year":"2024"},{"key":"10.1016\/j.visinf.2025.100271_b113","series-title":"Chartbench: A benchmark for complex visual reasoning in charts","author":"Xu","year":"2023"},{"key":"10.1016\/j.visinf.2025.100271_b114","series-title":"The dawn of LMMs: Preliminary explorations with GPT-4V (ision)","first-page":"1","author":"Yang","year":"2023"},{"key":"10.1016\/j.visinf.2025.100271_b115","series-title":"AskChart: Universal chart understanding through textual enhancement","author":"Yang","year":"2024"},{"issue":"2","key":"10.1016\/j.visinf.2025.100271_b116","doi-asserted-by":"crossref","first-page":"43","DOI":"10.1016\/j.visinf.2024.04.003","article-title":"Generative AI for visualization: State of the art and future directions","volume":"8","author":"Ye","year":"2024","journal-title":"Vis. Informatics"},{"issue":"2","key":"10.1016\/j.visinf.2025.100271_b117","doi-asserted-by":"crossref","first-page":"839","DOI":"10.1109\/TVCG.2020.3030410","article-title":"Revisiting the modifiable areal unit problem in deep traffic prediction with visual analytics","volume":"27","author":"Zeng","year":"2021","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"1","key":"10.1016\/j.visinf.2025.100271_b118","doi-asserted-by":"crossref","first-page":"525","DOI":"10.1109\/TVCG.2024.3456159","article-title":"Advancing multimodal large language models in chart question answering with visualization-referenced instruction tuning","volume":"31","author":"Zeng","year":"2025","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b119","doi-asserted-by":"crossref","unstructured":"Zhang,\u00a0L., Hu,\u00a0A., Xu,\u00a0H., Yan,\u00a0M., Xu,\u00a0Y., Jin,\u00a0Q., Zhang,\u00a0J., Huang,\u00a0F., 2024. TinyChart: Efficient Chart Understanding with Program-of-Thoughts Learning and Visual Token Merging. In: Proc. EMNLP. pp. 1882\u20131898.","DOI":"10.18653\/v1\/2024.emnlp-main.112"},{"key":"10.1016\/j.visinf.2025.100271_b120","first-page":"1","article-title":"Lightva: Lightweight visual analytics with LLM agent-based task planning and execution","author":"Zhao","year":"2024","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"issue":"3","key":"10.1016\/j.visinf.2025.100271_b121","doi-asserted-by":"crossref","first-page":"1830","DOI":"10.1109\/TVCG.2024.3368060","article-title":"LEVA: Using large language models to enhance visual analytics","volume":"31","author":"Zhao","year":"2025","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"10.1016\/j.visinf.2025.100271_b122","first-page":"1","article-title":"ChartKG: A knowledge-graph-based representation for chart images","author":"Zhou","year":"2024","journal-title":"IEEE Trans. Vis. Comput. Graphics"}],"container-title":["Visual Informatics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2468502X25000543?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2468502X25000543?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,9]],"date-time":"2026-03-09T19:13:35Z","timestamp":1773083615000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S2468502X25000543"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12]]},"references-count":122,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["S2468502X25000543"],"URL":"https:\/\/doi.org\/10.1016\/j.visinf.2025.100271","relation":{},"ISSN":["2468-502X"],"issn-type":[{"value":"2468-502X","type":"print"}],"subject":[],"published":{"date-parts":[[2025,12]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"A survey of visual insight mining: Connecting data and insights via visualization","name":"articletitle","label":"Article Title"},{"value":"Visual Informatics","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.visinf.2025.100271","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 The Author(s). Published by Elsevier B.V. on behalf of Zhejiang University and Zhejiang University Press.","name":"copyright","label":"Copyright"}],"article-number":"100271"}}