{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,4]],"date-time":"2025-11-04T11:18:33Z","timestamp":1762255113890,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":49,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032083265","type":"print"},{"value":"9783032083272","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T00:00:00Z","timestamp":1760227200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T00:00:00Z","timestamp":1760227200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>Graph Neural Networks (GNNs) are a powerful technique for machine learning on graph-structured data, yet they pose challenges in interpretability. Existing GNN explanation methods usually yield technical outputs, such as subgraphs and feature importance scores, that are difficult to understand, thereby violating the purpose of explanations. Motivated by recent Explainable AI (XAI) research, we propose <jats:italic>GraphXAIN<\/jats:italic>, a method that generates natural language narratives explaining GNN prediction. GraphXAIN is a model- and explainer-agnostic method that uses Large Language Models (LLMs) to translate explanatory subgraphs and feature importance scores into coherent, story-like explanations of GNN predictions. Evaluation on real-world datasets demonstrates GraphXAIN\u2019s ability to improve graph explanations. A survey of machine learning researchers and practitioners reveals that GraphXAIN enhances four explainability dimensions: understandability, satisfaction, convincingness, and suitability for communicating model predictions. When combined with another graph explainer method, GraphXAIN further improves trustworthiness, insightfulness, confidence, and usability. Notably, 95% of participants find GraphXAIN to be a valuable addition to the current GNN explanation methods. By incorporating natural language narratives, our approach serves both graph practitioners and non-expert users by providing clearer and more effective explanations.\n<\/jats:p>","DOI":"10.1007\/978-3-032-08327-2_5","type":"book-chapter","created":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T18:07:10Z","timestamp":1760206030000},"page":"91-114","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["GraphXAIN: Narratives to\u00a0Explain Graph Neural Networks"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-1267-4263","authenticated-orcid":false,"given":"Mateusz","family":"Cedro","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8397-2937","authenticated-orcid":false,"given":"David","family":"Martens","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,12]]},"reference":[{"key":"5_CR1","unstructured":"Adebayo, J., Gilmer, J., Muelly, M., Goodfellow, I., Hardt, M., Kim, B.: Sanity checks for saliency maps. Adv. Neural Inf. Process. Syst. 31 (2018)"},{"issue":"5","key":"5_CR2","doi-asserted-by":"publisher","first-page":"2596","DOI":"10.1007\/s10618-023-00924-w","volume":"38","author":"H Baniecki","year":"2024","unstructured":"Baniecki, H., Parzych, D., Biecek, P.: The grammar of interactive explanatory model analysis. Data Min. Knowl. Disc. 38(5), 2596\u20132632 (2024)","journal-title":"Data Min. Knowl. Disc."},{"key":"5_CR3","unstructured":"Biecek, P., Samek, W.: Position: explain to question not to justify. In: Forty-First International Conference on Machine Learning (2024)"},{"issue":"5","key":"5_CR4","doi-asserted-by":"publisher","first-page":"2665","DOI":"10.1007\/s10618-023-00930-y","volume":"38","author":"D Brughmans","year":"2024","unstructured":"Brughmans, D., Leyman, P., Martens, D.: Nice: an algorithm for nearest instance counterfactual explanations. Data Min. Knowl. Disc. 38(5), 2665\u20132703 (2024)","journal-title":"Data Min. Knowl. Disc."},{"key":"5_CR5","unstructured":"Cedro, M., Chlebus, M.: Beyond the black box: do more complex deep learning models provide superior XAI explanations? arXiv preprint arXiv:2405.08658 (2024)"},{"key":"5_CR6","doi-asserted-by":"publisher","first-page":"13614","DOI":"10.1073\/pnas.1320645111","volume":"111","author":"MF Dahlstrom","year":"2014","unstructured":"Dahlstrom, M.F.: Using narratives and storytelling to communicate science with nonexpert audiences. Proc. Natl. Acad. Sci. 111, 13614\u201313620 (2014)","journal-title":"Proc. Natl. Acad. Sci."},{"key":"5_CR7","doi-asserted-by":"crossref","unstructured":"Dahlstrom, M.F., Scheufele, D.A.: (Escaping) the paradox of scientific storytelling. PLoS Biol. 16(10) (2018)","DOI":"10.1371\/journal.pbio.2006720"},{"key":"5_CR8","unstructured":"Fatemi, B., Halcrow, J., Perozzi, B.: Talk like a graph: encoding graphs for large language models. arXiv preprint arXiv:2310.04560 (2023)"},{"key":"5_CR9","doi-asserted-by":"publisher","first-page":"65","DOI":"10.1613\/jair.5477","volume":"61","author":"A Gatt","year":"2018","unstructured":"Gatt, A., Krahmer, E.: Survey of the state of the art in natural language generation: core tasks, applications and evaluation. J. Artif. Intell. Res. 61, 65\u2013170 (2018)","journal-title":"J. Artif. Intell. Res."},{"key":"5_CR10","unstructured":"Giorgi, F., Campagnano, C., Silvestri, F., Tolomei, G.: Natural language counterfactual explanations for graphs using large language models. arXiv preprint arXiv:2410.09295 (2024)"},{"key":"5_CR11","doi-asserted-by":"crossref","unstructured":"Goethals, S., Martens, D., Evgeniou, T.: Manipulation risks in Explainable AI: the implications of the disagreement problem. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 185\u2013200. Springer (2023)","DOI":"10.1007\/978-3-031-74633-8_12"},{"key":"5_CR12","doi-asserted-by":"crossref","unstructured":"Grice, H.: Logic and conversation. Syntax Semant. 3 (1975)","DOI":"10.1163\/9789004368811_003"},{"issue":"5","key":"5_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3236009","volume":"51","author":"R Guidotti","year":"2018","unstructured":"Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., Pedreschi, D.: A survey of methods for explaining black box models. ACM Comput. Surv. (CSUR) 51(5), 1\u201342 (2018)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"5_CR14","unstructured":"He, X., et al.: G-retriever: retrieval-augmented generation for textual graph understanding and question answering. arXiv preprint arXiv:2402.07630 (2024)"},{"key":"5_CR15","doi-asserted-by":"crossref","unstructured":"He, Y., Zheng, Z., Soga, P., Zhu, Y., Dong, Y., Li, J.: Explaining graph neural networks with large language models: a counterfactual perspective on molecule graphs. In: Findings of the Association for Computational Linguistics: EMNLP 2024, pp. 7079\u20137096 (2024)","DOI":"10.18653\/v1\/2024.findings-emnlp.415"},{"key":"5_CR16","unstructured":"Hinns, J., Martens, D.: Exposing image classifier shortcuts with counterfactual frequency (CoF) tables. arXiv preprint arXiv:2405.15661 (2024)"},{"key":"5_CR17","unstructured":"Hoffman, R.R., Mueller, S.T., Klein, G., Litman, J.: Metrics for explainable AI: challenges and prospects. arXiv preprint arXiv:1812.04608 (2018)"},{"key":"5_CR18","doi-asserted-by":"crossref","unstructured":"Holzinger, A., Saranti, A., Molnar, C., Biecek, P., Samek, W.: Explainable AI methods-a brief overview. In: International Workshop on Extending Explainable AI Beyond Deep Models and Classifiers, pp. 13\u201338. Springer (2020)","DOI":"10.1007\/978-3-031-04083-2_2"},{"key":"5_CR19","unstructured":"Ichmoukhamedov, T., Hinns, J., Martens, D.: How good is my story? Towards quantitative metrics for evaluating llm-generated XAI narratives. arXiv preprint arXiv:2412.10220 (2024)"},{"issue":"4","key":"5_CR20","doi-asserted-by":"publisher","first-page":"527","DOI":"10.1287\/isre.1080.0198","volume":"20","author":"U Kayande","year":"2009","unstructured":"Kayande, U., De Bruyn, A., Lilien, G.L., Rangaswamy, A., Van Bruggen, G.H.: How incorporating feedback mechanisms in a DSS affects DSS evaluations. Inf. Syst. Res. 20(4), 527\u2013546 (2009)","journal-title":"Inf. Syst. Res."},{"key":"5_CR21","unstructured":"Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016)"},{"key":"5_CR22","series-title":"Communications in Computer and Information Science","doi-asserted-by":"publisher","first-page":"447","DOI":"10.1007\/978-3-030-65965-3_30","volume-title":"ECML PKDD 2020 Workshops","author":"M Ku\u017aba","year":"2020","unstructured":"Ku\u017aba, M., Biecek, P.: What would you ask the machine learning model? Identification of user needs for model explanations based on human-model conversations. In: Koprinska, I., Gulla, J.A. (eds.) ECML PKDD 2020. CCIS, vol. 1323, pp. 447\u2013459. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-65965-3_30"},{"issue":"5","key":"5_CR23","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3696444","volume":"57","author":"A Longa","year":"2025","unstructured":"Longa, A., et al.: Explaining the explainers in graph neural networks: a comparative study. ACM Comput. Surv. 57(5), 1\u201337 (2025)","journal-title":"ACM Comput. Surv."},{"key":"5_CR24","unstructured":"Loshchilov, I.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"5_CR25","unstructured":"Lucic, A., Ter\u00a0Hoeve, M.A., Tolomei, G., De\u00a0Rijke, M., Silvestri, F.: CF-GNNExplainer: counterfactual explanations for graph neural networks. In: International Conference on Artificial Intelligence and Statistics, pp. 4499\u20134511. PMLR (2022)"},{"key":"5_CR26","unstructured":"Lundberg, S.: A unified approach to interpreting model predictions. arXiv preprint arXiv:1705.07874 (2017)"},{"key":"5_CR27","unstructured":"Madumal, P., Miller, T., Sonenberg, L., Vetere, F.: A grounded interaction protocol for explainable artificial intelligence. arXiv preprint arXiv:1903.02409 (2019)"},{"key":"5_CR28","doi-asserted-by":"crossref","unstructured":"Martens, D., Hinns, J., Dams, C., Vergouwen, M., Evgeniou, T.: Tell me a story! Narrative-driven XAI with large language models. Decis. Support Syst., 114402 (2025)","DOI":"10.1016\/j.dss.2025.114402"},{"issue":"1","key":"5_CR29","doi-asserted-by":"publisher","first-page":"73","DOI":"10.25300\/MISQ\/2014\/38.1.04","volume":"38","author":"D Martens","year":"2014","unstructured":"Martens, D., Provost, F.: Explaining data-driven document classifications. MIS Q. 38(1), 73\u2013100 (2014)","journal-title":"MIS Q."},{"key":"5_CR30","doi-asserted-by":"crossref","unstructured":"Michalski, R.S.: A theory and methodology of inductive learning. In: Machine Learning, pp. 83\u2013134. Elsevier (1983)","DOI":"10.1016\/B978-0-08-051054-5.50008-X"},{"issue":"2","key":"5_CR31","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1037\/h0043158","volume":"63","author":"GA Miller","year":"1956","unstructured":"Miller, G.A.: The magical number seven, plus or minus two: some limits on our capacity for processing information. Psychol. Rev. 63(2), 81 (1956)","journal-title":"Psychol. Rev."},{"key":"5_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.artint.2018.07.007","volume":"267","author":"T Miller","year":"2019","unstructured":"Miller, T.: Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1\u201338 (2019)","journal-title":"Artif. Intell."},{"key":"5_CR33","unstructured":"Miller, T., Howe, P., Sonenberg, L.: Explainable AI: beware of inmates running the asylum or: how i learnt to stop worrying and love the social and behavioural sciences. arXiv preprint arXiv:1712.00547 (2017)"},{"issue":"3\u20134","key":"5_CR34","first-page":"1","volume":"11","author":"S Mohseni","year":"2021","unstructured":"Mohseni, S., Zarei, N., Ragan, E.D.: A multidisciplinary survey and framework for design and evaluation of Explainable AI systems. ACM Trans. Interact. Intell. Syst. (TiiS) 11(3\u20134), 1\u201345 (2021)","journal-title":"ACM Trans. Interact. Intell. Syst. (TiiS)"},{"key":"5_CR35","unstructured":"Molnar, C.: Interpretable Machine Learning, 2 edn. (2022). https:\/\/christophm.github.io\/interpretable-ml-book"},{"key":"5_CR36","doi-asserted-by":"crossref","unstructured":"Pan, B., Xiong, Z., Wu, G., Zhang, Z., Zhang, Y., Zhao, L.: TagExplainer: narrating graph explanations for text-attributed graph learning models. arXiv preprint arXiv:2410.15268 (2024)","DOI":"10.1145\/3627673.3679830"},{"key":"5_CR37","doi-asserted-by":"crossref","unstructured":"Puduppully, R., Dong, L., Lapata, M.: Data-to-text generation with content selection and planning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a033, pp. 6908\u20136915 (2019)","DOI":"10.1609\/aaai.v33i01.33016908"},{"key":"5_CR38","unstructured":"Ribeiro, J., Cardoso, L., Santos, V., Carvalho, E., Carneiro, N., Alves, R.: How reliable and stable are explanations of XAI methods? arXiv preprint arXiv:2407.03108 (2024)"},{"key":"5_CR39","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., Guestrin, C.: \u201cWhy should I trust you?\u201d explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135\u20131144 (2016)","DOI":"10.1145\/2939672.2939778"},{"issue":"1","key":"5_CR40","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1109\/TNN.2008.2005605","volume":"20","author":"F Scarselli","year":"2008","unstructured":"Scarselli, F., Gori, M., Tsoi, A.C., Hagenbuchner, M., Monfardini, G.: The graph neural network model. IEEE Trans. Neural Networks 20(1), 61\u201380 (2008)","journal-title":"IEEE Trans. Neural Networks"},{"key":"5_CR41","unstructured":"Simonyan, K., Vedaldi, A., Zisserman, A.: Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034 (2013)"},{"key":"5_CR42","doi-asserted-by":"crossref","unstructured":"Sokol, K., Flach, P.A.: Conversational explanations of machine learning predictions through class-contrastive counterfactual statements. In: IJCAI, pp. 5785\u20135786 (2018)","DOI":"10.24963\/ijcai.2018\/836"},{"key":"5_CR43","unstructured":"Sundararajan, M., Taly, A., Yan, Q.: Axiomatic attribution for deep networks. In: International Conference on Machine Learning, pp. 3319\u20133328. PMLR (2017)"},{"key":"5_CR44","unstructured":"Veli\u010dkovi\u0107, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks. arXiv preprint arXiv:1710.10903 (2017)"},{"key":"5_CR45","first-page":"841","volume":"31","author":"S Wachter","year":"2017","unstructured":"Wachter, S., Mittelstadt, B., Russell, C.: Counterfactual explanations without opening the black box: automated decisions and the GDPR. Harv. JL Tech. 31, 841 (2017)","journal-title":"Harv. JL Tech."},{"key":"5_CR46","unstructured":"Xu, K., Hu, W., Leskovec, J., Jegelka, S.: How powerful are graph neural networks? arXiv preprint arXiv:1810.00826 (2018)"},{"key":"5_CR47","unstructured":"Ying, Z., Bourgeois, D., You, J., Zitnik, M., Leskovec, J.: GNNExplainer: generating explanations for graph neural networks. Adv. Neural Inf. Process. Syst. 32 (2019)"},{"key":"5_CR48","doi-asserted-by":"crossref","unstructured":"Yousuf, R.B., Defelice, N., Sharma, M., Xu, S., Ramakrishnan, N.: LLM augmentations to support analytical reasoning over multiple documents. In: 2024 IEEE International Conference on Big Data (BigData), pp. 1892\u20131901. IEEE (2024)","DOI":"10.1109\/BigData62323.2024.10826114"},{"key":"5_CR49","unstructured":"Zytek, A., Pid\u00f2, S., Veeramachaneni, K.: LLMs for XAI: future directions for explaining explanations. arXiv preprint arXiv:2405.06064 (2024)"}],"container-title":["Communications in Computer and Information Science","Explainable Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-08327-2_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T18:07:16Z","timestamp":1760206036000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-08327-2_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,12]]},"ISBN":["9783032083265","9783032083272"],"references-count":49,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-08327-2_5","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,12]]},"assertion":[{"value":"12 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The data and code used in this research, as well as the survey results, are available at: .","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Data and Code Availability"}},{"value":"This study involved a voluntary and anonymous online survey to evaluate the proposed explainable AI technique. No sensitive and personally identifiable information was collected. Informed consent was obtained via a consent form presented at the beginning of the survey, which informed participants that participation was voluntary and anonymous and that they could withdraw at any time.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Considerations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"xAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"World Conference on Explainable Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Istanbul","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"T\u00fcrkiye","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 July 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 July 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"xai2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/xaiworldconference.com\/2025\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}