{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T12:14:40Z","timestamp":1775132080543,"version":"3.50.1"},"reference-count":23,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T00:00:00Z","timestamp":1770249600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T00:00:00Z","timestamp":1770249600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s00530-025-02150-4","type":"journal-article","created":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T13:15:59Z","timestamp":1770297359000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Quantifying Factual Divergence in Generative Models: SHAP-LIME Based Hallucination Score for LLMs"],"prefix":"10.1007","volume":"32","author":[{"given":"Ijazul","family":"Haq","sequence":"first","affiliation":[]},{"given":"Muhammad","family":"Saqib","sequence":"additional","affiliation":[]},{"given":"Yingjie","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Irfan Ali","family":"Khan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,5]]},"reference":[{"issue":"4","key":"2150_CR1","doi-asserted-by":"publisher","DOI":"10.1098\/rsos.241776","volume":"12","author":"U Peters","year":"2025","unstructured":"Peters, U., Chin-Yee, B.: Generalization bias in large language model summarization of scientific research. Royal Society Open Science. 12(4), 241776 (2025)","journal-title":"Royal Society Open Science."},{"key":"2150_CR2","doi-asserted-by":"crossref","unstructured":"Gwon, H.J., Yoo, Y., Lee, J.Y.: Improving Generative Ai Reliability: A Qualitative Study of Hallucination in Large Language Models. Available at SSRN 5118420. (2025);","DOI":"10.2139\/ssrn.5118420"},{"key":"2150_CR3","doi-asserted-by":"crossref","unstructured":"Samek, W., M\u00fcller, K.R.: Towards explainable artificial intelligence. In: Explainable AI: interpreting, explaining and visualizing deep learning. Springer; (2019). p. 5\u201322","DOI":"10.1007\/978-3-030-28954-6_1"},{"issue":"1","key":"2150_CR4","doi-asserted-by":"publisher","first-page":"2400304","DOI":"10.1002\/aisy.202400304","volume":"7","author":"AM Salih","year":"2025","unstructured":"Salih, A.M., Raisi-Estabragh, Z., Galazzo, I.B., Radeva, P., Petersen, S.E., Lekadir, K., et al.: A perspective on explainable artificial intelligence methods: SHAP and LIME. Advanced Intelligent Systems. 7(1), 2400304 (2025)","journal-title":"Advanced Intelligent Systems."},{"key":"2150_CR5","doi-asserted-by":"crossref","unstructured":"Alkaissi, H., McFarlane, S.I.: Artificial hallucinations in ChatGPT: implications in scientific writing. Cureus. 2023;15(2)","DOI":"10.7759\/cureus.35179"},{"issue":"12","key":"2150_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3571730","volume":"55","author":"Z Ji","year":"2023","unstructured":"Ji, Z., Lee, N., Frieske, R., Yu, T., Su, D., Xu, Y., et al.: Survey of hallucination in natural language generation. ACM Comput. Surv. 55(12), 1\u201338 (2023)","journal-title":"ACM Comput. Surv."},{"issue":"8","key":"2150_CR7","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al.: Language models are unsupervised multitask learners. OpenAI blog. 1(8), 9 (2019)","journal-title":"OpenAI blog."},{"issue":"240","key":"2150_CR8","first-page":"1","volume":"24","author":"A Chowdhery","year":"2023","unstructured":"Chowdhery, A., Narang, S., Devlin, J., Bosma, M., Mishra, G., Roberts, A., et al.: Palm: Scaling language modeling with pathways. J. Mach. Learn. Res. 24(240), 1\u2013113 (2023)","journal-title":"J. Mach. Learn. Res."},{"issue":"70","key":"2150_CR9","first-page":"1","volume":"25","author":"HW Chung","year":"2024","unstructured":"Chung, H.W., Hou, L., Longpre, S., Zoph, B., Tay, Y., Fedus, W., et al.: Scaling instruction-finetuned language models. J. Mach. Learn. Res. 25(70), 1\u201353 (2024)","journal-title":"J. Mach. Learn. Res."},{"key":"2150_CR10","doi-asserted-by":"crossref","unstructured":"Li, W., Gendron, C., Toxtli, C.: Infusing Human Feedback into Intermediate Prompting Steps of Large Language Models. In: World Congress in Computer Science, Computer Engineering & Applied Computing. Springer; (2024). p. 52\u201364","DOI":"10.1007\/978-3-031-85628-0_4"},{"key":"2150_CR11","unstructured":"Tanneru, S.H., Ley, D., Agarwal, C., Lakkaraju, H.: On the Difficulty of Faithful Chain-of-Thought Reasoning in Large Language Models. In: Proceedings of the Trustworthy Multi-modal Foundation Models and AI Agents (TiFA) Workshop; (2024). p. n.pag. To appear"},{"issue":"12","key":"2150_CR12","first-page":"17","volume":"8","author":"M Abdelrahman","year":"2024","unstructured":"Abdelrahman, M.: Hallucination in Low-Resource Languages: Amplified Risks and Mitigation Strategies for Multilingual LLMs. Journal of Applied Big Data Analytics, Decision-Making, and Predictive Modelling Systems. 8(12), 17\u201324 (2024)","journal-title":"Journal of Applied Big Data Analytics, Decision-Making, and Predictive Modelling Systems."},{"key":"2150_CR13","doi-asserted-by":"publisher","first-page":"391","DOI":"10.1162\/tacl_a_00373","volume":"9","author":"AR Fabbri","year":"2021","unstructured":"Fabbri, A.R., Kry\u015bci\u0144ski, W., McCann, B., Xiong, C., Socher, R., Radev, D.: Summeval: Re-evaluating summarization evaluation. Transactions of the Association for Computational Linguistics. 9, 391\u2013409 (2021)","journal-title":"Transactions of the Association for Computational Linguistics."},{"key":"2150_CR14","first-page":"9459","volume":"33","author":"P Lewis","year":"2020","unstructured":"Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin, V., Goyal, N., et al.: Retrieval-augmented generation for knowledge-intensive nlp tasks. Adv. Neural. Inf. Process. Syst. 33, 9459\u20139474 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"1","key":"2150_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10791-025-09683-2","volume":"28","author":"RT Sairaj","year":"2025","unstructured":"Sairaj, R.T., Balasundaram, S.: Ensemble learning with RAG model to reduce redundant question topics in auto-generated exam questions. Discover Computing. 28(1), 1\u201317 (2025)","journal-title":"Discover Computing."},{"key":"2150_CR16","doi-asserted-by":"crossref","unstructured":"Tharaniya\u00a0Sairaj, R., Balasundaram, S.: Ontology Mapping for Retrieval Augmented Modelling to Reduce Factual Hallucinations in pretrained language Model-Based Auto-Generated Questions. Applied Ontology. 2025;p. 15705838251343009","DOI":"10.1177\/15705838251343009"},{"key":"2150_CR17","doi-asserted-by":"crossref","unstructured":"Al\u00a0Bataineh, A., Sickler, R., Kurcz, K., Pedersen, K.: AI-generated vs. human text: Introducing a new dataset for benchmarking and analysis. IEEE Transactions on Artificial Intelligence. (2025);","DOI":"10.1109\/TAI.2025.3544183"},{"key":"2150_CR18","unstructured":"Lundberg, S.M., Lee, S.I.: A unified approach to interpreting model predictions. Advances in neural information processing systems. (2017);30"},{"key":"2150_CR19","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., Guestrin, C.: \" Why should i trust you?\" Explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining; (2016). p. 1135\u20131144","DOI":"10.1145\/2939672.2939778"},{"key":"2150_CR20","doi-asserted-by":"crossref","unstructured":"Lei, T., Barzilay, R., Jaakkola, T.: Rationalizing neural predictions. arXiv preprint arXiv:1606.04155. (2016);","DOI":"10.18653\/v1\/D16-1011"},{"issue":"4","key":"2150_CR21","doi-asserted-by":"publisher","first-page":"777","DOI":"10.1162\/coli_a_00486","volume":"49","author":"H Rashkin","year":"2023","unstructured":"Rashkin, H., Nikolaev, V., Lamm, M., Aroyo, L., Collins, M., Das, D., et al.: Measuring attribution in natural language generation models. Comput. Linguist. 49(4), 777\u2013840 (2023)","journal-title":"Comput. Linguist."},{"key":"2150_CR22","first-page":"24466","volume":"36","author":"S Yang","year":"2023","unstructured":"Yang, S., Zhang, S., Xia, C., Feng, Y., Xiong, C., Zhou, M.: Preference-grounded token-level guidance for language model fine-tuning. Adv. Neural. Inf. Process. Syst. 36, 24466\u201324496 (2023)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2150_CR23","doi-asserted-by":"crossref","unstructured":"Shankar, S., Zamfirescu-Pereira, J., Hartmann, B., Parameswaran, A., Arawjo, I.: Who validates the validators? aligning llm-assisted evaluation of llm outputs with human preferences. In: Proceedings of the 37th Annual ACM Symposium on User Interface Software and Technology; (2024). p. 1\u201314","DOI":"10.1145\/3654777.3676450"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02150-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-02150-4","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02150-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T11:37:49Z","timestamp":1775129869000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-02150-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,5]]},"references-count":23,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["2150"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-02150-4","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,5]]},"assertion":[{"value":"15 July 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 February 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"146"}}