{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T03:08:53Z","timestamp":1774667333612,"version":"3.50.1"},"reference-count":74,"publisher":"MDPI AG","issue":"11","license":[{"start":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T00:00:00Z","timestamp":1730332800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["BDCC"],"abstract":"<jats:p>Artificial intelligence (AI) encompasses the development of systems that perform tasks typically requiring human intelligence, such as reasoning and learning. Despite its widespread use, AI often raises trust issues due to the opacity of its decision-making processes. This challenge has led to the development of explainable artificial intelligence (XAI), which aims to enhance user understanding and trust by providing clear explanations of AI decisions and processes. This paper reviews existing XAI research, focusing on its application in the healthcare sector, particularly in medical and medicinal contexts. Our analysis is organized around key properties of XAI\u2014understandability, comprehensibility, transparency, interpretability, and explainability\u2014providing a comprehensive overview of XAI techniques and their practical implications.<\/jats:p>","DOI":"10.3390\/bdcc8110149","type":"journal-article","created":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T06:55:37Z","timestamp":1730357737000},"page":"149","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":39,"title":["Exploring the Landscape of Explainable Artificial Intelligence (XAI): A Systematic Review of Techniques and Applications"],"prefix":"10.3390","volume":"8","author":[{"given":"Sayda Umma","family":"Hamida","sequence":"first","affiliation":[{"name":"Department of Computer Science and Engineering, Daffodil International University, Birulia, Dhaka 1216, Bangladesh"}]},{"given":"Mohammad Jabed Morshed","family":"Chowdhury","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Daffodil International University, Birulia, Dhaka 1216, Bangladesh"},{"name":"Department of Computer Science and IT, La Trobe University, Melbourne, VIC 3086, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2227-9244","authenticated-orcid":false,"given":"Narayan Ranjan","family":"Chakraborty","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Daffodil International University, Birulia, Dhaka 1216, Bangladesh"}]},{"given":"Kamanashis","family":"Biswas","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Daffodil International University, Birulia, Dhaka 1216, Bangladesh"},{"name":"Faculty of Law and Business, Australian Catholic University, Brisbane, QLD 4014, Australia"}]},{"given":"Shahrab Khan","family":"Sami","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Shah Jalal University of Science and Technology, Sylhet 3114, Bangladesh"}]}],"member":"1968","published-online":{"date-parts":[[2024,10,31]]},"reference":[{"key":"ref_1","unstructured":"Knight, W. (2020). AI\u2019s Language Problem, MIT Technology Review. Available online: https:\/\/www.technologyreview.com\/2016\/08\/09\/158125\/ais-language-problem\/."},{"key":"ref_2","unstructured":"Innovation, Science and Economic Development Canada (2021, March 22). Government of Canada Announces Next Steps in Safeguarding Research. Canada.ca, Available online: https:\/\/www.canada.ca\/en\/innovation-science-economic-development\/news\/2021\/03\/government-of-canada-announces-next-steps-in-safeguarding-research.html."},{"key":"ref_3","unstructured":"IBM (2021, September 06). Explainable AI. Available online: https:\/\/www.ibm.com\/watson\/explainable-ai."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Wiener, N. (1961). Cybernetics: Or the Control and Communication in the Animal and the Machine. Analog VLSI: Signal and Information Processing, MIT Press. Available online: https:\/\/direct.mit.edu\/books\/oa-monograph\/4581\/Cybernetics-or-Control-and-Communication-in-the.","DOI":"10.1037\/13140-000"},{"key":"ref_5","first-page":"436","article-title":"Deep Learning","volume":"521","author":"LeCun","year":"2015","journal-title":"Res. Net"},{"key":"ref_6","unstructured":"Patil, R. (2022, May 05). Industry Trends in AI\u2014Topic 3\u2014Explainable AI. Available online: https:\/\/www.linkedin.com\/pulse\/industry-trends-ai-topic-3-explainable-rajesh-patil."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"154096","DOI":"10.1109\/ACCESS.2019.2949286","article-title":"Black-Box vs. White-Box: Understanding Their Advantages and Weaknesses From a Practical Point of View","volume":"7","year":"2019","journal-title":"IEEE Access"},{"key":"ref_8","unstructured":"(2021, September 13). Transparency and Responsibility in Artificial Intelligence. Available online: https:\/\/www2.deloitte.com\/content\/dam\/Deloitte\/nl\/Documents\/innovatie\/deloitte-nl-innovation-bringing-transparency-and-ethics-into-ai.pdf."},{"key":"ref_9","unstructured":"Haasdijk, E. (2021, September 13). A Call for Transparency and Responsibility in Artificial Intelligence. Available online: https:\/\/www2.deloitte.com\/nl\/nl\/pages\/innovatie\/artikelen\/a-call-for-transparency-and-responsibility-in-artificial-intelligence.html."},{"key":"ref_10","unstructured":"(2021, September 13). Model Transparency and Explainability. Ople.Ai., Available online: https:\/\/ople.ai\/ai-blog\/model-transparency-and-explainability\/."},{"key":"ref_11","unstructured":"Choudhury, A. (2021, September 16). Explainability vs. Interpretability In Artificial Intelligence and Machine Learning. Available online: https:\/\/analyticsindiamag.com\/explainability-vs-interpretability-in-artificial-intelligence-and-machine-learning\/."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"103367","DOI":"10.1016\/j.artint.2020.103367","article-title":"Interestingness Elements for Explainable Reinforcement Learning: Understanding Agents\u2019 Capabilities and Limitations","volume":"288","author":"Sequeira","year":"2020","journal-title":"Artif. Intell."},{"key":"ref_13","unstructured":"(2021, September 16). Model Interpretability. Available online: https:\/\/www.datarobot.com\/wiki\/interpretability\/."},{"key":"ref_14","unstructured":"Shen, O. (2021, September 13). Interpretability in Machine Learning: An Overview. Available online: https:\/\/thegradient.pub\/interpretability-in-ml-a-broad-overview\/."},{"key":"ref_15","unstructured":"(2021, September 13). Model Interpretability. Available online: https:\/\/www.darpa.mil\/attachments\/XAIProgramUpdate.pdf."},{"key":"ref_16","unstructured":"(2024, September 26). Systematic Reviews\u2014Subject and Research Guides at Macquarie University. Available online: https:\/\/libguides.mq.edu.au\/systematic_reviews\/prisma_screen#:~:text=The%20PRISMA%20flow%20diagram%20visually,recorded%20at%20the%20different%20stages."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Carvalho, D.V., Pereira, E.M., and Cardoso, J.S. (2019). Machine Learning Interpretability: A Survey on Methods and Metrics. Electronics, 8.","DOI":"10.3390\/electronics8080832"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Meske, C., and Bunde, E. (2020). Transparency and Trust in Human-AI-Interaction: The Role of Model-Agnostic Explanations in Computer Vision-Based Decision Support. Artificial Intelligence in HCI, Springer.","DOI":"10.1007\/978-3-030-50334-5_4"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"113941","DOI":"10.1016\/j.eswa.2020.113941","article-title":"Post-Hoc Explanation of Black-Box Classifiers Using Confident Itemsets","volume":"165","author":"Moradi","year":"2021","journal-title":"Expert Syst. Appl."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1561\/1500000066","article-title":"Explainable Recommendation: A Survey and New Perspectives","volume":"14","author":"Zhang","year":"2020","journal-title":"Found. Trends\u00ae Inf. Retr."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"7","DOI":"10.1017\/eds.2022.7","article-title":"Neural Network Attribution Methods for Problems in Geoscience: A Novel Synthetic Benchmark Dataset","volume":"1","author":"Mamalakis","year":"2022","journal-title":"Environ. Data Sci."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"281","DOI":"10.1016\/j.future.2022.03.009","article-title":"The Explainability Paradox: Challenges for XAI in Digital Pathology","volume":"133","author":"Evans","year":"2022","journal-title":"Future Gener. Comput. Syst."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Matin, S.S., and Pradhan, B. (2021). Earthquake-Induced Building-Damage Mapping Using Explainable AI (XAI). Sensors, 21.","DOI":"10.3390\/s21134489"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"149","DOI":"10.1002\/9781119910619.ch7","article-title":"Explainable Artificial Intelligence for Cybersecurity","volume":"103","author":"Femi","year":"2023","journal-title":"Wirel. Commun. Cybersecur."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"103471","DOI":"10.1016\/j.artint.2021.103471","article-title":"Using Ontologies to Enhance Human Understandability of Global Post-Hoc Explanations of Black-Box Models","volume":"296","author":"Confalonieri","year":"2021","journal-title":"Artif. Intell."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"103507","DOI":"10.1016\/j.artint.2021.103507","article-title":"\u201cThat\u2019s (Not) the Output I Expected!\u201d On the Role of End User Expectations in Creating Explanations of AI Systems","volume":"298","author":"Riveiro","year":"2021","journal-title":"Artif. Intell."},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"103404","DOI":"10.1016\/j.artint.2020.103404","article-title":"Evaluating XAI: A Comparison of Rule-Based and Example-Based Explanations","volume":"291","author":"Nieuwburg","year":"2021","journal-title":"Artif. Intell."},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Dobrovolskis, A., Kazanavi\u010dius, E., and Ki\u017eauskien\u0117, L. (2023). Building XAI-Based Agents for IoT Systems. Appl. Sci., 13.","DOI":"10.3390\/app13064040"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Singh, A., Sengupta, S., and Lakshminarayanan, V. (2020). Explainable Deep Learning Models in Medical Image Analysis. J. Imaging, 6.","DOI":"10.3390\/jimaging6060052"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Linardatos, P., Papastefanopoulos, V., and Kotsiantis, S. (2021). Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy, 23.","DOI":"10.3390\/e23010018"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Sahoh, B., and Choksuriwong, A. (2021). Beyond Deep Event Prediction: Deep Event Understanding Based on Explainable Artificial Intelligence. Explainable AI and Other Applications of Intelligent Computing, Springer.","DOI":"10.1007\/978-3-030-64949-4_4"},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"106916","DOI":"10.1016\/j.knosys.2021.106916","article-title":"Classification by Ordinal Sums of Conjunctive and Disjunctive Functions for Explainable AI and Interpretable Machine Learning Solutions","volume":"220","author":"Hudec","year":"2021","journal-title":"Knowl.-Based Syst."},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Wells, L., and Bednarz, T. (2021). Explainable AI and Reinforcement Learning\u2014A Systematic Review of Current Approaches and Trends. Front. Artif. Intell., 4.","DOI":"10.3389\/frai.2021.550030"},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Zhou, J., Gandomi, A.H., Chen, F., and Holzinger, A. (2021). Evaluating the Quality of Machine Learning Explanations: A Survey on Methods and Metrics. Electronics, 10.","DOI":"10.3390\/electronics10050593"},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"103457","DOI":"10.1016\/j.artint.2021.103457","article-title":"GLocalX\u2014From Local to Global Explanations of Black Box AI Models","volume":"294","author":"Setzu","year":"2021","journal-title":"Artif. Intell."},{"key":"ref_36","first-page":"1","article-title":"Explainability for Artificial Intelligence in Healthcare: A Multidisciplinary Perspective","volume":"20","author":"Amann","year":"2020","journal-title":"BMC Med. Inf. Decis. Mak."},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"de Sousa, I.P., Vellasco, M.M.B.R., and Costa da Silva, E. (2021). Explainable Artificial Intelligence for Bias Detection in COVID CT-Scan Classifiers. Sensors, 21.","DOI":"10.3390\/s21165657"},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Ogrezeanu, I., Vizitu, A., Ciu\u0219del, C., Puiu, A., Coman, S., Boldi\u0219or, C., Itu, A., Robert, D., Moldoveanu, F., and Suciu, C. (2022). Privacy-Preserving and Explainable AI in Industrial Applications. Appl. Sci., 12.","DOI":"10.3390\/app12136395"},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Merry, M., Riddle, P., and Warren, J. (2021). A Mental Models Approach for Defining Explainable Artificial Intelligence. BMC Med. Inform. Decis. Mak., 21.","DOI":"10.1186\/s12911-021-01703-7"},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Theunissen, M., and Browning, J. (2022). Putting Explainable AI in Context: Institutional Explanations for Medical AI. Ethics Inf. Technol., 24.","DOI":"10.1007\/s10676-022-09649-8"},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"124","DOI":"10.3390\/biomedinformatics2010008","article-title":"Quantified Explainability: Convolutional Neural Network Focus Assessment in Arrhythmia Detection","volume":"2","author":"Varandas","year":"2022","journal-title":"BioMedInformatics"},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"102472","DOI":"10.1016\/j.inffus.2024.102472","article-title":"Explainable AI-Driven IoMT Fusion: Unravelling Techniques, Opportunities, and Challenges with Explainable AI in Healthcare","volume":"110","author":"Wani","year":"2024","journal-title":"Inf. Fusion"},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"35728","DOI":"10.1109\/ACCESS.2024.3370848","article-title":"Generative Adversarial Networks (GANs) in Medical Imaging: Advancements, Applications, and Challenges","volume":"12","author":"Islam","year":"2024","journal-title":"IEEE Access"},{"key":"ref_44","unstructured":"Manesh, A. (2024, September 17). Interpretable Machine Learning (IML) with XGBoost and Additive Tools. Medium, Available online: https:\/\/medium.com\/@anicomanesh\/interpretable-machine-learning-iml-with-xgboost-and-additive-tools-42258fb1f14."},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Chen, T., and Guestrin, C. (2016). XGBoost: A Scalable Tree Boosting System. arXiv, 13.","DOI":"10.1145\/2939672.2939785"},{"key":"ref_46","unstructured":"Zhang, H. (2019). Explainable AI: A Brief Survey on History, Research Areas, Approaches and Challenges. Natural Language Processing and Chinese Computing, Springer."},{"key":"ref_47","first-page":"089443932098011","article-title":"Will Algorithms Blind People? The Effect of Explainable AI and Decision-Makers\u2019 Experience on AI-Supported Decision-Making in Government","volume":"40","author":"Janssen","year":"2020","journal-title":"Soc. Sci. Comput. Rev."},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"1494","DOI":"10.3390\/make6030071","article-title":"A Cognitive Load Theory (CLT) Analysis of Machine Learning Explainability, Transparency, Interpretability, and Shared Interpretability","volume":"6","author":"Fox","year":"2024","journal-title":"Mach. Learn. Knowl. Extr."},{"key":"ref_49","doi-asserted-by":"crossref","first-page":"101805","DOI":"10.1016\/j.inffus.2023.101805","article-title":"Explainable Artificial Intelligence (XAI): What We Know and What Is Left to Attain Trustworthy Artificial Intelligence","volume":"99","author":"Ali","year":"2023","journal-title":"Inf. Fusion"},{"key":"ref_50","doi-asserted-by":"crossref","first-page":"104405","DOI":"10.1016\/j.trc.2023.104405","article-title":"Towards Explainable Motion Prediction Using Heterogeneous Graph Representations","volume":"157","author":"Limeros","year":"2023","journal-title":"Transp. Res. Part C Emerg. Technol."},{"key":"ref_51","doi-asserted-by":"crossref","first-page":"82","DOI":"10.1016\/j.inffus.2019.12.012","article-title":"Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges Toward Responsible AI","volume":"58","author":"Bennetot","year":"2020","journal-title":"Inf. Fusion"},{"key":"ref_52","doi-asserted-by":"crossref","first-page":"102493","DOI":"10.1016\/j.ijhcs.2020.102493","article-title":"Interpretable Confidence Measures for Decision Support Systems","volume":"144","author":"Schoonderwoerd","year":"2020","journal-title":"Int. J. Hum.-Comput. Stud."},{"key":"ref_53","doi-asserted-by":"crossref","unstructured":"Marton, S., L\u00fcdtke, S., and Bartelt, C. (2022). Explanations for Neural Networks by Neural Networks. Appl. Sci., 12.","DOI":"10.3390\/app12030980"},{"key":"ref_54","doi-asserted-by":"crossref","first-page":"29","DOI":"10.1016\/j.inffus.2021.07.016","article-title":"Unbox the Black-Box for the Medical Explainable AI via Multi-Modal and Multi-Centre Data Fusion: A Mini-Review, Two Showcases and Beyond","volume":"77","author":"Yang","year":"2022","journal-title":"Inf. Fusion"},{"key":"ref_55","doi-asserted-by":"crossref","first-page":"101666","DOI":"10.1016\/j.giq.2021.101666","article-title":"The Perils and Pitfalls of Explainable AI: Strategies for Explaining Algorithmic Decision-Making","volume":"39","author":"Bruijn","year":"2021","journal-title":"Gov. Inf. Q."},{"key":"ref_56","doi-asserted-by":"crossref","first-page":"e16110","DOI":"10.1016\/j.heliyon.2023.e16110","article-title":"Essential Properties and Explanation Effectiveness of Explainable Artificial Intelligence in Healthcare: A Systematic Review","volume":"9","author":"Jung","year":"2023","journal-title":"Heliyon"},{"key":"ref_57","doi-asserted-by":"crossref","unstructured":"Loh, H.W., Ooi, C.P., Seoni, S., Barua, P.D., Molinari, F., and Acharya, U.R. (2022). Application of Explainable Artificial Intelligence for Healthcare: A Systematic Review of the Last Decade (2011\u20132022). Comput. Methods Programs Biomed., 226.","DOI":"10.1016\/j.cmpb.2022.107161"},{"key":"ref_58","doi-asserted-by":"crossref","unstructured":"Bresso, E., Monnin, P., Bousquet, C., Calvier, F.E., Ndiaye, N.C., Petitpain, N., Sma\u00efl-Tabbone, M., and Coulet, A. (2021). Investigating ADR Mechanisms with Explainable AI: A Feasibility Study with Knowledge Graph Mining. BMC Med. Inform. Decis. Mak., 21.","DOI":"10.1186\/s12911-021-01518-6"},{"key":"ref_59","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1007\/s10676-022-09632-3","article-title":"Explanatory Pragmatism: A Context-Sensitive Framework for Explainable Medical AI","volume":"24","author":"Nyrup","year":"2022","journal-title":"Ethics Inf. Technol."},{"key":"ref_60","doi-asserted-by":"crossref","unstructured":"Gashi, M., Mutlu, B., and Thalmann, S. (2023). Impact of Interdependencies: Multi-Component System Perspective Toward Predictive Maintenance Based on Machine Learning and XAI. Appl. Sci., 13.","DOI":"10.3390\/app13053088"},{"key":"ref_61","doi-asserted-by":"crossref","unstructured":"Antoniadi, A.M., Du, Y., Guendouz, Y., Wei, L., Mazo, C., Becker, B.A., and Mooney, C. (2021). Current Challenges and Future Opportunities for XAI in Machine Learning-Based Clinical Decision Support Systems: A Systematic Review. Appl. Sci., 11.","DOI":"10.3390\/app11115088"},{"key":"ref_62","doi-asserted-by":"crossref","first-page":"245","DOI":"10.1613\/jair.1.12228","article-title":"A Survey on the Explainability of Supervised Machine Learning","volume":"70","author":"Burkart","year":"2021","journal-title":"J. Artif. Intell. Res."},{"key":"ref_63","doi-asserted-by":"crossref","first-page":"e5","DOI":"10.1017\/S0269888921000011","article-title":"Argumentation and Explainable Artificial Intelligence: A Survey","volume":"36","author":"Vassiliades","year":"2021","journal-title":"Knowl. Eng. Rev."},{"key":"ref_64","doi-asserted-by":"crossref","unstructured":"Tritscher, J., Krause, A., and Hotho, A. (2023). Feature Relevance XAI in Anomaly Detection: Reviewing Approaches and Challenges. Front. Artif. Intell., 6.","DOI":"10.3389\/frai.2023.1099521"},{"key":"ref_65","doi-asserted-by":"crossref","first-page":"102558","DOI":"10.1016\/j.irfa.2023.102558","article-title":"Prediction and Interpretation of Daily NFT and DeFi Prices Dynamics: Inspection Through Ensemble Machine Learning & XAI","volume":"87","author":"Ghosh","year":"2023","journal-title":"Int. Rev. Financ. Anal."},{"key":"ref_66","doi-asserted-by":"crossref","first-page":"129169","DOI":"10.1109\/ACCESS.2020.3009852","article-title":"Vision-Based Fault Diagnostics Using Explainable Deep Learning With Class Activation Maps","volume":"8","author":"Sun","year":"2020","journal-title":"IEEE Access"},{"key":"ref_67","doi-asserted-by":"crossref","first-page":"611","DOI":"10.1007\/s13244-018-0639-9","article-title":"Convolutional Neural Networks: An Overview and Application in Radiology","volume":"9","author":"Yamashita","year":"2018","journal-title":"Insights Imaging"},{"key":"ref_68","doi-asserted-by":"crossref","unstructured":"Berloco, F., Marvulli, P.M., Suglia, V., Colucci, S., Pagano, G., Palazzo, L., Aliani, M., Castellana, G., Guido, P., and D\u2019Addio, G. (2024). Enhancing Survival Analysis Model Selection through XAI(t) in Healthcare. Appl. Sci., 14.","DOI":"10.3390\/app14146084"},{"key":"ref_69","doi-asserted-by":"crossref","first-page":"28896","DOI":"10.1109\/ACCESS.2023.3255403","article-title":"Interpretable Classification of Pneumonia Infection Using Explainable AI (XAI-ICP)","volume":"11","author":"Sheu","year":"2023","journal-title":"IEEE Access"},{"key":"ref_70","doi-asserted-by":"crossref","unstructured":"Jinad, R., Islam, A., and Shashidhar, N. (2024). Interpretability and Transparency of Machine Learning in File Fragment Analysis with Explainable Artificial Intelligence. Electronics, 13.","DOI":"10.3390\/electronics13132438"},{"key":"ref_71","doi-asserted-by":"crossref","first-page":"52138","DOI":"10.1109\/ACCESS.2018.2870052","article-title":"Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)","volume":"6","author":"Adadi","year":"2018","journal-title":"IEEE Access"},{"key":"ref_72","first-page":"100118","article-title":"Systematic Literature Review on Opportunities, Challenges, and Future Research Recommendations of Artificial Intelligence in Education","volume":"4","author":"Chiu","year":"2023","journal-title":"CAEAI"},{"key":"ref_73","unstructured":"Vilone, G., and Longo, L. (2020). Explainable Artificial Intelligence: A Systematic Review. arXiv."},{"key":"ref_74","unstructured":"(2024). AI Tools. Latest AI Tools and Technologies Every Mobile App Developer Should Know. Appmysite, Available online: https:\/\/www.appmysite.com\/blog\/top-ai-tools-for-mobile-app-developers\/."}],"container-title":["Big Data and Cognitive Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2504-2289\/8\/11\/149\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T16:25:19Z","timestamp":1760113519000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2504-2289\/8\/11\/149"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,31]]},"references-count":74,"journal-issue":{"issue":"11","published-online":{"date-parts":[[2024,11]]}},"alternative-id":["bdcc8110149"],"URL":"https:\/\/doi.org\/10.3390\/bdcc8110149","relation":{},"ISSN":["2504-2289"],"issn-type":[{"value":"2504-2289","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,31]]}}}