{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T03:05:53Z","timestamp":1767323153853,"version":"3.48.0"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032139603","type":"print"},{"value":"9783032139610","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-13961-0_23","type":"book-chapter","created":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T03:02:30Z","timestamp":1767322950000},"page":"225-235","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Interpretability-Aware Pruning for\u00a0Efficient Medical Image Analysis"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-1253-6269","authenticated-orcid":false,"given":"Nikita","family":"Malik","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0001-4525-4464","authenticated-orcid":false,"given":"Pratinav","family":"Seth","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0006-5465-4247","authenticated-orcid":false,"given":"Neeraj K.","family":"Singh","sequence":"additional","affiliation":[]},{"given":"Chintan","family":"Chitroda","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9416-3497","authenticated-orcid":false,"given":"Vinay K.","family":"Sankarapu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,2]]},"reference":[{"key":"23_CR1","unstructured":"Arrieta, A.B., et al.: Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI (2019). https:\/\/arxiv.org\/abs\/1910.10045"},{"key":"23_CR2","doi-asserted-by":"crossref","unstructured":"Atakishiyev, S., Salameh, M., Goebel, R.: Safety implications of explainable artificial intelligence in end-to-end autonomous driving (2025). https:\/\/arxiv.org\/abs\/2403.12176","DOI":"10.1109\/TITS.2025.3574738"},{"key":"23_CR3","unstructured":"Blalock, D., Ortiz, J.J.G., Frankle, J., Guttag, J.V.: What is the state of neural network pruning? In: Proceedings of Machine Learning and Systems, vol. 2, pp. 129\u2013146 (2020)"},{"issue":"1","key":"23_CR4","doi-asserted-by":"publisher","first-page":"10200","DOI":"10.1038\/s41598-020-67076-5","volume":"10","author":"XP Burgos-Artizzu","year":"2020","unstructured":"Burgos-Artizzu, X.P., et al.: Evaluation of deep convolutional neural networks for automatic classification of common maternal fetal ultrasound planes. Sci. Rep. 10(1), 10200 (2020)","journal-title":"Sci. Rep."},{"key":"23_CR5","unstructured":"Dosovitskiy, A., et al.: An image is worth $$16 \\times 16$$ words: transformers for image recognition at scale. In: International Conference on Learning Representations (ICLR) (2021)"},{"key":"23_CR6","doi-asserted-by":"crossref","unstructured":"Esteva, A., et al.: Deep learning-enabled medical computer vision. NPJ Digit. Med. 4(1), 5 (2021)","DOI":"10.1038\/s41746-020-00376-2"},{"key":"23_CR7","unstructured":"Frankle, J., Carbin, M.: The lottery ticket hypothesis: finding sparse, trainable neural networks. In: International Conference on Learning Representations (ICLR) (2019)"},{"key":"23_CR8","unstructured":"Fresz, B., et al.: The contribution of XAI for the safe development and certification of AI: an expert-based analysis (2024). https:\/\/arxiv.org\/abs\/2408.02379"},{"key":"23_CR9","unstructured":"Gale, T., Elsen, E., Hooker, S.: The state of sparsity in deep neural networks. arXiv preprint arXiv:1902.09574 (2019)"},{"key":"23_CR10","unstructured":"Gohel, P., Singh, P., Mohanty, M.: Explainable AI: current status and future directions (2021). https:\/\/arxiv.org\/abs\/2107.07045"},{"key":"23_CR11","unstructured":"Han, S., Pool, J., Tran, J., Dally, W.J.: Learning both weights and connections for efficient neural networks. In: Advances in Neural Information Processing Systems (NeurIPS) (2015)"},{"key":"23_CR12","unstructured":"Hassibi, B., Stork, D.G.: Second order derivatives for network pruning: optimal brain surgeon. In: Advances in Neural Information Processing Systems (NeurIPS), pp. 164\u2013171 (1993)"},{"key":"23_CR13","doi-asserted-by":"crossref","unstructured":"Hatefi, S.M.V., Dreyer, M., Achtibat, R., Wiegand, T., Samek, W., Lapuschkin, S.: Pruning by explaining revisited: optimizing attribution methods to prune CNNs and transformers (2024). https:\/\/arxiv.org\/abs\/2408.12568","DOI":"10.1007\/978-3-031-92648-8_10"},{"key":"23_CR14","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"23_CR15","unstructured":"Holzinger, A., Biemann, C., Pattichis, C.S., Kell, D.B.: What do we need to build explainable AI systems for the medical domain? arXiv preprint arXiv:1712.09923 (2017)"},{"key":"23_CR16","unstructured":"Kumar, S.: Covid19-Pneumonia-normal chest X-Ray images (2022)"},{"key":"23_CR17","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1016\/j.media.2017.07.005","volume":"42","author":"G Litjens","year":"2017","unstructured":"Litjens, G., et al.: A survey on deep learning in medical image analysis. Med. Image Anal. 42, 60\u201388 (2017)","journal-title":"Med. Image Anal."},{"key":"23_CR18","unstructured":"Liu, Z., Sun, M., Zhou, T., Huang, G., Darrell, T.: Rethinking the value of network pruning. In: International Conference on Learning Representations (ICLR) (2019)"},{"key":"23_CR19","doi-asserted-by":"crossref","unstructured":"Montavon, G., Binder, A., Lapuschkin, S., Samek, W., M\u00fcller, K.: Layer-wise relevance propagation: an overview. In: Explainable AI (2019). https:\/\/api.semanticscholar.org\/CorpusID:202579539","DOI":"10.1007\/978-3-030-28954-6_10"},{"key":"23_CR20","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1016\/j.patcog.2016.11.008","volume":"65","author":"G Montavon","year":"2017","unstructured":"Montavon, G., Lapuschkin, S., Binder, A., Samek, W., M\u00fcller, K.R.: Explaining nonlinear classification decisions with deep taylor decomposition. Pattern Recogn. 65, 211\u2013222 (2017)","journal-title":"Pattern Recogn."},{"key":"23_CR21","doi-asserted-by":"crossref","unstructured":"Pogorelov, K., et\u00a0al.: Kvasir: a multi-class image dataset for computer aided gastrointestinal disease detection. In: Proceedings of the 8th ACM on Multimedia Systems Conference, pp. 164\u2013169 (2017)","DOI":"10.1145\/3083187.3083212"},{"key":"23_CR22","unstructured":"Rajpurkar, P., et al.: Mura: large dataset for abnormality detection in musculoskeletal radiographs. arXiv preprint arXiv:1712.06957 (2017)"},{"key":"23_CR23","unstructured":"Sabih, M., Hannig, F., Teich, J.: Utilizing explainable AI for quantization and pruning of deep neural networks. arXiv abs\/2008.09072 (2020). https:\/\/api.semanticscholar.org\/CorpusID:221186873"},{"key":"23_CR24","unstructured":"Sanh, V., Wolf, T., Rush, A.M.: Movement pruning: adaptive sparsity by fine-tuning. In: Advances in Neural Information Processing Systems, vol.\u00a033, pp. 20378\u201320389 (2020)"},{"key":"23_CR25","doi-asserted-by":"crossref","unstructured":"Sankarapu, V.K., Chitroda, C., Rathore, Y., Singh, N.K., Seth, P.: Dlbacktrace: a model agnostic explainability for any deep learning models (2025). https:\/\/arxiv.org\/abs\/2411.12643","DOI":"10.1109\/IJCNN64981.2025.11228632"},{"key":"23_CR26","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1146\/annurev-bioeng-071516-044442","volume":"19","author":"D Shen","year":"2017","unstructured":"Shen, D., Wu, G., Suk, H.I.: Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 19, 221\u2013248 (2017)","journal-title":"Annu. Rev. Biomed. Eng."},{"key":"23_CR27","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"23_CR28","unstructured":"Sundararajan, M., Taly, A., Yan, Q.: Axiomatic attribution for deep networks. In: Precup, D., Teh, Y.W. (eds.) Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6\u201311 August 2017. Proceedings of Machine Learning Research, vol.\u00a070, pp. 3319\u20133328. PMLR (2017). https:\/\/api.semanticscholar.org\/CorpusID:16747630"},{"key":"23_CR29","unstructured":"Zhu, M., Gupta, S.: To prune, or not to prune: exploring the efficacy of pruning for model compression. In: International Conference on Learning Representations (ICLR) (2018)"}],"container-title":["Lecture Notes in Computer Science","Efficient Medical Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-13961-0_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T03:02:33Z","timestamp":1767322953000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-13961-0_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9783032139603","9783032139610"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-13961-0_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"2 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors are affiliated with AryaXAI.com, part of Aurionpro Solutions Limited. DL-Backtrace was introduced by AryaXAI. The authors declare no further competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"EMA4MICCAI 2025","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Efficient Medical Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Daejeon","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Korea (Republic of)","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ema4miccai2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/sites.google.com\/view\/ema4miccai2025","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}