{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T15:56:49Z","timestamp":1774540609458,"version":"3.50.1"},"reference-count":54,"publisher":"MDPI AG","issue":"6","license":[{"start":{"date-parts":[[2023,3,16]],"date-time":"2023-03-16T00:00:00Z","timestamp":1678924800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Korea Agency for Technology and Standards","award":["K_G012002234001"],"award-info":[{"award-number":["K_G012002234001"]}]},{"name":"Korea Agency for Technology and Standards","award":["K_G012002236201"],"award-info":[{"award-number":["K_G012002236201"]}]},{"name":"Korea Agency for Technology and Standards","award":["2020-08460006"],"award-info":[{"award-number":["2020-08460006"]}]},{"name":"Gachon University research fund","award":["K_G012002234001"],"award-info":[{"award-number":["K_G012002234001"]}]},{"name":"Gachon University research fund","award":["K_G012002236201"],"award-info":[{"award-number":["K_G012002236201"]}]},{"name":"Gachon University research fund","award":["2020-08460006"],"award-info":[{"award-number":["2020-08460006"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Deep learning has achieved remarkably positive results and impacts on medical diagnostics in recent years. Due to its use in several proposals, deep learning has reached sufficient accuracy to implement; however, the algorithms are black boxes that are hard to understand, and model decisions are often made without reason or explanation. To reduce this gap, explainable artificial intelligence (XAI) offers a huge opportunity to receive informed decision support from deep learning models and opens the black box of the method. We conducted an explainable deep learning method based on ResNet152 combined with Grad\u2013CAM for endoscopy image classification. We used an open-source KVASIR dataset that consisted of a total of 8000 wireless capsule images. The heat map of the classification results and an efficient augmentation method achieved a high positive result with 98.28% training and 93.46% validation accuracy in terms of medical image classification.<\/jats:p>","DOI":"10.3390\/s23063176","type":"journal-article","created":{"date-parts":[[2023,3,17]],"date-time":"2023-03-17T02:59:26Z","timestamp":1679021966000},"page":"3176","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":70,"title":["Endoscopic Image Classification Based on Explainable Deep Learning"],"prefix":"10.3390","volume":"23","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7665-9432","authenticated-orcid":false,"given":"Doniyorjon","family":"Mukhtorov","sequence":"first","affiliation":[{"name":"Department of IT Convergence Engineering, Gachon University, Sujeong-Gu, Seongnam-Si 461-701, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1691-8268","authenticated-orcid":false,"given":"Madinakhon","family":"Rakhmonova","sequence":"additional","affiliation":[{"name":"Department of IT Convergence Engineering, Gachon University, Sujeong-Gu, Seongnam-Si 461-701, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6223-4502","authenticated-orcid":false,"given":"Shakhnoza","family":"Muksimova","sequence":"additional","affiliation":[{"name":"Department of IT Convergence Engineering, Gachon University, Sujeong-Gu, Seongnam-Si 461-701, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0184-7599","authenticated-orcid":false,"given":"Young-Im","family":"Cho","sequence":"additional","affiliation":[{"name":"Department of IT Convergence Engineering, Gachon University, Sujeong-Gu, Seongnam-Si 461-701, Republic of Korea"}]}],"member":"1968","published-online":{"date-parts":[[2023,3,16]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"53","DOI":"10.1016\/j.compmedimag.2018.09.004","article-title":"SD-CNN: A shallow-deep CNN for improved breast cancer diagnosis","volume":"70","author":"Gao","year":"2018","journal-title":"Comput. Med. Imaging Graph."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"781","DOI":"10.1148\/radiol.2019190613","article-title":"A Roadmap for Foundational Research on Artificial Intelligence in Medical Imaging","volume":"291","author":"Langlotz","year":"2019","journal-title":"Radiology"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"6536","DOI":"10.3934\/mbe.2019326","article-title":"Convolutional neural networks for computer-aided detection or diagnosis in medical image analysis: An overview","volume":"16","author":"Gao","year":"2019","journal-title":"Math. Biosci. Eng."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"303","DOI":"10.1016\/j.media.2016.07.007","article-title":"Large scale deep learning for computer aided detection of mammographic lesions","volume":"35","author":"Kooi","year":"2017","journal-title":"Med. Image Anal."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"221e48","DOI":"10.1146\/annurev-bioeng-071516-044442","article-title":"Deep learning in medical image analysis","volume":"19","author":"Shen","year":"2017","journal-title":"Annu. Rev. Biomed. Eng."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Liang, P., Shi, W., and Zhang, X. (2018). Remote Sensing Image Classification Based on Stacked Denoising Autoencoder. Remote Sens., 10.","DOI":"10.3390\/rs10010016"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Song, L., Wang, L., Sun, H., Cui, C., and Yu, Z. (2022). Fatigue Performance Prediction of RC Beams Based on Optimized Machine Learning Technology. Materials, 15.","DOI":"10.3390\/ma15186349"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Yang, S., Lee, G., and Huang, L. (2022). Deep Learning-Based Dynamic Computation Task Offloading for Mobile Edge Computing Networks. Sensors, 22.","DOI":"10.3390\/s22114088"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Urtnasan, E., Park, J.-U., Lee, J.-H., Koh, S.-B., and Lee, K.-J. (2022). Deep Learning for Automatic Detection of Periodic Limb Movement Disorder Based on Electrocardiogram Signals. Diagnostics, 12.","DOI":"10.3390\/diagnostics12092149"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Mad Sahad, S., Tan, N.W., Sajid, M., Jones, E.A., and Abdul Latiff, A.H. (2022). Enhancing Channelized Feature Interpretability Using Deep Learning Predictive Modeling. Appl. Sci., 12.","DOI":"10.3390\/app12189032"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Ahmed, M., Mumtaz, R., Anwar, Z., Shaukat, A., Arif, O., and Shafait, F. (2022). A Multi\u2013Step Approach for Optically Active and Inactive Water Quality Parameter Estimation Using Deep Learning and Remote Sensing. Water, 14.","DOI":"10.3390\/w14132112"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Tsuneki, M., Abe, M., and Kanavati, F. (2022). A Deep Learning Model for Prostate Adenocarcinoma Classification in Needle Biopsy Whole-Slide Images Using Transfer Learning. Diagnostics, 12.","DOI":"10.3390\/diagnostics12030768"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Habibollahi, R., Seydi, S.T., Hasanlou, M., and Mahdianpari, M. (2022). TCD-Net: A Novel Deep Learning Framework for Fully Polarimetric Change Detection Using Transfer Learning. Remote Sens., 14.","DOI":"10.3390\/rs14030438"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"22","DOI":"10.3390\/make4010002","article-title":"A Transfer Learning Evaluation of Deep Neural Networks for Image Classification","volume":"4","author":"Zengeler","year":"2022","journal-title":"Mach. Learn. Knowl. Extr."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Asam, M., Hussain, S.J., Mohatram, M., Khan, S.H., Jamal, T., Zafar, A., Khan, A., Ali, M.U., and Zahoora, U. (2021). Detection of Exceptional Malware Variants Using Deep Boosted Feature Spaces and Machine Learning. Appl. Sci., 11.","DOI":"10.3390\/app112110464"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Javed Awan, M., Mohd Rahim, M.S., Salim, N., Mohammed, M.A., Garcia-Zapirain, B., and Abdulkareem, K.H. (2021). Efficient Detection of Knee Anterior Cruciate Ligament from Magnetic Resonance Imaging Using Deep Learning Approach. Diagnostics, 11.","DOI":"10.3390\/diagnostics11010105"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Alkhelaiwi, M., Boulila, W., Ahmad, J., Koubaa, A., and Driss, M. (2021). An Efficient Approach Based on Privacy-Preserving Deep Learning for Satellite Image Classification. Remote Sens., 13.","DOI":"10.3390\/rs13112221"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Chen, H.-C., Widodo, A.M., Wisnujati, A., Rahaman, M., Lin, J.C.-W., Chen, L., and Weng, C.-E. (2022). AlexNet Convolutional Neural Network for Disease Detection and Classification of Tomato Leaf. Electronics, 11.","DOI":"10.3390\/electronics11060951"},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Nicholas, P.J., To, A., Tanglay, O., Young, I.M., Sughrue, M.E., and Doyen, S. (2022). Using a ResNet-18 Network to Detect Features of Alzheimer\u2019s Disease on Functional Magnetic Resonance Imaging: A Failed Replication. Comment on Odusami et al. Analysis of Features of Alzheimer\u2019s Disease: Detection of Early Stage from Functional Brain Changes in Magnetic Resonance Images Using a Finetuned ResNet18 Network. Diagnostics 2021, 11, 1071. Diagnostics, 12.","DOI":"10.3390\/diagnostics12051094"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Ullah, Z., Lodhi, B.A., and Hur, J. (2020). Detection and Identification of Demagnetization and Bearing Faults in PMSM Using Transfer Learning-Based VGG. Energies, 13.","DOI":"10.3390\/en13153834"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"454","DOI":"10.3758\/s13423-020-01825-5","article-title":"Artificial cognition: How experimental psychology can help generate explainable artificial intelligence","volume":"28","author":"Taylor","year":"2021","journal-title":"Psychon. Bull. Rev."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Markus, A., Kors, J., and Rijnbeek, P. (2020). The role of explainability in creating trustworthy artificial intelligence for health care: A comprehensive survey of the terminology, design choices, and evaluation strategies. arXiv.","DOI":"10.1016\/j.jbi.2020.103655"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Global Burden of Disease Cancer Collaboration, Fitzmaurice, C., Akinyemiju, T.F., Al Lami, F.H., Alam, T., Alizadeh-Navaei, R., Allen, C., Alsharif, U., Alvis-Guzman, N., and Amini, E. (2018). Global, regional, and national cancer incidence, mortality, years of life lost, years lived with disability, and disability-adjusted life-years for 29 cancer groups, 1990 to 2016: A systematic analysis for the global burden of disease study. JAMA Oncol., 4, 1553\u20131568.","DOI":"10.1200\/JCO.2018.36.15_suppl.1568"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Paradisa, R.H., Bustamam, A., Mangunwardoyo, W., Victor, A.A., Yudantha, A.R., and Anki, P. (2022). Deep Feature Vectors Concatenation for Eye Disease Detection Using Fundus Image. Electronics, 11.","DOI":"10.3390\/electronics11010023"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Xiao, C., Liu, Z., Zhang, T., and Zhang, X. (2021). Deep Learning Method for Fault Detection of Wind Turbine Converter. Appl. Sci., 11.","DOI":"10.3390\/app11031280"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Li, X., Du, J., Yang, J., and Li, S. (2022). When Mobilenetv2 Meets Transformer: A Balanced Sheep Face Recognition Model. Agriculture, 12.","DOI":"10.3390\/agriculture12081126"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Wang, C., and Xiao, Z. (2021). Potato Surface Defect Detection Based on Deep Transfer Learning. Agriculture, 11.","DOI":"10.3390\/agriculture11090863"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Mogan, J.N., Lee, C.P., Lim, K.M., and Muthu, K.S. (2022). VGG16-MLP: Gait Recognition with Fine-Tuned VGG-16 and Multilayer Perceptron. Appl. Sci., 12.","DOI":"10.3390\/app12157639"},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"283","DOI":"10.1038\/s41597-020-00622-y","article-title":"HyperKvasir, a comprehensive multi-class image and video dataset for gastrointestinal endoscopy","volume":"7","author":"Borgli","year":"2020","journal-title":"Sci. Data"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., and Van Der Maaten, L. (2017, January 21\u201326). Densely connected convolutional networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Liu, P., Wei, Y., Wang, Q., Chen, Y., and Xie, J. (2020). Research on Post-Earthquake Landslide Extraction Algorithm Based on Improved U-Net Model. Remote Sens., 12.","DOI":"10.3390\/rs12050894"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Metzger, U., Michel, A.-J., Ardelean, M.-A., and Metzger, R.P. (2022). Transanal Endoscopic-Assisted Pull-Through Colectomy for Children with High Intestinal Aganglionosis. Children, 9.","DOI":"10.3390\/children9050588"},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Yu, X., Tang, S., Cheang, C.F., Yu, H.H., and Choi, I.C. (2022). Multi-Task Model for Esophageal Lesion Analysis Using Endoscopic Images: Classification with Image Retrieval and Segmentation with Attention. Sensors, 22.","DOI":"10.3390\/s22010283"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Neupane, B., Horanont, T., and Aryal, J. (2021). Deep Learning-Based Semantic Segmentation of Urban Features in Satellite Images: A Review and Meta-Analysis. Remote Sens., 13.","DOI":"10.3390\/rs13040808"},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"653","DOI":"10.1007\/s10120-018-0793-2","article-title":"Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images","volume":"21","author":"Hirasawa","year":"2018","journal-title":"Gastric Cancer"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Cao, G., Song, W., and Zhao, Z. (2019, January 24\u201325). Gastric cancer diagnosis with mask R-CNN. Proceedings of the 2019 11th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC), Hangzhou, China.","DOI":"10.1109\/IHMSC.2019.00022"},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"126","DOI":"10.1007\/s10120-019-00992-2","article-title":"Convolutional neural network for the diagnosis of early gastric cancer based on magnifying narrow band imaging","volume":"23","author":"Li","year":"2020","journal-title":"Gastric Cancer"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"106","DOI":"10.1016\/j.ebiom.2017.10.014","article-title":"Application of convolutional neural networks in the diagnosis of helicobacter pylori infection based on endoscopic images","volume":"25","author":"Shichijo","year":"2017","journal-title":"eBioMedicine"},{"key":"ref_40","first-page":"462","article-title":"Artificial intelligence diagnosis of helicobacter pylori infection using blue laser imaging-bright and linked color imaging: A single-center prospective study","volume":"31","author":"Nakashima","year":"2018","journal-title":"Ann. Gastroenterol."},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Tajbakhsh, N., Gurudu, S.R., and Liang, J. (2015, January 16\u201319). Automatic polyp detection in colonoscopy videos using an ensemble of convolutional neural networks. Proceedings of the 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI), Brooklyn, NY, USA.","DOI":"10.1109\/ISBI.2015.7163821"},{"key":"ref_42","unstructured":"Simonyan, K., Vedaldi, A., and Zisserman, A. (2014). Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv."},{"key":"ref_43","unstructured":"K\u00fcmmerer, M., Theis, L., and Bethge, M. (2014). Deep Gaze I: Boosting Saliency Prediction with Feature Maps Trained on ImageNet. arXiv."},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017, January 22\u201329). Grad-cam: Visual explanations from deep networks via gradient-based localization. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.74"},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Esmaeili, M., Vettukattil, R., Banitalebi, H., Krogh, N.R., and Geitung, J.T. (2021). Explainable Artificial Intelligence for Human-Machine Interaction in Brain Tumor Localization. J. Pers. Med., 11.","DOI":"10.3390\/jpm11111213"},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"Wang, H., Wang, Z., Du, M., Yang, F., Zhang, Z., Ding, S., Mardziel, P., and Hu, X. (2020, January 14\u201319). Score-CAM: Score-weighted visual explanations for convolutional neural networks. Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA.","DOI":"10.1109\/CVPRW50498.2020.00020"},{"key":"ref_47","doi-asserted-by":"crossref","first-page":"5875","DOI":"10.1109\/TIP.2021.3089943","article-title":"LayerCAM: Exploring hierarchical class activation maps for localization","volume":"30","author":"Jiang","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"ref_48","unstructured":"Fu, R., Hu, Q., Dong, X., Guo, Y., Gao, Y., and Li, B. (2020). Axiom-based grad-cam: Towards accurate visualization and explanation of cnns. arXiv."},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., and Chen, L.C. (2018, January 18\u201322). Mobilenetv2: Inverted residuals and linear bottlenecks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00474"},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"Gjestang, H.L., Hicks, S.A., Thambawita, V., Halvorsen, P., and Riegler, M.A. (2021, January 7\u20139). A self-learning teacher-student framework for gastrointestinal image classification. Proceedings of the 2021 IEEE 34th International Symposium on Computer-Based Medical Systems (CBMS), Aveiro, Portugal.","DOI":"10.1109\/CBMS52027.2021.00087"},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"Thambawita, V., Str\u00fcmke, I., Hicks, S.A., Halvorsen, P., Parasa, S., and Riegler, M.A. (2021). Impact of Image Resolution on Deep Learning Performance in Endoscopy Image Classification: An Experimental Study Using a Large Dataset of Endoscopic Images. Diagnostics, 11.","DOI":"10.3390\/diagnostics11122183"},{"key":"ref_52","doi-asserted-by":"crossref","unstructured":"Srivastava, A., Tomar, N.K., Bagci, U., and Jha, D. (2022). Video Capsule Endoscopy Classification using Focal Modulation Guided Convolutional Neural Network. arXiv.","DOI":"10.1109\/CBMS55023.2022.00064"},{"key":"ref_53","doi-asserted-by":"crossref","unstructured":"Pozdeev, A.A., Obukhova, N.A., and Motyko, A.A. (2019, January 28\u201331). Automatic analysis of endoscopic images for polyps detection and segmentation. Proceedings of the IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (EIConRus),  Saint Petersburg and Moscow, Russia.","DOI":"10.1109\/EIConRus.2019.8657018"},{"key":"ref_54","doi-asserted-by":"crossref","unstructured":"Fonoll\u00e1, R., van der Sommen, F., Schreuder, R.M., Schoon, E.J., and de With, P.H. (2019, January 8\u201311). Multi-modal classification of polyp malignancy using CNN features with balanced class augmentation. Proceedings of the IEEE 16th International Symposium on Biomedical Imaging, Venice, Italy.","DOI":"10.1109\/ISBI.2019.8759320"}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/23\/6\/3176\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T18:56:44Z","timestamp":1760122604000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/23\/6\/3176"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,3,16]]},"references-count":54,"journal-issue":{"issue":"6","published-online":{"date-parts":[[2023,3]]}},"alternative-id":["s23063176"],"URL":"https:\/\/doi.org\/10.3390\/s23063176","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,3,16]]}}}