{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T12:13:12Z","timestamp":1775650392480,"version":"3.50.1"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T00:00:00Z","timestamp":1772496000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T00:00:00Z","timestamp":1775606400000},"content-version":"vor","delay-in-days":36,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/100000308","name":"Carnegie Corporation of New York","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100000308","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Discov Artif Intell"],"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>Cervical cancer remains a major global health challenge, particularly in low-resource settings where access to timely and accurate diagnostic services is limited. Although Pap smear screening has significantly reduced cervical cancer mortality, manual cytological assessment is labor-intensive, prone to inter-observer variability, and dependent on expert availability. To address these limitations, this study presents a hybrid classification framework that integrates deep feature extraction with interpretable rule-based learning for cervical cancer diagnosis from Pap smear images. High-level visual features are extracted using a pre-trained InceptionV3 convolutional neural network, while classification is performed using Convolutional Tsetlin Machines (CTMs), which employ logical clauses learned through Tsetlin automata to enable transparent decision-making. The proposed framework was evaluated on a publicly available liquid-based cytology dataset comprising four diagnostic classes: Negative for Intraepithelial Malignancy, Low Squamous Intraepithelial Lesion, High Squamous Intraepithelial Lesion, and Squamous Cell Carcinoma. Experimental results obtained under stratified cross-validation demonstrate strong and consistent performance, achieving an accuracy of 99.96%, precision of 98.99%, recall of 98.96%, and an F1-score of 98.98%. Beyond high classification accuracy, the intrinsic interpretability of CTMs allows inspection of learned logical rules, supporting clinical transparency and trust. These findings indicate that combining deep transfer learning with interpretable machine learning offers a promising direction for reliable and explainable cervical cancer screening systems, particularly in resource-constrained healthcare environments.<\/jats:p>","DOI":"10.1007\/s44163-026-01057-x","type":"journal-article","created":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T17:07:35Z","timestamp":1772557655000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhanced cervical cancer classification using convolutional tsetlin machines with transfer learning"],"prefix":"10.1007","volume":"6","author":[{"given":"Emmanuel","family":"Ahishakiye","sequence":"first","affiliation":[]},{"given":"Leonard","family":"Nkalubo","sequence":"additional","affiliation":[]},{"given":"Fredrick","family":"Kanobe","sequence":"additional","affiliation":[]},{"given":"Danison","family":"Taremwa","sequence":"additional","affiliation":[]},{"given":"Bartha Alexandra","family":"Nantongo","sequence":"additional","affiliation":[]},{"given":"Shallon","family":"Ahimbisibwe","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,3,3]]},"reference":[{"key":"1057_CR1","unstructured":"WHO. Cervical cancer. Accessed: Jul. 30, 2024. [Online]. Available: https:\/\/www.who.int\/news-room\/fact-sheets\/detail\/cervical-cancer#:~:text=Cervical cancer is the fourth,350 000 deaths in 2022."},{"issue":"11","key":"1057_CR2","doi-asserted-by":"publisher","first-page":"31312","DOI":"10.7759\/cureus.31312","volume":"14","author":"S Basoya","year":"2022","unstructured":"Basoya S, Anjankar A Cervical Cancer: Early Detection and Prevention in Reproductive Age Group. Cureus. 2022;14(11):31312. https:\/\/doi.org\/10.7759\/cureus.31312.","journal-title":"Cureus"},{"key":"1057_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3389\/fonc.2022.851367","volume":"12","author":"X Hou","year":"2022","unstructured":"Hou X, Shen G, Zhou L, Li Y, Wang T, Ma X. Artificial intelligence in cervical cancer screening and diagnosis. Front Oncol. 2022;12:1\u201313. https:\/\/doi.org\/10.3389\/fonc.2022.851367","journal-title":"Front Oncol"},{"key":"1057_CR4","doi-asserted-by":"publisher","DOI":"10.4103\/apjon.apjon","author":"PL Sachan","year":"2018","unstructured":"Sachan PL, Singh M, Patel ML, Sachan R. Cervical cancer screening using pap smear test and clinical correlation. Med Forum Mon. 2018. https:\/\/doi.org\/10.4103\/apjon.apjon.","journal-title":"Med Forum Mon"},{"key":"1057_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3390\/diagnostics13193079","volume":"13","author":"YM Lee","year":"2023","unstructured":"Lee YM, Lee B, Cho NH, Park JH. Beyond the microscope: a technological overture for cervical cancer detection. Diagnostics. 2023;13:1\u201317. https:\/\/doi.org\/10.3390\/diagnostics13193079.","journal-title":"Diagnostics"},{"key":"1057_CR6","doi-asserted-by":"publisher","unstructured":"Zhao X, Wang L, Zhang Y, Han X, Deveci M, Parmar M. A review of convolutional neural networks in computer vision. no 4 Springer Neth. 2024. p. 57. https:\/\/doi.org\/10.1007\/s10462-024-10721-6.","DOI":"10.1007\/s10462-024-10721-6"},{"key":"1057_CR7","doi-asserted-by":"publisher","first-page":"327","DOI":"10.1016\/j.aej.2024.05.102","volume":"103","author":"KAA Mahmoud","year":"2024","unstructured":"Mahmoud KAA, Badr MM, Elmalhy NA, Hamdy RA, Ahmed S, Mordi AA. Transfer learning by fine-tuning pre-trained convolutional neural network architectures for switchgear fault detection using thermal imaging. Alexandria Eng J. 2024;103:327\u201342. https:\/\/doi.org\/10.1016\/j.aej.2024.05.102","journal-title":"Alexandria Eng J"},{"key":"1057_CR8","doi-asserted-by":"publisher","DOI":"10.1186\/s40537-021-00444-8","author":"L Alzubaidi","year":"2021","unstructured":"Alzubaidi L, et al Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. J Big Data. 2021. https:\/\/doi.org\/10.1186\/s40537-021-00444-8.","journal-title":"J Big Data"},{"key":"1057_CR9","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-023-10466-8","author":"SF Ahmed","year":"2023","unstructured":"Ahmed SF, et al. Deep learning modelling techniques: current progress, applications, advantages, and challenges. no 0123456789 Springer Neth. 2023. https:\/\/doi.org\/10.1007\/s10462-023-10466-8.","journal-title":"no 0123456789 Springer Neth"},{"key":"1057_CR10","unstructured":"Granmo O-C. The Tsetlin machine - a game theoretic bandit driven approach to optimal pattern recognition with propositional logic, no. 1, 2018. Available: http:\/\/arxiv.org\/abs\/1804.01508"},{"key":"1057_CR11","unstructured":"Granmo O-C, Glimsdal S, Jiao L, Goodwin M, Omlin CW, Berge GT. The Convolutional Tsetlin Machine, no. d, 2019. Available: http:\/\/arxiv.org\/abs\/1905.09688"},{"key":"1057_CR12","doi-asserted-by":"publisher","first-page":"115134","DOI":"10.1109\/ACCESS.2019.2935416","volume":"7","author":"GT Berge","year":"2019","unstructured":"Berge GT, Granmo OC, Tveit TO, Goodwin M, Jiao L, Matheussen BV. Using the tsetlin machine to learn human-interpretable rules for high-accuracy text categorization with medical applications. IEEE Access. 2019;7:115134\u201346. https:\/\/doi.org\/10.1109\/ACCESS.2019.2935416.","journal-title":"IEEE Access"},{"issue":"4","key":"1057_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1111\/exsy.12873","volume":"40","author":"R Saha","year":"2023","unstructured":"Saha R, Granmo OC, Goodwin M. Using Tsetlin Machine to discover interpretable rules in natural language processing applications. Expert Syst. 2023;40(4):1\u201312. https:\/\/doi.org\/10.1111\/exsy.12873.","journal-title":"Expert Syst"},{"key":"1057_CR14","doi-asserted-by":"publisher","first-page":"564","DOI":"10.1007\/978-3-030-22999-3_49","volume":"11606 LNAI","author":"KD Abeyrathna","year":"2019","unstructured":"Abeyrathna KD, Granmo OC, Zhang X, Goodwin M. A scheme for continuous input to the tsetlin machine with applications to forecasting disease outbreaks. Lect Notes Comput Sci (including Subser Lect Notes Artif Intell Lect Notes Bioinformatics). 2019;11606 LNAI:564\u201378. https:\/\/doi.org\/10.1007\/978-3-030-22999-3_49.","journal-title":"Lect Notes Comput Sci (including Subser Lect Notes Artif Intell Lect Notes Bioinformatics)"},{"key":"1057_CR15","unstructured":"Nymark PS. Tsetlin Machine for analysis of healthcare data: a comparison with standard machine learning algorithms. Nowegian University of Life Sciences. 2021."},{"issue":"1","key":"1057_CR16","doi-asserted-by":"publisher","first-page":"121","DOI":"10.1007\/s10844-021-00682-5","volume":"59","author":"R Saha","year":"2022","unstructured":"Saha R, Granmo OC, Zadorozhny VI, Goodwin M. A relational tsetlin machine with applications to natural language understanding. J Intell Inf Syst. 2022;59(1):121\u201348. https:\/\/doi.org\/10.1007\/s10844-021-00682-5.","journal-title":"J Intell Inf Syst"},{"issue":"4","key":"1057_CR17","doi-asserted-by":"publisher","first-page":"247","DOI":"10.5468\/OGS.25041","volume":"68","author":"SS Abrar","year":"2025","unstructured":"Abrar SS, Isa SAM, Hairon SM, Ismail MP, Kadir MNBNA. Recent advances in applications of machine learning in cervical cancer research: a focus on prediction models. Obstet Gynecol Sci. 2025;68(4):247. https:\/\/doi.org\/10.5468\/OGS.25041","journal-title":"Obstet Gynecol Sci"},{"key":"1057_CR18","doi-asserted-by":"publisher","first-page":"100787","DOI":"10.1016\/J.EIJ.2025.100787","volume":"32","author":"EHI Eliwa","year":"2025","unstructured":"Eliwa EHI, Abd El-Hafeez T A robust deep learning pipeline for multi-class cervical cancer cell identification. Egypt Inf J. 2025;32:100787. https:\/\/doi.org\/10.1016\/J.EIJ.2025.100787.","journal-title":"Egypt Inf J"},{"key":"1057_CR19","doi-asserted-by":"publisher","first-page":"31","DOI":"10.1016\/j.cmpb.2016.10.001","volume":"138","author":"K Bora","year":"2017","unstructured":"Bora K, Chowdhury M, Mahanta LB, Kundu MK, Das AK. Automated classification of Pap smear images to detect cervical dysplasia. Comput Methods Programs Biomed. 2017;138:31\u201347. https:\/\/doi.org\/10.1016\/j.cmpb.2016.10.001.","journal-title":"Comput Methods Programs Biomed"},{"issue":"2","key":"1057_CR20","doi-asserted-by":"publisher","first-page":"233","DOI":"10.1109\/TITB.2010.2087030","volume":"15","author":"ME Plissiti","year":"2011","unstructured":"Plissiti ME, Nikou C, Charchanti A. Automated detection of cell nuclei in pap smear images using morphological reconstruction and clustering. IEEE Trans Inf Technol Biomed. 2011;15(2):233\u201341. https:\/\/doi.org\/10.1109\/TITB.2010.2087030","journal-title":"IEEE Trans Inf Technol Biomed"},{"issue":"11","key":"1057_CR21","doi-asserted-by":"publisher","first-page":"3447","DOI":"10.31557\/APJCP.2019.20.11.3447","volume":"20","author":"PB Shanthi","year":"2019","unstructured":"Shanthi PB, Faruqi F, Hareesha KS, Kudva R. Deep Convolution Neural Network for malignancy detection and classification in microscopic uterine cervix cell images. Asian Pac J Cancer Prev. 2019;20(11):3447\u201356. https:\/\/doi.org\/10.31557\/APJCP.2019.20.11.3447.","journal-title":"Asian Pac J Cancer Prev"},{"issue":"4","key":"1057_CR22","doi-asserted-by":"publisher","first-page":"65","DOI":"10.1007\/S11063-025-11770-W","volume":"257","author":"S Dayalane","year":"2025","unstructured":"Dayalane S, Murugesan S, Mathivanan SK, Rajadurai H, Panneer Selvam K, Shah MA. Cervical cancer classification using deep learning approach using colposcopy images. Neural Process. Lett. 2025;257(4):65. https:\/\/doi.org\/10.1007\/S11063-025-11770-W.","journal-title":"Lett"},{"issue":"1","key":"1057_CR23","doi-asserted-by":"publisher","first-page":"3240","DOI":"10.1038\/s41598-025-87953-1","volume":"151","author":"Y Zhang","year":"2025","unstructured":"Zhang Y, Ning C, Yang W. An automatic cervical cell classification model based on improved DenseNet. Sci Rep. 2025;15(1):3240. https:\/\/doi.org\/10.1038\/s41598-025-87953-1.","journal-title":"Sci Rep"},{"issue":"1","key":"1057_CR24","doi-asserted-by":"publisher","first-page":"3945","DOI":"10.1038\/s41598-024-74531-0","volume":"151","author":"H Kaur","year":"2025","unstructured":"Kaur H, Sharma R, Kaur J Comparison of deep transfer learning models for classification of cervical cancer from pap smear images. Sci Rep. 2025;151(1):3945. https:\/\/doi.org\/10.1038\/s41598-024-74531-0.","journal-title":"Sci Rep"},{"key":"1057_CR25","doi-asserted-by":"publisher","DOI":"10.1007\/s44163-024-00185-6","author":"E Ahishakiye","year":"2024","unstructured":"Ahishakiye E, Kanobe F Optimizing cervical cancer classification using transfer learning with deep gaussian processes and support vector machines. Discov Artif Intell. 2024. https:\/\/doi.org\/10.1007\/s44163-024-00185-6.","journal-title":"Discov Artif Intell"},{"key":"1057_CR26","doi-asserted-by":"publisher","first-page":"101587","DOI":"10.1016\/J.IMU.2024.101587","volume":"51","author":"ID Mienye","year":"2024","unstructured":"Mienye ID, et al A survey of explainable artificial intelligence in healthcare: concepts, applications, and challenges. Inf Med Unlocked. 2024;51:101587. https:\/\/doi.org\/10.1016\/J.IMU.2024.101587.","journal-title":"Inf Med Unlocked"},{"issue":"3","key":"1057_CR27","doi-asserted-by":"publisher","first-page":"247","DOI":"10.1109\/JPROC.2021.3060483","volume":"109","author":"W Samek","year":"2021","unstructured":"Samek W, Montavon G, Lapuschkin S, Anders CJ, M\u00fcller KR. Explaining Deep neural networks and beyond: a review of methods and applications. Proc IEEE. 2021;109(3):247\u201378. https:\/\/doi.org\/10.1109\/JPROC.2021.3060483","journal-title":"Proc IEEE"},{"key":"1057_CR28","doi-asserted-by":"publisher","first-page":"100178","DOI":"10.1016\/J.ACAGS.2024.100178","volume":"23","author":"AS Antonini","year":"2024","unstructured":"Antonini AS, et al Machine Learning model interpretability using SHAP values: application to Igneous Rock Classification task. Appl Comput Geosci. 2024;23:100178. https:\/\/doi.org\/10.1016\/J.ACAGS.2024.100178.","journal-title":"Appl Comput Geosci"},{"key":"1057_CR29","doi-asserted-by":"publisher","first-page":"110787","DOI":"10.1016\/J.EJRAD.2023.110787","volume":"162","author":"K Borys","year":"2023","unstructured":"Borys K, et al Explainable AI in medical imaging: an overview for clinical practitioners \u2013 Saliency-based XAI approaches. Eur J Radiol. 2023;162:110787. https:\/\/doi.org\/10.1016\/J.EJRAD.2023.110787.","journal-title":"Eur J Radiol"},{"key":"1057_CR30","doi-asserted-by":"publisher","first-page":"110786","DOI":"10.1016\/j.ejrad.2023.110786","volume":"162","author":"K Borys","year":"2023","unstructured":"Borys K, et al Explainable AI in medical imaging: an overview for clinical practitioners \u2013 Beyond saliency-based XAI approaches. Eur J Radiol. 2023;162:110786. https:\/\/doi.org\/10.1016\/j.ejrad.2023.110786.","journal-title":"Eur J Radiol"},{"key":"1057_CR31","doi-asserted-by":"publisher","first-page":"109370","DOI":"10.1016\/J.COMPELECENG.2024.109370","volume":"118","author":"Z Sadeghi","year":"2024","unstructured":"Sadeghi Z, et al A review of Explainable Artificial Intelligence in healthcare. Comput Electr Eng Aug. 2024;118:109370. https:\/\/doi.org\/10.1016\/J.COMPELECENG.2024.109370.","journal-title":"Comput Electr Eng Aug"},{"issue":"17","key":"1057_CR32","doi-asserted-by":"publisher","first-page":"2154","DOI":"10.3390\/HEALTHCARE13172154","volume":"13","author":"Q Abbas","year":"2025","unstructured":"Abbas Q, Jeong W, Lee SW Explainable AI in clinical decision support systems: a meta-analysis of methods, applications, and usability challenges. Healthcare. 2025;13(17):2154. https:\/\/doi.org\/10.3390\/HEALTHCARE13172154.","journal-title":"Healthcare"},{"key":"1057_CR33","unstructured":"Hou J et al. Self-explainable ai for medical image analysis: a survey and new outlooks. 2024. p. 1. Accessed: Dec. 21, 2025. Available: https:\/\/arxiv.org\/pdf\/2410.02331"},{"issue":"12","key":"1057_CR34","doi-asserted-by":"publisher","first-page":"9620","DOI":"10.21037\/QIMS-24-723\/COIF","volume":"14","author":"Z Teng","year":"2024","unstructured":"Teng Z, et al. Dec., A literature review of artificial intelligence (AI) for medical image segmentation: from AI and explainable AI to trustworthy. AI Quant Imaging Med Surg. 2024;14(12):9620\u201352. https:\/\/doi.org\/10.21037\/QIMS-24-723\/COIF","journal-title":"Imaging Med Surg"},{"key":"1057_CR35","doi-asserted-by":"publisher","DOI":"10.17632\/zddtpgzv63.4","author":"E Hussain","year":"2019","unstructured":"Hussain E Liquid based cytology pap smear images for multi-class diagnosis of cervical cancer. Mendeley Data. 2019. https:\/\/doi.org\/10.17632\/zddtpgzv63.4.","journal-title":"Mendeley Data"},{"key":"1057_CR36","doi-asserted-by":"publisher","unstructured":"Ribeiro MT, Singh S, Guestrin C. \u2018Why should i trust you?\u2019: Explaining the predictions of any classifier. In NAACL-HLT 2016\u20132016 conference of the north american chapter of the association for computational linguistics: human language technologies. 2016. pp. 97\u2013101. https:\/\/doi.org\/10.18653\/v1\/n16-3020","DOI":"10.18653\/v1\/n16-3020"},{"key":"1057_CR37","first-page":"4766","volume":"2017","author":"SM Lundberg","year":"2017","unstructured":"Lundberg SM, Lee SI. A unified approach to interpreting model predictions. Adv Neural Inf Process Syst 2017;2017:4766\u20134775.","journal-title":"Adv Neural Inf Process Syst"},{"issue":"5","key":"1057_CR38","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat Mach Intell. 2019;1(5):206\u201315. https:\/\/doi.org\/10.1038\/s42256-019-0048-x","journal-title":"Nat Mach Intell"},{"key":"1057_CR39","doi-asserted-by":"publisher","unstructured":"Ghorbani A, Abid A, Zou J. Interpretation of neural networks is fragile. In Proceedings of AAAI Conference on Artificial Intelligence. 2019. vol. 33, no. 01, pp. 3681\u20133688 https:\/\/doi.org\/10.1609\/AAAI.V33I01.33013681.","DOI":"10.1609\/AAAI.V33I01.33013681"},{"issue":"6","key":"1057_CR40","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1042\/BSR20181769","volume":"38","author":"M Wu","year":"2018","unstructured":"Wu M, Yan C, Liu H, Liu Q, Yin Y. Automatic classification of cervical cancer from cytological images by using convolutional neural network. Biosci Rep. 2018;38(6):1\u20139. https:\/\/doi.org\/10.1042\/BSR20181769.","journal-title":"Biosci Rep"},{"key":"1057_CR41","doi-asserted-by":"publisher","DOI":"10.22514\/ejgo.2025.012","author":"I Pacal","year":"2025","unstructured":"Pacal I. Investigating deep learning approaches for cervical cancer diagnosis: a focus on modern image-based models. Eur J Gynaecol Oncol. 2025. https:\/\/doi.org\/10.22514\/ejgo.2025.012.","journal-title":"Eur J Gynaecol Oncol"},{"key":"1057_CR42","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3389\/frobt.2024.1444763","volume":"11","author":"M Ennab","year":"2024","unstructured":"Ennab M, Mcheick H. Enhancing interpretability and accuracy of AI models in healthcare: a comprehensive review on challenges and future directions. Front Robot AI 2024;11:1\u201316 https:\/\/doi.org\/10.3389\/frobt.2024.1444763","journal-title":"Front Robot AI"}],"container-title":["Discover Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44163-026-01057-x","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44163-026-01057-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44163-026-01057-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T11:27:17Z","timestamp":1775647637000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44163-026-01057-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,3]]},"references-count":42,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2026,12]]}},"alternative-id":["1057"],"URL":"https:\/\/doi.org\/10.1007\/s44163-026-01057-x","relation":{},"ISSN":["2731-0809"],"issn-type":[{"value":"2731-0809","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3,3]]},"assertion":[{"value":"23 November 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 February 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 March 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics and consent to participate"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"The authors declare no competing interests.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"301"}}