{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,20]],"date-time":"2025-05-20T17:48:10Z","timestamp":1747763290576,"version":"3.40.5"},"reference-count":39,"publisher":"Wiley","license":[{"start":{"date-parts":[[2023,12,22]],"date-time":"2023-12-22T00:00:00Z","timestamp":1703203200000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Applied Computational Intelligence and Soft Computing"],"published-print":{"date-parts":[[2023,12,22]]},"abstract":"<jats:p>This paper investigates the performance enhancement of base classifiers within the AdaBoost framework applied to medical datasets. Adaptive boosting (AdaBoost), being an instance of boosting, combines other classifiers to enhance their performance. We conducted a comprehensive experiment to assess the efficacy of twelve base classifiers with the AdaBoost framework, namely, Bayes network, decision stump, ZeroR, decision tree, Na\u00efve Bayes, J-48, voted perceptron, random forest, bagging, random tree, stacking, and AdaBoost itself. The experiments are carried out on five datasets from the medical domain based on various types of cancers, i.e., global cancer map (GCM), lymphoma-I, lymphoma-II, leukaemia, and embryonal tumours. The evaluation focuses on the accuracy, precision, and efficiency of the base classifiers in the AdaBoost framework. The results show that the performance of Na\u00efve Bayes, Bayes network, and voted perceptron is highly improved compared to the rest of the base classifiers, attaining accuracies as high as 94.74%, 97.78%, and 97.78%, respectively. The results also show that in most cases, the base classifiers perform better with AdaBoost compared to their performance, i.e., for voted perceptron, the accuracy is improved up to 13.34%.For bagging, it is improved by up to 7%. This research aims to identify such base classifiers with optimal boosting capabilities within the AdaBoost framework for medical datasets. The significance of these results is that they provide insight into the performance of the base classifiers when used in the boosting framework to enhance the classification performance of classifiers in scenarios where individual classifiers do not perform up to the mark.<\/jats:p>","DOI":"10.1155\/2023\/5542049","type":"journal-article","created":{"date-parts":[[2023,12,22]],"date-time":"2023-12-22T22:20:08Z","timestamp":1703283608000},"page":"1-10","source":"Crossref","is-referenced-by-count":3,"title":["Performance Augmentation of Base Classifiers Using Adaptive Boosting Framework for Medical Datasets"],"prefix":"10.1155","volume":"2023","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2809-1157","authenticated-orcid":true,"given":"Durr e","family":"Nayab","sequence":"first","affiliation":[{"name":"Department of Computer Systems Engineering, University of Engineering and Technology, Peshawar 25120, Pakistan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3323-2732","authenticated-orcid":true,"given":"Rehan Ullah","family":"Khan","sequence":"additional","affiliation":[{"name":"Department of Information Technology, College of Computer, Qassim University, Buraydah, Saudi Arabia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5323-6661","authenticated-orcid":true,"given":"Ali Mustafa","family":"Qamar","sequence":"additional","affiliation":[{"name":"Department of Computer Science, College of Computer, Qassim University, Buraydah, Saudi Arabia"}]}],"member":"311","reference":[{"first-page":"278","article-title":"AdaBoost Gabor Fisher classifier for face recognition","author":"S. Shan","key":"1"},{"key":"2","doi-asserted-by":"publisher","DOI":"10.1016\/j.ejor.2019.10.015"},{"key":"3","doi-asserted-by":"publisher","DOI":"10.1016\/j.biosystemseng.2020.03.021"},{"key":"4","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2020.103770"},{"key":"5","doi-asserted-by":"publisher","DOI":"10.1016\/j.acalib.2021.102450"},{"key":"6","doi-asserted-by":"publisher","DOI":"10.1006\/jcss.1997.1504"},{"issue":"1","key":"7","first-page":"511","article-title":"Rapid object detection using a boosted cascade of simple","volume":"1","author":"P. Viola","year":"2001","journal-title":"Proceedings of CVPR"},{"issue":"1","key":"8","first-page":"0","article-title":"Multi-class AdaBoost","volume":"4","author":"J. Zhu","year":"2006","journal-title":"Machine Learning"},{"key":"9","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2007.07.001"},{"issue":"11","key":"10","first-page":"771","article-title":"Palmprint recognition: a naive bayesian approach","volume":"31","author":"Z. Nisar","year":"2014","journal-title":"World Applied Sciences Journal"},{"key":"11","doi-asserted-by":"publisher","DOI":"10.1088\/1757-899x\/546\/5\/052068"},{"issue":"1","key":"12","first-page":"440","article-title":"The optimality of Na\u00efve Bayes","volume":"3","author":"H. Zhang","year":"2004","journal-title":"Machine Learning"},{"key":"13","doi-asserted-by":"publisher","DOI":"10.1016\/j.envsoft.2018.09.016"},{"first-page":"301","article-title":"An implementation of Na\u00efve Bayes classifier","author":"F. J. Yang","key":"14"},{"first-page":"489","article-title":"Ranking algorithm for named\u2013entity extraction: boosting and the Voted Perceptron","author":"M. Collens","key":"15"},{"key":"16","doi-asserted-by":"publisher","DOI":"10.1023\/a:1007662407062"},{"key":"17","doi-asserted-by":"publisher","DOI":"10.1002\/jemt.23326"},{"issue":"13","key":"18","first-page":"41","article-title":"Real-time training of voted perceptron for classification of EEG data","volume":"10","author":"I. Marti\u0161ius","year":"2013","journal-title":"International Journal of Artificial Intelligence"},{"key":"19","first-page":"70","article-title":"Bayesian networks","volume":"5","author":"B. Gal","year":"2007","journal-title":"Neural Networks"},{"key":"20","doi-asserted-by":"publisher","DOI":"10.1023\/a:1007465528199"},{"key":"21","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2012.12.051"},{"key":"22","doi-asserted-by":"publisher","DOI":"10.1504\/ijcse.2018.090450"},{"issue":"1","key":"23","first-page":"1","article-title":"Feature selection with conjunctions of Decision Stumps and learning from microarray data","volume":"9","author":"M. Shah","year":"2010","journal-title":"Machine Learning"},{"issue":"1","key":"24","first-page":"148","article-title":"Local additive regression of decision stumps","volume":"3","author":"S. B. kotsiants","year":"2006","journal-title":"Neural Networks"},{"first-page":"233","article-title":"Induction of one-level decision trees","author":"W. Iba","key":"25"},{"issue":"2","key":"26","first-page":"1","article-title":"Building decision trees from decision stumps","volume":"2","author":"L. Peter","year":"2010","journal-title":"Forum American Bar Association"},{"key":"27","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-011-9272-4"},{"key":"28","doi-asserted-by":"publisher","DOI":"10.1023\/a:1009887311454"},{"first-page":"114","article-title":"A practical differentially private random decision tree classifier","author":"G. Jagannathan","key":"29"},{"issue":"2","key":"30","first-page":"121","article-title":"A Bagging method using Decision Tress in the role of base classifiers","volume":"3","author":"K. Machova","year":"2006","journal-title":"Cybernetics"},{"key":"31","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2019.103319"},{"key":"32","doi-asserted-by":"publisher","DOI":"10.22266\/ijies2018.0228.05"},{"key":"33","doi-asserted-by":"publisher","DOI":"10.1109\/access.2021.3067816"},{"issue":"3","key":"34","first-page":"2039","article-title":"Consistency of Random Forests and other averaging classifiers","volume":"9","author":"L. Devroye","year":"2008","journal-title":"Journal of Machine Learning Research"},{"issue":"5","key":"35","first-page":"272","article-title":"Random forests and decision trees","volume":"9","author":"J. Ali","year":"2012","journal-title":"International Journal of Computer Science Issues (IJCSI)"},{"key":"36","doi-asserted-by":"publisher","DOI":"10.14569\/ijacsa.2011.021212"},{"key":"37","doi-asserted-by":"publisher","DOI":"10.1007\/s10586-017-1109-8"},{"issue":"10","key":"38","first-page":"426","article-title":"Analysis and prediction of diabetes diseases using machine learning algorithm: ensemble approach","volume":"4","author":"R. Joshi","year":"2017","journal-title":"International Research Journal of Engineering and Technology"},{"volume-title":"Cancer Gene Expression Datasets","year":"2015","author":"X. Zhang","key":"39"}],"container-title":["Applied Computational Intelligence and Soft Computing"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/downloads.hindawi.com\/journals\/acisc\/2023\/5542049.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/acisc\/2023\/5542049.xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/acisc\/2023\/5542049.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,22]],"date-time":"2023-12-22T22:20:23Z","timestamp":1703283623000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.hindawi.com\/journals\/acisc\/2023\/5542049\/"}},"subtitle":[],"editor":[{"given":"Kalapraveen","family":"Bagadi","sequence":"additional","affiliation":[]}],"short-title":[],"issued":{"date-parts":[[2023,12,22]]},"references-count":39,"alternative-id":["5542049","5542049"],"URL":"https:\/\/doi.org\/10.1155\/2023\/5542049","relation":{},"ISSN":["1687-9732","1687-9724"],"issn-type":[{"type":"electronic","value":"1687-9732"},{"type":"print","value":"1687-9724"}],"subject":[],"published":{"date-parts":[[2023,12,22]]}}}