{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T08:33:31Z","timestamp":1768725211621,"version":"3.49.0"},"reference-count":34,"publisher":"MDPI AG","issue":"2","license":[{"start":{"date-parts":[[2021,1,26]],"date-time":"2021-01-26T00:00:00Z","timestamp":1611619200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Algorithms"],"abstract":"<jats:p>Boosting of the ensemble learning model has made great progress, but most of the methods are Boosting the single mode. For this reason, based on the simple multiclass enhancement framework that uses local similarity as a weak learner, it is extended to multimodal multiclass enhancement Boosting. First, based on the local similarity as a weak learner, the loss function is used to find the basic loss, and the logarithmic data points are binarized. Then, we find the optimal local similarity and find the corresponding loss. Compared with the basic loss, the smaller one is the best so far. Second, the local similarity of the two points is calculated, and then the loss is calculated by the local similarity of the two points. Finally, the text and image are retrieved from each other, and the correct rate of text and image retrieval is obtained, respectively. The experimental results show that the multimodal multi-class enhancement framework with local similarity as the weak learner is evaluated on the standard data set and compared with other most advanced methods, showing the experience proficiency of this method.<\/jats:p>","DOI":"10.3390\/a14020037","type":"journal-article","created":{"date-parts":[[2021,1,26]],"date-time":"2021-01-26T12:03:57Z","timestamp":1611662637000},"page":"37","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["The Study of Multiple Classes Boosting Classification Method Based on Local Similarity"],"prefix":"10.3390","volume":"14","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5127-8403","authenticated-orcid":false,"given":"Shixun","family":"Wang","sequence":"first","affiliation":[{"name":"College of Computer and Information Engineering, Henan Normal University, Xinxiang 453007, China"},{"name":"Engineering Lab of Intelligence Business &amp; Internet of Things, Xinzheng 451191, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2320-3743","authenticated-orcid":false,"given":"Qiang","family":"Chen","sequence":"additional","affiliation":[{"name":"College of Computer and Information Engineering, Henan Normal University, Xinxiang 453007, China"}]}],"member":"1968","published-online":{"date-parts":[[2021,1,26]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"175","DOI":"10.1016\/j.ins.2015.08.030","article-title":"Fast image classification by boosting fuzzy classifiers","volume":"327","author":"Korytkowski","year":"2016","journal-title":"Inf. Sci."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"1793","DOI":"10.1109\/TGRS.2015.2488681","article-title":"Scene classification via a gradient boosting random convolutional network framework","volume":"54","author":"Zhang","year":"2016","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_3","unstructured":"Ioannis, E.L., and Pintelas, P. (2020). On ensemble techniques of weight-constrained neural networks. Evol. Syst., 1\u201313."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"101847","DOI":"10.1016\/j.artmed.2020.101847","article-title":"Early temporal prediction of type 2 diabetes risk condition from a general practitioner electronic health record: A multiple instance boosting approach","volume":"105","author":"Bernardini","year":"2020","journal-title":"Artif. Intell. Med."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"4648","DOI":"10.1109\/TIP.2017.2718189","article-title":"Action recognition using 3d histograms of texture and a multi-class boosting classifier","volume":"26","author":"Zhang","year":"2017","journal-title":"IEEE Trans. Image Process."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"1951","DOI":"10.1007\/s10994-019-05797-z","article-title":"Boosting as a kernel-based method","volume":"108","author":"Aravkin","year":"2016","journal-title":"Mach. Learn."},{"key":"ref_7","first-page":"835357","article-title":"Analysis of generalization ability for different adaboost variants based on classification and regression trees","volume":"2015","author":"Nagahashi","year":"2015","journal-title":"J. Electr. Comput. Eng."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"119","DOI":"10.1006\/jcss.1997.1504","article-title":"A decision-theoretic generalization of on-line learning and an application to boosting","volume":"55","author":"Freund","year":"1997","journal-title":"J. Comput. Syst. Sci."},{"key":"ref_9","first-page":"32","article-title":"A literature review on supervised machine learning algorithms and boosting process","volume":"169","author":"Praveena","year":"2017","journal-title":"Int. J. Comput. Appl."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"351","DOI":"10.1016\/j.neucom.2020.03.064","article-title":"AdaBoost-CNN: An adaptive boosting algorithm for convolutional neural networks to classify multi-class imbalanced datasets using transfer learning","volume":"404","author":"Taherkhani","year":"2020","journal-title":"Neurocomputing"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Livieris, I., Kiriakidou, N., Kanavos, A., Tampakas, V., and Pintelas, P. (2018). On ensemble ssl algorithms for credit scoring problem. Informatics, 5.","DOI":"10.3390\/informatics5040040"},{"key":"ref_12","first-page":"437","article-title":"A theory of multiclass boosting","volume":"14","author":"Mukherjee","year":"2011","journal-title":"J. Mach. Learn. Res."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Masnadi-Shirazi, H., Mahadevan, V., and Vasconcelos, N. (2010, January 13\u201318). On the design of robust classifiers for computer vision. Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA.","DOI":"10.1109\/CVPR.2010.5540136"},{"key":"ref_14","unstructured":"Appel, R., and Perona, P. (2017, January 6\u201311). A simple multi-class boosting framework with theoretical guarantees and empirical proficiency. Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Piewak, F., Pinggera, P., Schfer, M., Peter, D., Schwarz, B., Schneider, N., Enzweiler, M., Pfeiffer, D., and Zollner, M. (2018, January 8\u201314). Boosting lidar-based semantic labeling by cross-modal training data generation. Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany.","DOI":"10.1007\/978-3-030-11024-6_39"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Shen, C., and Hao, Z. (2011, January 20\u201325). A direct formulation for totally-corrective multi-class boosting. Proceedings of the Computer Vision & Pattern Recognition, Providence, RI, USA.","DOI":"10.1109\/CVPR.2011.5995554"},{"key":"ref_17","unstructured":"Saberian, M., and Vasconcelos, N. (2011, January 12\u201314). Multiclass Boosting: Theory and Algorithms. Proceedings of the 24th International Conference on Neural Information Processing Systems, Granada, Spain."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.neucom.2019.05.040","article-title":"Multimodal multiclass boosting and its application to cross-modal retrieval","volume":"357","author":"Wang","year":"2019","journal-title":"Neurocomputing"},{"key":"ref_19","first-page":"113","article-title":"Reducing multiclass to binary: A unifying approach for margin classifiers","volume":"1","author":"Allwein","year":"2000","journal-title":"J. Mach. Learn. Res."},{"key":"ref_20","unstructured":"Coxeter, H.S.M. (1973). Regular Polytopes, Dover Publications."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"263","DOI":"10.1613\/jair.105","article-title":"Solving multiclass learning problems via error-correcting output codes","volume":"2","author":"Dietterich","year":"1995","journal-title":"J. Artif. Intell. Res."},{"key":"ref_22","unstructured":"Duda, R.O., Hart, P.E., and Stork, D.G. (2004). Pattern Classification, Wiley."},{"key":"ref_23","first-page":"189","article-title":"Multiclass boosting for weak classifiers","volume":"6","author":"Pfeiffer","year":"2005","journal-title":"J. Mach. Learn. Res."},{"key":"ref_24","first-page":"2551","article-title":"Vc theory of large margin multi-category classifiers","volume":"8","author":"Guermeur","year":"2007","journal-title":"J. Mach. Learn. Res."},{"key":"ref_25","unstructured":"Wiley, R. (1999). An Introduction to Derivatives, Harcourt Brace College Publishers."},{"key":"ref_26","unstructured":"Mason, L., Baxter, J., Bartlett, P., and Frean, M. (December, January 29). Boosting Algorithms as Gradient Descent. Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA."},{"key":"ref_27","first-page":"131","article-title":"Evidence contrary to the statistical view of boosting","volume":"9","author":"Mease","year":"2008","journal-title":"C4 Programs Mach. Learn."},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Saberian, M.J., Masnadi-Shirazi, H., and Vasconcelos, N. (2011, January 20\u201325). TaylorBoost: First and second-order boosting algorithms with explicit margin control. Proceedings of the Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA.","DOI":"10.1109\/CVPR.2011.5995605"},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"297","DOI":"10.1023\/A:1007614523901","article-title":"Improved boosting algorithms using confidence-rated predictions","volume":"37","author":"Schapire","year":"1999","journal-title":"Mach. Learn."},{"key":"ref_30","first-page":"3185","article-title":"Statistical learning theory","volume":"41","author":"Vapnik","year":"1998","journal-title":"Encycl. Ences Learn."},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Sun, Y., Todorovic, S., Li, J., and Wu, D. (2005, January 7\u201311). Unifying the error-correcting and output-code AdaBoost within the margin framework. Machine Learning. Proceedings of the Twenty-Second International Conference (ICML 2005), Bonn, Germany.","DOI":"10.1145\/1102351.1102461"},{"key":"ref_32","unstructured":"Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep Learning, The MIT Press."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"2089","DOI":"10.1109\/TPAMI.2014.2315792","article-title":"Structboost: Boosting methods for predicting structured output variables","volume":"36","author":"Shen","year":"2014","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_34","unstructured":"Appel, R., Burgos-Artizzu, X., and Perona, P. (2016). Improved Multi-Class Cost-Sensitive Boosting via Estimation of the Minimum-Risk Class. arXiv."}],"container-title":["Algorithms"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1999-4893\/14\/2\/37\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T05:15:43Z","timestamp":1760159743000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1999-4893\/14\/2\/37"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,1,26]]},"references-count":34,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2021,2]]}},"alternative-id":["a14020037"],"URL":"https:\/\/doi.org\/10.3390\/a14020037","relation":{},"ISSN":["1999-4893"],"issn-type":[{"value":"1999-4893","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,1,26]]}}}