{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,4]],"date-time":"2026-02-04T03:14:45Z","timestamp":1770174885596,"version":"3.49.0"},"reference-count":47,"publisher":"SAGE Publications","issue":"1","license":[{"start":{"date-parts":[[2020,6,24]],"date-time":"2020-06-24T00:00:00Z","timestamp":1592956800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/journals.sagepub.com\/page\/policies\/text-and-data-mining-license"}],"content-domain":{"domain":["journals.sagepub.com"],"crossmark-restriction":true},"short-container-title":["Journal of Intelligent &amp; Fuzzy Systems"],"published-print":{"date-parts":[[2020,7,17]]},"abstract":"<jats:p>\n                    Learning from high-dimensional imbalanced data is prevalent in many vital real-world applications, which poses a severe challenge to traditional data mining and machine learning algorithms. The existing works generally use dimension reduction methods to deal with the curse of dimensionality, then apply traditional imbalance learning techniques to combat the problem of class imbalance. However, dimensionality reduction may cause the loss of useful information, especially for the minority classes. This paper introduces an ensemble-based method, HIBoost, to directly handle the imbalanced learning problem in high dimensional space. HIBoost takes into account the inherent high-dimensional hubness phenomenon, i.e., high-dimensional data tends to contain the singular points (hubs and anti-hubs) which frequently or rarely occur in\n                    <jats:italic>k<\/jats:italic>\n                    -nearest neighbors of other points. For the singular hubs and anti-hubs induced by high dimension, HIBoost introduces a discount factor to restrict the weight growth of them in the process of updating weight, so that the risk of over fitting can be reduced when training component classifiers. For class imbalance problem, HIBoost uses SMOTE to balance the training data in each iteration so as to alleviate the prediction bias of component classifiers. Experimental results based on sixteen high-dimensional imbalanced data sets demonstrate the effectiveness of HIBoost.\n                  <\/jats:p>","DOI":"10.3233\/jifs-190821","type":"journal-article","created":{"date-parts":[[2020,6,26]],"date-time":"2020-06-26T17:51:04Z","timestamp":1593193864000},"page":"133-144","update-policy":"https:\/\/doi.org\/10.1177\/sage-journals-update-policy","source":"Crossref","is-referenced-by-count":5,"title":["HIBoost: A hubness-aware ensemble learning algorithm for high-dimensional imbalanced data classification"],"prefix":"10.1177","volume":"39","author":[{"given":"Qin","family":"Wu","sequence":"first","affiliation":[{"name":"College of Information Science and Engineering, Hunan University, Changsha, P.R. China"}]},{"given":"Yaping","family":"Lin","sequence":"additional","affiliation":[{"name":"College of Information Science and Engineering, Hunan University, Changsha, P.R. China"}]},{"given":"Tuanfei","family":"Zhu","sequence":"additional","affiliation":[{"name":"College of Computer Engineering and Applied Mathematics, Changsha University, Changsha, China"}]},{"given":"Yue","family":"Zhang","sequence":"additional","affiliation":[{"name":"College of Information Science and Engineering, Hunan University, Changsha, P.R. China"}]}],"member":"179","published-online":{"date-parts":[[2020,6,24]]},"reference":[{"key":"e_1_3_2_2_2","first-page":"2487","article-title":"Hubs in space: Popular nearest neighbors in high-dimensional data","volume":"11","author":"Radovanovi\u0107 M.","year":"2010","unstructured":"Radovanovi\u0107M., NanopoulosA. and Ivanovi\u0107M., Hubs in space: Popular nearest neighbors in high-dimensional data, Journal of Machine Learning Research 11 (2010), 2487\u20132531.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_3_2","doi-asserted-by":"crossref","unstructured":"HeH. and GarciaE.A. Learning from imbalanced data IEEE Transactions on Knowledge & Data Engineering (2008) 1263\u20131284.","DOI":"10.1109\/TKDE.2008.239"},{"key":"e_1_3_2_4_2","doi-asserted-by":"crossref","unstructured":"BlagusR. and LusaL. Class prediction for high-dimensional class-imbalanced data BMC Bioinformatics (2010) 1\u201317.","DOI":"10.1186\/1471-2105-11-523"},{"key":"e_1_3_2_5_2","doi-asserted-by":"crossref","unstructured":"LinW.J. and ChenJ.J. Class-imbalanced classifiers for high-dimensional data Briefings in Bioinformatics (2013) 13\u201326.","DOI":"10.1093\/bib\/bbs006"},{"key":"e_1_3_2_6_2","doi-asserted-by":"crossref","unstructured":"MaldonadoS. WeberR. and FamiliF. Feature selection for high-dimensional class-imbalanced data sets using Support Vector Machines Information Sciences (2014) 228\u2013246.","DOI":"10.1016\/j.ins.2014.07.015"},{"key":"e_1_3_2_7_2","doi-asserted-by":"crossref","unstructured":"MoayedikiaA. OngK.L. and BooY.L. Feature selection for high dimensional imbalanced class data using harmony search Engineering Applications of Artificial Intelligence (2017) 38\u201349.","DOI":"10.1016\/j.engappai.2016.10.008"},{"key":"e_1_3_2_8_2","doi-asserted-by":"crossref","unstructured":"TomasevN. RadovanovicM. and MladenicD. The role of hubness in clustering high-dimensional data IEEE Transactions on Knowledge and Data Engineering (2014) 739\u2013751.","DOI":"10.1109\/TKDE.2013.25"},{"key":"e_1_3_2_9_2","doi-asserted-by":"crossref","unstructured":"BreimanL. Bagging predictors Machine Learning (1996) 123\u2013140.","DOI":"10.1007\/BF00058655"},{"key":"e_1_3_2_10_2","doi-asserted-by":"crossref","unstructured":"SvetnikV. WangT. and TongC. Boosting: an ensemble learning tool for compound classification and QSAR modeling Journal of Chemical Information & Modeling (2005) 786\u2013799.","DOI":"10.1021\/ci0500379"},{"key":"e_1_3_2_11_2","doi-asserted-by":"crossref","unstructured":"ZhangZ. LiY. and WangC. et al. An Ensemble Learning Method for Wireless Multimedia Device Identification[J] Security and Communication Networks (2018).","DOI":"10.1155\/2018\/5264526"},{"key":"e_1_3_2_12_2","doi-asserted-by":"crossref","unstructured":"GalarM. FernandezA. and BarrenecheaE. A Review on Ensembles for the Class Imbalance Problem: Bagging Boosting and Hybrid-Based Approaches IEEE Transactions on Systems Man & Cybernetics (2012) 463\u2013484.","DOI":"10.1109\/TSMCC.2011.2161285"},{"key":"e_1_3_2_13_2","doi-asserted-by":"crossref","unstructured":"ChawlaN.V. BowyerK.W. HallL.O. and KegelmeyerW.P. SMOTE: Synthetic Minority Over-Sampling Technique Journal of Artificial Intelligence Research (2002) 321\u2013357.","DOI":"10.1613\/jair.953"},{"key":"e_1_3_2_14_2","unstructured":"PachetF. and AucouturierJ.J. Improving timbre similarity: How high is the sky Journal of Negative Results in Speech and Audio Sciences (2004) 1\u201313."},{"key":"e_1_3_2_15_2","doi-asserted-by":"crossref","unstructured":"HicklinR.A. WatsonC.I. and UleryB. The Myth of the Goats: How Many People Have Fingerprints that are Hard to Match NIST (2005).","DOI":"10.6028\/NIST.IR.7271"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10586-018-1772-4"},{"key":"e_1_3_2_17_2","doi-asserted-by":"crossref","unstructured":"TomasevN. A probabilistic approach to nearest-neighbor classification: naive hubness bayesian kNN ACM International Conference on Information and Knowledge Management (2011) 2173\u20132176.","DOI":"10.1145\/2063576.2063919"},{"key":"e_1_3_2_18_2","doi-asserted-by":"crossref","unstructured":"BuzaK. and Pe\u0161kaL. Drug-target interaction prediction with Bipartite Local Models and hubness-aware regression Neuro-Computing (2017) 284\u2013293.","DOI":"10.1016\/j.neucom.2017.04.055"},{"key":"e_1_3_2_19_2","doi-asserted-by":"crossref","unstructured":"BuzaK. Classification of gene expression data: a hubness-aware semi-supervised approach Computer Methods and Pro-Grams in Biomedicine (2016) 105\u2013113.","DOI":"10.1016\/j.cmpb.2016.01.016"},{"key":"e_1_3_2_20_2","unstructured":"ZengD. XiaoY. WangJ. et al. Distant Supervised Relation Extraction with Cost-Sensitive Loss[J]."},{"key":"e_1_3_2_21_2","doi-asserted-by":"crossref","unstructured":"WeissG.M. Mining with rarity: a unifying framework SIGKDD Explorations (2004) 7\u201319.","DOI":"10.1145\/1007730.1007734"},{"key":"e_1_3_2_22_2","doi-asserted-by":"crossref","unstructured":"SunY. KamelM.S. and WongA.K.C. Cost-sensitive boosting for classification of imbalanced data Pattern Recognition (2007) 3358\u20133378.","DOI":"10.1016\/j.patcog.2007.04.009"},{"key":"e_1_3_2_23_2","doi-asserted-by":"crossref","unstructured":"ZhuT. LinY. and LiuY. Minority oversampling for imbalanced ordinal regression Knowledge-Based Systems (2019) 140\u2013155.","DOI":"10.1016\/j.knosys.2018.12.021"},{"key":"e_1_3_2_24_2","doi-asserted-by":"crossref","unstructured":"HongyuG. and HernaL.V. Learning from Imbalanced Data Sets with Boosting and Data Generation: The Data Boost-IM Approach SIGKDD Explorations (2004) 30\u201339.","DOI":"10.1145\/1007730.1007736"},{"key":"e_1_3_2_25_2","doi-asserted-by":"crossref","unstructured":"VoigtT. FriedR. and BackesM. Threshold optimization for classification in imbalanced data in a problemof gamma-ray astronomy Advances in Data Analysis and Classification (2014) 195\u2013216.","DOI":"10.1007\/s11634-014-0167-5"},{"key":"e_1_3_2_26_2","doi-asserted-by":"crossref","unstructured":"WangB. and PineauJ. Online bagging and boosting for imbalanced data streams IEEE Transactions on Knowledge & Data Engineering (2016) 3353\u20133366","DOI":"10.1109\/TKDE.2016.2609424"},{"key":"e_1_3_2_27_2","doi-asserted-by":"crossref","unstructured":"WuD. WangZ. and ChenY. Mixed-kernel based weighted extreme learningmachine for inertial sensor based human activity recognitionwith imbalanced dataset Neuro-Computing (2016) 35\u201349.","DOI":"10.1016\/j.neucom.2015.11.095"},{"key":"e_1_3_2_28_2","doi-asserted-by":"crossref","unstructured":"KrawczykB. Wo\u017aniakM. and SchaeferG. Cost-sensitive decision tree ensembles for effective imbalanced classification Applied Soft Computing (2014) 554\u2013562.","DOI":"10.1016\/j.asoc.2013.08.014"},{"key":"e_1_3_2_29_2","doi-asserted-by":"crossref","unstructured":"WangS. ChenH. and YaoX. Negative correlation learning for classification ensembles IJCNN (2010) 1\u20138.","DOI":"10.1109\/IJCNN.2010.5596702"},{"key":"e_1_3_2_30_2","doi-asserted-by":"crossref","unstructured":"HaixiangG. YijingL. and ShangJ. Learning from class-imbalanced data: Review of methods and applications Expert Systems with Applications (2017) 220\u2013239.","DOI":"10.1016\/j.eswa.2016.12.035"},{"key":"e_1_3_2_31_2","doi-asserted-by":"crossref","unstructured":"FuY. ZhangH. and BaiY. An Under-sampling Method: Based on Principal Component Analysis and Comprehensive Evaluation Model QRS Companion (2016) 414\u2013415.","DOI":"10.1109\/QRS-C.2016.68"},{"key":"e_1_3_2_32_2","doi-asserted-by":"crossref","unstructured":"HaJ. and LeeJ.S. A new under-sampling method using genetic algorithm for imbalanced data classification Proceedings of the 10th International Conference on Ubiqui-tous Information Management and Communication (2016).","DOI":"10.1145\/2857546.2857643"},{"key":"e_1_3_2_33_2","doi-asserted-by":"crossref","unstructured":"NgW.W.Y. HuJ. and YeungD.S. Diversified sensitivity-based under sampling for imbalance classification problems IEEE transactions on Cybernetics (2015) 2402\u20132412.","DOI":"10.1109\/TCYB.2014.2372060"},{"key":"e_1_3_2_34_2","doi-asserted-by":"crossref","unstructured":"ZhuT. LinY. and LiuY. Synthetic minority oversampling technique for multiclass imbalance problems Pattern Recognition (2017) 327\u2013340.","DOI":"10.1016\/j.patcog.2017.07.024"},{"key":"e_1_3_2_35_2","doi-asserted-by":"crossref","unstructured":"BaruaS. IslamM.M. and YaoX. MWMOTE-Majority Weighted Minority Oversampling Technique for Imbalanced Data Set Learning IEEE Transactions on Knowledge & Data Engineering (2013) 405\u2013425.","DOI":"10.1109\/TKDE.2012.232"},{"key":"e_1_3_2_36_2","doi-asserted-by":"crossref","unstructured":"YoungW.A. NyklS.L. and WeckmanG.R. Using Voronoi diagrams to improve classification performances when mode-ling imbalanced datasets Neural Computing and Applications (2015) 1041\u20131054.","DOI":"10.1007\/s00521-014-1780-0"},{"key":"e_1_3_2_37_2","doi-asserted-by":"crossref","unstructured":"S\u00e1ezJ.A. LuengoJ. and StefanowskiJ. SMOTE-IPF: Addressing the noisy and borderline examples problem in imbalanced classification by a re-sampling method with filtering Information Sciences (2015) 184\u2013203.","DOI":"10.1016\/j.ins.2014.08.051"},{"key":"e_1_3_2_38_2","doi-asserted-by":"crossref","unstructured":"ChawlaN.V. LazarevicA. HallL.O. and BowyerK. SMOTE Boost: Improving prediction of the minority class in boosting PKDD (2003) 107\u2013119.","DOI":"10.1007\/978-3-540-39804-2_12"},{"key":"e_1_3_2_39_2","doi-asserted-by":"crossref","unstructured":"SeiffertC. KhoshgoftaarT. Van HulseJ. and NapolitanoA. Rusboost:Ahybrid approach to alleviating class imbalance IEEE Trans Systems Man and Cybernetics (2010) 185\u2013197.","DOI":"10.1109\/TSMCA.2009.2029559"},{"key":"e_1_3_2_40_2","doi-asserted-by":"crossref","unstructured":"FreundY. and SchapireR. Experiments with a New Boosting Algorithm Proceedings of the 13th International Conference on Machine Learning (1996) 325\u2013332.","DOI":"10.1145\/238061.238163"},{"key":"e_1_3_2_41_2","doi-asserted-by":"crossref","unstructured":"TangB. and HeH. GIR-based ensemble sampling approaches for imbalanced learning Pattern Recognition (2017) 306\u2013319.","DOI":"10.1016\/j.patcog.2017.06.019"},{"key":"e_1_3_2_42_2","doi-asserted-by":"crossref","unstructured":"GalarM. Fern\u00e1ndezA. and BarrenecheaE. EUSBoost: Enhancing ensembles for highly imbalanced data-sets by evolutionary under sampling Pattern Recognition (2013) 3460\u20133471.","DOI":"10.1016\/j.patcog.2013.05.006"},{"key":"e_1_3_2_43_2","doi-asserted-by":"crossref","unstructured":"D\u00edez-PastorJ.F. Juan Garc\u00eda-OsorioJ. and C\u00e9sar Ran-dom Balance: Ensembles of variable priors classifiers for imbalanced data Knowledge-Based Systems (2015) 96\u2013111.","DOI":"10.1016\/j.knosys.2015.04.022"},{"key":"e_1_3_2_44_2","unstructured":"LiA.H. and BradicJ. Boosting in the presence of outliers: adaptive classification with nonconvex loss functions Journal of the American Statistical Association (2018) 1\u201315."},{"key":"e_1_3_2_45_2","unstructured":"AsuncionA. and NewmanD. UCI machine learning repository (2007). http:\/\/ergodicity.net\/tag\/ma chine-learning\/."},{"key":"e_1_3_2_46_2","unstructured":"https:\/\/www.openml.org\/."},{"key":"e_1_3_2_47_2","first-page":"865","article-title":"Nearest neighbors in high-dimensional data: The emergence and influence of hubs[C]\/\/","volume":"2009","author":"Radovanovi\u0107 M.","unstructured":"Radovanovi\u0107M., NanopoulosA. and Ivanovi\u0107M., Nearest neighbors in high-dimensional data: The emergence and influence of hubs[C]\/\/, Proceedings of the 26th Annual International Conference on Machine Learning, ACM, 2009, 865\u2013872.","journal-title":"Proceedings of the 26th Annual International Conference on Machine Learning"},{"key":"e_1_3_2_48_2","unstructured":"CorderG.W. and ForemanD.I. Nonparametric statistics: a step-by-step approach John Wiley & Sons (2014)."}],"container-title":["Journal of Intelligent &amp; Fuzzy Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/JIFS-190821","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/full-xml\/10.3233\/JIFS-190821","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/JIFS-190821","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T13:19:15Z","timestamp":1770124755000},"score":1,"resource":{"primary":{"URL":"https:\/\/journals.sagepub.com\/doi\/10.3233\/JIFS-190821"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,6,24]]},"references-count":47,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2020,7,17]]}},"alternative-id":["10.3233\/JIFS-190821"],"URL":"https:\/\/doi.org\/10.3233\/jifs-190821","relation":{},"ISSN":["1064-1246","1875-8967"],"issn-type":[{"value":"1064-1246","type":"print"},{"value":"1875-8967","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,6,24]]}}}