{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T08:58:56Z","timestamp":1764665936338,"version":"3.46.0"},"reference-count":37,"publisher":"MDPI AG","issue":"12","license":[{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Entropy"],"abstract":"<jats:p>As a foundational analytical tool, the discernibility matrix plays a pivotal role in the systematic reduction of knowledge in rough set-based systems. Recent advancements in rough set theory have witnessed the proliferation of discernibility matrix-based knowledge reduction algorithms, with notable applications in classical, neighborhood, covering, and fuzzy rough set models. However, the quadratic growth of the discernibility matrix\u2019s complexity (relative to domain size) imposes fundamental scalability limits, rendering it inefficient for real-world applications with massive datasets. To address this issue, we introduced a discernibility hashing strategy to limit the growth scale of the discernibility attributes and proposed a feature selection algorithm via discernibility hash based on rough set theory. First, on the premise of keeping the information of the original discernibility matrix unchanged, the method maps the discernibility attribute set of all objects to the storage unit through a hash function and records the number of collisions to construct a discernibility hash. By using this mapping, the two-dimensional matrix space can be reduced to a one-dimensional hash space, which greatly removes invalid and redundant elements. Secondly, based on the discernibility hash, an efficient knowledge reduction algorithm is proposed. The algorithm avoids invalid and redundant element attribute sets to participate in the knowledge reduction process and improves the efficiency of the algorithm. Finally, the experimental results show that the method is superior to the discernibility matrix method in terms of storage space and running time.<\/jats:p>","DOI":"10.3390\/e27121222","type":"journal-article","created":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T08:49:22Z","timestamp":1764665362000},"page":"1222","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Accelerated Feature Selection via Discernibility Hashing: A Rough Set Approach"],"prefix":"10.3390","volume":"27","author":[{"given":"Sheng","family":"Luo","sequence":"first","affiliation":[{"name":"School of Computer and Information, Shanghai Polytechnic University, Shanghai 201209, China"},{"name":"Artificial Intelligence Institute, Shanghai Polytechnic University, Shanghai 201209, China"}]},{"given":"Linxiang","family":"Shi","sequence":"additional","affiliation":[{"name":"School of Computer and Information, Shanghai Polytechnic University, Shanghai 201209, China"},{"name":"Artificial Intelligence Institute, Shanghai Polytechnic University, Shanghai 201209, China"}]},{"given":"Lin","family":"Chen","sequence":"additional","affiliation":[{"name":"School of Computer and Information, Shanghai Polytechnic University, Shanghai 201209, China"},{"name":"Artificial Intelligence Institute, Shanghai Polytechnic University, Shanghai 201209, China"}]},{"given":"Xiaolin","family":"Cao","sequence":"additional","affiliation":[{"name":"School of Computer and Information, Shanghai Polytechnic University, Shanghai 201209, China"},{"name":"Artificial Intelligence Institute, Shanghai Polytechnic University, Shanghai 201209, China"}]}],"member":"1968","published-online":{"date-parts":[[2025,12,1]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"44","DOI":"10.1016\/j.inffus.2020.01.005","article-title":"Overview and comparative study of dimensionality reduction techniques for high dimensional data","volume":"59","author":"Ayesha","year":"2020","journal-title":"Inf. Fusion"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"16351","DOI":"10.1109\/TNNLS.2025.3565320","article-title":"An Efficient and Robust Feature Selection Approach Based on Zentropy Measure and Neighborhood-Aware Model","volume":"36","author":"Yuan","year":"2025","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"7278","DOI":"10.1007\/s10489-024-05481-3","article-title":"Semi-supervised feature selection based on discernibility matrix and mutual information","volume":"54","author":"Qian","year":"2024","journal-title":"Appl. Intell."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"121099","DOI":"10.1016\/j.ins.2024.121099","article-title":"A novel adaptive neighborhood rough sets based on sparrow search algorithm and feature selection","volume":"679","author":"Liu","year":"2024","journal-title":"Inf. Sci."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"373","DOI":"10.1016\/j.ins.2020.02.015","article-title":"A neighborhood rough set model with nominal metric embedding","volume":"520","author":"Sheng","year":"2020","journal-title":"Inf. Sci."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1016\/j.ins.2006.06.003","article-title":"Rudiments of rough sets","volume":"177","author":"Pawlak","year":"2007","journal-title":"Inf. Sci."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"68","DOI":"10.1016\/j.ins.2020.11.021","article-title":"Attribute reduction with fuzzy rough self-information measures","volume":"549","author":"Wang","year":"2021","journal-title":"Inf. Sci."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"867","DOI":"10.1016\/j.ins.2008.11.020","article-title":"Discernibility matrix simplification for constructing attribute reducts","volume":"179","author":"Yao","year":"2009","journal-title":"Inf. Sci."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"611","DOI":"10.1007\/s10115-012-0589-3","article-title":"Discernibility matrix simplification with new attribute dependency functions for incomplete information systems","volume":"37","author":"Lang","year":"2013","journal-title":"Knowl. Inf. Syst."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"414","DOI":"10.1016\/j.ins.2023.01.091","article-title":"An ensemble classifier through rough set reducts for handling data with evidential attributes","volume":"635","author":"Trabelsi","year":"2023","journal-title":"Inf. Sci."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"1311","DOI":"10.1109\/TFUZZ.2024.3522466","article-title":"Learning Operator-Valued Kernels From Multilabel Datesets With Fuzzy Rough Sets","volume":"33","author":"Wang","year":"2025","journal-title":"IEEE Trans. Fuzzy Syst."},{"key":"ref_12","first-page":"2165","article-title":"Rough set theory for document clustering: A review","volume":"32","author":"Vidhya","year":"2017","journal-title":"J. Intell. Fuzzy Syst."},{"key":"ref_13","first-page":"509","article-title":"Cluster Analysis using Rough Set Theory","volume":"9","author":"Singh","year":"2017","journal-title":"J. Inform. Math. Sci."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"204","DOI":"10.1016\/j.knosys.2015.05.017","article-title":"Generalized attribute reduct in rough set theory","volume":"91","author":"Jia","year":"2016","journal-title":"Knowl.-Based Syst."},{"key":"ref_15","first-page":"100","article-title":"On reduct construction algorithms","volume":"4062","author":"Yao","year":"2008","journal-title":"Trans. Comput. Sci."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"3061","DOI":"10.1007\/s40815-023-01577-z","article-title":"Multi-Granularity Probabilistic Rough Fuzzy Sets for Interval-Valued Fuzzy Decision Systems","volume":"25","author":"Li","year":"2023","journal-title":"Int. J. Fuzzy Syst."},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"77","DOI":"10.1080\/03081070512331318329","article-title":"An uncertainty measure in partition-based fuzzy rough sets","volume":"34","author":"Mi","year":"2005","journal-title":"Int. J. Gen. Syst."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"75","DOI":"10.3233\/FI-2021-2014","article-title":"On Variable Precision Generalized Rough Sets and Incomplete Decision Tables","volume":"179","author":"Syau","year":"2021","journal-title":"Fundam. Informaticae"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"1131","DOI":"10.1109\/TKDE.2007.1044","article-title":"On Three Types of Covering-Based Rough Sets","volume":"19","author":"Zhu","year":"2007","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Ma, L., and Li, M. (2025). Covering rough set models, fuzzy rough set models and soft rough set models induced by covering similarity. Inf. Sci., 689.","DOI":"10.1016\/j.ins.2024.121520"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"3577","DOI":"10.1016\/j.ins.2008.05.024","article-title":"Neighborhood Rough Set Based Heterogeneous Feature Subset Selection","volume":"178","author":"Hu","year":"2008","journal-title":"Inf. Sci."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"233","DOI":"10.1016\/j.ins.2003.08.005","article-title":"Constructive and axiomatic approaches of fuzzy approximation operators","volume":"159","author":"Wu","year":"2004","journal-title":"Inf. Sci."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"1649","DOI":"10.1109\/TKDE.2010.260","article-title":"Kernelized Fuzzy Rough Sets and Their Applications","volume":"23","author":"Hu","year":"2011","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"289","DOI":"10.3233\/FUN-2005-68305","article-title":"A Comparative Study of Algebra Viewpoint and Information Viewpoint in Attribute Reduction","volume":"68","author":"Wang","year":"2005","journal-title":"Fundam. Informaticae"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"108949","DOI":"10.1016\/j.ijar.2023.108949","article-title":"Attribute reduction based on fusion information entropy","volume":"160","author":"Ji","year":"2023","journal-title":"Int. J. Approx. Reason."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"111","DOI":"10.1016\/j.ins.2021.08.067","article-title":"Granular-conditional-entropy-based attribute reduction for partially labeled data with proxy labels","volume":"580","author":"Gao","year":"2021","journal-title":"Inf. Sci."},{"key":"ref_27","first-page":"219","article-title":"Information entropy based attribute reduction for incomplete heterogeneous data","volume":"43","author":"Wang","year":"2022","journal-title":"J. Intell. Fuzzy Syst."},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"142","DOI":"10.1016\/j.knosys.2017.10.033","article-title":"Discernibility matrix based incremental attribute reduction for dynamic data","volume":"140","author":"Wei","year":"2018","journal-title":"Knowl.-Based Syst."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"20","DOI":"10.1016\/j.neucom.2018.01.094","article-title":"Compressed binary discernibility matrix based incremental attribute reduction algorithm for group dynamic data","volume":"344","author":"Ma","year":"2019","journal-title":"Neurocomputing"},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.ijar.2019.11.010","article-title":"Discernibility matrix based incremental feature selection on fused decision tables","volume":"118","author":"Liu","year":"2020","journal-title":"Int. J. Approx. Reason."},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Skowron, A., and Rauszer, C. (1992). The Discernibility Matrices and Functions in Information Systems, Kluwer.","DOI":"10.1007\/978-94-015-7975-9_21"},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"1092","DOI":"10.1016\/j.neucom.2007.09.003","article-title":"A novel condensing tree structure for rough set feature selection","volume":"71","author":"Yang","year":"2008","journal-title":"Neurocomputing"},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"474","DOI":"10.23919\/CJE.2014.10851196","article-title":"Constructing Rough Set Based Unbalanced Binary Tree for Feature Selection","volume":"23","author":"Lu","year":"2014","journal-title":"Chin. J. Electron."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"2233","DOI":"10.1007\/s00500-015-1638-0","article-title":"Minimal attribute reduction with rough set based on compactness discernibility information tree","volume":"20","author":"Jiang","year":"2016","journal-title":"Soft Comput."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"1457","DOI":"10.1109\/TKDE.2004.96","article-title":"Semantics-preserving dimensionality reduction: Rough and fuzzy-rough-based approaches","volume":"16","author":"Jensen","year":"2004","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.asoc.2008.05.006","article-title":"Dimensionality reduction based on rough set theory: A review","volume":"9","author":"Thangavel","year":"2009","journal-title":"Appl. Soft Comput."},{"key":"ref_37","first-page":"7","article-title":"Rough set theory and its applications","volume":"3","author":"Pawlak","year":"1998","journal-title":"J. Telecommun. Inf. Technol."}],"container-title":["Entropy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1099-4300\/27\/12\/1222\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T08:54:13Z","timestamp":1764665653000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1099-4300\/27\/12\/1222"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,1]]},"references-count":37,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["e27121222"],"URL":"https:\/\/doi.org\/10.3390\/e27121222","relation":{},"ISSN":["1099-4300"],"issn-type":[{"value":"1099-4300","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,1]]}}}