{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T06:30:54Z","timestamp":1772087454346,"version":"3.50.1"},"reference-count":36,"publisher":"MDPI AG","issue":"2","license":[{"start":{"date-parts":[[2026,2,21]],"date-time":"2026-02-21T00:00:00Z","timestamp":1771632000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Entropy"],"abstract":"<jats:p>Within the framework of evidence theory, maximum entropy is regarded as a measure of total uncertainty that satisfies a comprehensive set of mathematical properties and behavioral requirements. However, its practical applicability is severely questioned due to the high computational complexity of its calculation, which involves the manipulation of the power set of the frame of discernment. In the literature, attempts have been made to reduce this complexity by restricting the computation to singleton elements, leading to a formulation based on reachable probability intervals. Although this approach relies on a less specific representation of evidential information, it has been shown to provide an equivalent maximum entropy value under certain conditions. In this paper, we present an experimental comparative study of two algorithms for calculating maximum entropy in evidence theory: the classical algorithm, which operates directly on belief functions, and an alternative algorithm based on reachable probability intervals. Through numerical experiments, we demonstrate that the differences between these approaches are less pronounced than previously suggested in the literature. Depending on the type of information representations to which it is applied, the original algorithm based on belief functions can be more efficient than the one using the reachable probability interval approach. This is an interesting result, and a reason for choosing one algorithm over the other depending on the situation.<\/jats:p>","DOI":"10.3390\/e28020247","type":"journal-article","created":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T10:00:52Z","timestamp":1771840852000},"page":"247","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Comparison of Algorithms to Achieve the Maximum Entropy in the Theory of Evidence"],"prefix":"10.3390","volume":"28","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9018-5165","authenticated-orcid":false,"given":"Joaqu\u00edn","family":"Abell\u00e1n","sequence":"first","affiliation":[{"name":"Department of Computer Science and Artificial Intelligence, University of Granada, 18071 Granada, Spain"}]},{"given":"Aina","family":"L\u00f3pez-Gay","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Artificial Intelligence, University of Granada, 18071 Granada, Spain"}]},{"given":"Maria Isabel A.","family":"Ben\u00edtez","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Artificial Intelligence, University of Granada, 18071 Granada, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8257-8196","authenticated-orcid":false,"given":"Francisco Javier G.","family":"Castellano","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Artificial Intelligence, University of Granada, 18071 Granada, Spain"}]}],"member":"1968","published-online":{"date-parts":[[2026,2,21]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Walley, P. (1991). Statistical Reasoning with Imprecise Probabilities, Taylor & Francis. Chapman & Hall\/CRC Monographs on Statistics & Applied Probability.","DOI":"10.1007\/978-1-4899-3472-7"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"325","DOI":"10.1214\/aoms\/1177698950","article-title":"Upper and Lower Probabilities Induced by a Multivalued Mapping","volume":"38","author":"Dempster","year":"1967","journal-title":"Ann. Math. Stat."},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Shafer, G. (1976). A Mathematical Theory of Evidence, Princeton University Press.","DOI":"10.1515\/9780691214696"},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"37","DOI":"10.1016\/S0305-0483(99)00033-X","article-title":"The Dempster\u2013Shafer theory of evidence: An alternative approach to multicriteria decision modelling","volume":"28","author":"Beynon","year":"2000","journal-title":"Omega"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Yager, R.R., and Liu, L. (2008). A k-Nearest Neighbor Classification Rule Based on Dempster-Shafer Theory. Classic Works of the Dempster-Shafer Theory of Belief Functions, Springer.","DOI":"10.1007\/978-3-540-44792-4"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"569","DOI":"10.1109\/3468.618256","article-title":"A target identification comparison of Bayesian and Dempster-Shafer multisensor fusion","volume":"27","author":"Buede","year":"1997","journal-title":"IEEE Trans. Syst. Man Cybern. Part A Syst. Humans"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"292","DOI":"10.1109\/ICIP.1994.413578","article-title":"Human face recognition using Dempster-Shafer theory","volume":"Volume 2","author":"Ip","year":"1994","journal-title":"Proceedings of 1st International Conference on Image Processing"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"379","DOI":"10.1002\/j.1538-7305.1948.tb01338.x","article-title":"A Mathematical Theory of Communication","volume":"27","author":"Shannon","year":"1948","journal-title":"Bell Syst. Tech. J."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"249","DOI":"10.1080\/03081078308960825","article-title":"Entropy and specificity in a mathematical theory of evidence","volume":"9","author":"Yager","year":"1983","journal-title":"Int. J. Gen. Syst."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Klir, G., and Wierman, M. (1999). Uncertainty-Based Information: Elements of Generalized Information Theory, Physica-Verlag. Studies in Fuzziness and Soft Computing.","DOI":"10.1007\/978-3-7908-1869-7"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"733","DOI":"10.1080\/03081070802082486","article-title":"Requirements for total uncertainty measures in Dempster-Shafer theory of evidence","volume":"37","author":"Masegosa","year":"2008","journal-title":"Int. J. Gen. Syst."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"939","DOI":"10.1109\/PROC.1982.12425","article-title":"On the rationale of maximum-entropy methods","volume":"70","author":"Jaynes","year":"1982","journal-title":"Proc. IEEE"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Provencher Langlois, G., Buch, J., and Darbon, J. (2024). Efficient First-Order Algorithms for Large-Scale, Non-Smooth Maximum Entropy Models with Application to Wildfire Science. Entropy, 26.","DOI":"10.3390\/e26080691"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Stoyanov, J.M., Tagliani, A., and Novi Inverardi, P.L. (2024). Maximum Entropy Criterion for Moment Indeterminacy of Probability Densities. Entropy, 26.","DOI":"10.3390\/e26020121"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"423","DOI":"10.1016\/j.inffus.2012.03.003","article-title":"Ensembles of decision trees based on imprecise probabilities and uncertainty measures","volume":"14","year":"2013","journal-title":"Inf. Fusion"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Klir, G.J. (2005). Uncertainty and Information: Foundations of Generalized Information Theory, John Wiley And Sons, Inc.","DOI":"10.1002\/0471755575"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"5","DOI":"10.1023\/A:1016784627561","article-title":"On Measuring Uncertainty and Uncertainty-Based Information: Recent Developments","volume":"32","author":"Klir","year":"2001","journal-title":"Ann. Math. Artif. Intell."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"587","DOI":"10.1142\/S021848850300234X","article-title":"Maximum of Entropy for Credal Sets","volume":"11","author":"Abellan","year":"2003","journal-title":"Int. J. Uncertain. Fuzziness-Knowl.-Based"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"405","DOI":"10.1080\/03081079408935225","article-title":"Measuring total uncertainty in Dempster-Shafer Theory: A novel aaproach","volume":"22","author":"Harmanec","year":"1994","journal-title":"Int. J. Gen. Syst."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"3259","DOI":"10.1109\/TFUZZ.2020.3016723","article-title":"On the Maximum Entropy Negation of a Complex-Valued Distribution","volume":"29","author":"Xiao","year":"2021","journal-title":"IEEE Trans. Fuzzy Syst."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"377","DOI":"10.1142\/S0218488594000316","article-title":"Calculating maximum-entropy probability densities for belief functions","volume":"02","author":"Meyerowitz","year":"1994","journal-title":"Int. J. Uncertain. Fuzziness-Knowl.-Based Syst."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"549","DOI":"10.1016\/j.chaos.2016.07.014","article-title":"Deng entropy","volume":"91","author":"Deng","year":"2016","journal-title":"Chaos Solitons Fractals"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"195","DOI":"10.1016\/j.chaos.2016.12.024","article-title":"Analyzing properties of Deng entropy in the theory of evidence","volume":"95","year":"2017","journal-title":"Chaos Solitons Fractals"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"1186","DOI":"10.1109\/TSMC.2017.2770128","article-title":"Critique of Recent Uncertainty Measures Developed Under the Evidence Theory and Belief Intervals","volume":"50","year":"2020","journal-title":"IEEE Trans. Syst. Man Cybern. Syst."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"1215","DOI":"10.1002\/int.10143","article-title":"Building classification trees using the total uncertainty criterion","volume":"18","author":"Moral","year":"2003","journal-title":"Int. J. Intell. Syst."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"6827","DOI":"10.1016\/j.eswa.2012.01.013","article-title":"Bagging schemes on the presence of class noise in classification","volume":"39","author":"Masegosa","year":"2012","journal-title":"Expert Syst. Appl."},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"118017","DOI":"10.1109\/ACCESS.2020.3003715","article-title":"Maximum of Entropy for Belief Intervals Under Evidence Theory","volume":"8","year":"2020","journal-title":"IEEE Access"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"535","DOI":"10.1002\/j.1538-7305.1928.tb01236.x","article-title":"Transmission of Information1","volume":"7","author":"Hartley","year":"1928","journal-title":"Bell Syst. Tech. J."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"279","DOI":"10.1080\/03081078508934893","article-title":"A note on measures of specificity for fuzzy sets","volume":"10","author":"Dubois","year":"1985","journal-title":"Int. J. Gen. Syst."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"69","DOI":"10.1142\/S021848859300005X","article-title":"Maximum entropy algorithms for uncertainty measures","volume":"01","author":"Maeda","year":"1993","journal-title":"Int. J. Uncertain. Fuzziness-Knowl.-Based Syst."},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"205","DOI":"10.1109\/TSMCA.2009.2030962","article-title":"Notes on reducing algorithm complexity for computing an aggregate uncertainty measure","volume":"40","author":"Huynh","year":"2010","journal-title":"IEEE Trans. Syst. Man Cybern. Part A Syst. Hum."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"18284","DOI":"10.1109\/ACCESS.2019.2896286","article-title":"An Improved Deng Entropy and Its Application in Pattern Recognition","volume":"7","author":"Cui","year":"2019","journal-title":"IEEE Access"},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"120758","DOI":"10.1109\/ACCESS.2019.2937679","article-title":"The Maximum Deng Entropy","volume":"7","author":"Kang","year":"2019","journal-title":"IEEE Access"},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"47713","DOI":"10.1109\/ACCESS.2020.2979060","article-title":"Power Law and Dimension of the Maximum Value for Belief Distribution With the Maximum Deng Entropy","volume":"8","author":"Zhu","year":"2020","journal-title":"IEEE Access"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Pan, Q., Zhou, D., Tang, Y., Li, X., and Huang, J. (2019). A Novel Belief Entropy for Measuring Uncertainty in Dempster-Shafer Evidence Theory Framework Based on Plausibility Transformation and Weighted Hartley Entropy. Entropy, 21.","DOI":"10.3390\/e21020163"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Ji, D., Yang, X., Fei, L., and Zhai, C. (2019). An Improved Belief Entropy to Measure Uncertainty of Basic Probability Assignments Based on Deng Entropy and Belief Interval. Entropy, 21.","DOI":"10.3390\/e21111122"}],"container-title":["Entropy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1099-4300\/28\/2\/247\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T05:16:38Z","timestamp":1772082998000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1099-4300\/28\/2\/247"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,21]]},"references-count":36,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2026,2]]}},"alternative-id":["e28020247"],"URL":"https:\/\/doi.org\/10.3390\/e28020247","relation":{},"ISSN":["1099-4300"],"issn-type":[{"value":"1099-4300","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,21]]}}}