{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T10:22:12Z","timestamp":1774952532523,"version":"3.50.1"},"reference-count":38,"publisher":"MDPI AG","issue":"11","license":[{"start":{"date-parts":[[2021,10,29]],"date-time":"2021-10-29T00:00:00Z","timestamp":1635465600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Entropy"],"abstract":"<jats:p>There is no generally accepted definition for conditional Tsallis entropy. The standard definition of (unconditional) Tsallis entropy depends on a parameter \u03b1 that converges to the Shannon entropy as \u03b1 approaches 1. In this paper, we describe three proposed definitions of conditional Tsallis entropy suggested in the literature\u2014their properties are studied and their values, as a function of \u03b1, are compared. We also consider another natural proposal for conditional Tsallis entropy and compare it with the existing ones. Lastly, we present an online tool to compute the four conditional Tsallis entropies, given the probability distributions and the value of the parameter \u03b1.<\/jats:p>","DOI":"10.3390\/e23111427","type":"journal-article","created":{"date-parts":[[2021,11,1]],"date-time":"2021-11-01T22:21:08Z","timestamp":1635805268000},"page":"1427","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["On Conditional Tsallis Entropy"],"prefix":"10.3390","volume":"23","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1199-2220","authenticated-orcid":false,"given":"Andreia","family":"Teixeira","sequence":"first","affiliation":[{"name":"CINTESIS\u2014Centre for Health Technology and Services Research, Faculty of Medicine, University of Porto, 4200-450 Porto, Portugal"},{"name":"MEDCIDS\u2014Department of Community Medicine, Information and Decision in Health, Faculty of Medicine, University of Porto, 4200-450 Porto, Portugal"},{"name":"ADiT-LAB, Instituto Polit\u00e9cnico de Viana do Castelo, Rua Escola Industrial e Comercial Nun\u2019\u00c1lvares, 4900-347 Viana do Castelo, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8792-959X","authenticated-orcid":false,"given":"Andr\u00e9","family":"Souto","sequence":"additional","affiliation":[{"name":"LASIGE, Faculdade de Ci\u00eancias da Universidade de Lisboa, Campo Grande, 1749-016 Lisboa, Portugal"},{"name":"Departamento de Inform\u00e1tica, Faculdade de Ci\u00eancias da Universidade de Lisboa, Campo Grande, 1749-016 Lisboa, Portugal"},{"name":"Instituto de Telecomunica\u00e7\u00f5es, Av. Rovisco Pais, n 1, 1049-001 Lisboa, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9988-594X","authenticated-orcid":false,"given":"Lu\u00eds","family":"Antunes","sequence":"additional","affiliation":[{"name":"Computer Science Department, Faculty of Sciences, University of Porto, Rua do Campo Alegre, 4169-007 Porto, Portugal"}]}],"member":"1968","published-online":{"date-parts":[[2021,10,29]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"479","DOI":"10.1007\/BF01016429","article-title":"Possible generalization of Boltzmann-Gibbs statistics","volume":"52","author":"Tsallis","year":"1988","journal-title":"J. Stat. Phys."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"36","DOI":"10.1016\/S0019-9958(70)80040-7","article-title":"Generalized information functions","volume":"16","year":"1970","journal-title":"Inf. Control"},{"key":"ref_3","first-page":"30","article-title":"Quantification method of classification processes. concept of structural \u03b1-entropy","volume":"3","author":"Havrda","year":"1967","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"221","DOI":"10.1103\/RevModPhys.50.221","article-title":"General properties of entropy","volume":"50","author":"Wehrl","year":"1978","journal-title":"Rev. Mod. Phys."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Cover, T., and Thomas, J.A. (2006). Elements of Information Theory, Wiley. [2nd ed.].","DOI":"10.1002\/047174882X"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"379","DOI":"10.1002\/j.1538-7305.1948.tb01338.x","article-title":"A Mathematical Theory of Communication","volume":"27","author":"Shannon","year":"1948","journal-title":"Bell Syst. Tech. J."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"1765","DOI":"10.3390\/e13101765","article-title":"The Nonadditive Entropy Sq and Its Applications in Physics and Elsewhere: Some Remarks","volume":"13","author":"Tsallis","year":"2011","journal-title":"Entropy"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Gell-Mann, M., and Tsallis, C. (2004). Distributions of high-frequency stock-market observables. Nonextensive Entropy\u2014Interdisciplinary Applications, Oxford University Press.","DOI":"10.1093\/oso\/9780195159769.001.0001"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Ibrahim, R.W., and Darus, M. (2018). Analytic Study of Complex Fractional Tsallis\u2019 Entropy with Applications in CNNs. Entropy, 20.","DOI":"10.3390\/e20100722"},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"2426","DOI":"10.1016\/j.camwa.2010.08.038","article-title":"A novel automatic microcalcification detection technique using Tsallis entropy and a type II fuzzy index","volume":"60","author":"Mohanalin","year":"2010","journal-title":"Comput. Math. Appl."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"545","DOI":"10.1007\/s100510050217","article-title":"Sensitivity to initial conditions in the Bak-Sneppen model of biological evolution","volume":"1","author":"Tamarit","year":"1998","journal-title":"Eur. Phys. J. B"},{"key":"ref_12","unstructured":"(2018, November 08). Group of Statistical Physics. Available online: http:\/\/tsallis.cat.cbpf.br\/biblio.htm."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Ribeiro, M., Henriques, T., Castro, L., Souto, A., Antunes, L., Costa-Santos, C., and Teixeira, A. (2021). The Entropy Universe. Entropy, 23.","DOI":"10.3390\/e23020222"},{"key":"ref_14","first-page":"547","article-title":"On measures of information and entropy","volume":"1","year":"1961","journal-title":"Berkeley Symp. Math. Statist. Prob."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"023302","DOI":"10.1063\/1.2165744","article-title":"Information theoretical properties of Tsallis entropies","volume":"47","author":"Furuichi","year":"2006","journal-title":"J. Math. Phys."},{"key":"ref_16","first-page":"37","article-title":"Conditional Tsallis Entropy","volume":"13","author":"Manije","year":"2013","journal-title":"Cyb. Inf. Technol."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Heinrich, F., Ramzan, F., Rajavel, F.A., Schmitt, A.O., and G\u00fcltas, M. (2021). MIDESP: Mutual Information-Based Detection of Epistatic SNP Pairs for Qualitative and Quantitative Phenotypes. Biology, 10.","DOI":"10.3390\/biology10090921"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"e366","DOI":"10.7717\/peerj-cs.366","article-title":"Renyi entropy driven hierarchical graph clustering","volume":"7","author":"Oggier","year":"2021","journal-title":"PeerJ Comput. Sci."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"128","DOI":"10.1016\/j.ins.2021.02.059","article-title":"Information space of multi-sensor networks","volume":"565","author":"Tao","year":"2021","journal-title":"Inf. Sci."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"012301","DOI":"10.1103\/PhysRevA.62.012301","article-title":"Distinguishability of states and von Neumann entropy","volume":"62","author":"Jozsa","year":"2000","journal-title":"Phys. Rev. A"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Hassani, H., Unger, S., and Entezarian, M. (2021). Information content measurement of esg factors via entropy and its impact on society and security. Information, 12.","DOI":"10.3390\/info12100391"},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"656","DOI":"10.1002\/j.1538-7305.1949.tb00928.x","article-title":"Communication theory of secrecy systems","volume":"28","author":"Shannon","year":"1949","journal-title":"Bell Syst. Tech. J."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Bhotto, M.Z.A., and Antoniou, A. (2009, January 24\u201327). A new normalized minimum-error entropy algorithm with reduced computational complexity. Proceedings of the 2009 IEEE International Symposium on Circuits and Systems, Taipei, Taiwan.","DOI":"10.1109\/ISCAS.2009.5118324"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"595","DOI":"10.3390\/e13030595","article-title":"Entropy measures vs. Kolmogorov complexity","volume":"13","author":"Teixeira","year":"2011","journal-title":"Entropy"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Teixeira, A., Souto, A., Matos, A., and Antunes, L. (2010, January 13\u201318). Entropy measures vs. algorithmic information. Proceedings of the 2010 IEEE International Symposium on Information Theory, Austin, TX, USA.","DOI":"10.1109\/ISIT.2010.5513643"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Edgar, T., and Manz, D. (2017). Chapter 2-Science and Cyber Security. Research Methods for Cyber Security, Syngress.","DOI":"10.1016\/B978-0-12-805349-2.00002-9"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Huang, L., Shen, Y., Zhang, G., and Luo, H. (2015, January 14\u201316). Information system security risk assessment based on multidimensional cloud model and the entropy theory. Proceedings of the 2015 IEEE 5th International Conference on Electronics Information and Emergency Communication, Beijing, China.","DOI":"10.1109\/ICEIEC.2015.7284476"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"48","DOI":"10.26599\/TST.2019.9010047","article-title":"HTDet: A clustering method using information entropy for hardware Trojan detection","volume":"26","author":"Lu","year":"2021","journal-title":"Tsinghua Sci. Technol."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"2121","DOI":"10.1016\/j.bpj.2017.08.057","article-title":"Building Predictive Models of Genetic Circuits Using the Principle of Maximum Caliber","volume":"113","author":"Firman","year":"2017","journal-title":"Biophys J."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"363","DOI":"10.1111\/j.2006.0030-1299.14714.x","article-title":"Entropy and diversity","volume":"113","author":"Jost","year":"2006","journal-title":"Oikos"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Roach TNF (2020). Use and Abuse of Entropy in Biology: A Case for Caliber. Entropy, 22.","DOI":"10.3390\/e22121335"},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"688","DOI":"10.1038\/163688a0","article-title":"Measurement of diversity","volume":"163","author":"Simpson","year":"1949","journal-title":"Nature"},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"137","DOI":"10.1016\/j.physa.2015.09.067","article-title":"Weighted permutation entropy based on different symbolic approaches for financial time series","volume":"443","author":"Yin","year":"2016","journal-title":"Phys. A Stat. Mech. Its Appl."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Castiglioni, P., Parati, G., and Faini, A. (2019). Information-Domain Analysis of Cardiovascular Complexity: Night and Day Modulations of Entropy and the Effects of Hypertension. Entropy, 21.","DOI":"10.3390\/e21060550"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Polizzotto, N.R., Takahashi, T., Walker, C.P., and Cho, R.Y. (2016). Wide Range Multiscale Entropy Changes through Development. Entropy, 18.","DOI":"10.3390\/e18010012"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Prabhu, K.P., and Martis, R.J. (2020, January 2\u20134). Diagnosis of Schizophrenia using Kolmogorov Complexity and Sample Entropy. Proceedings of the 2020 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT), Bangalore, India.","DOI":"10.1109\/CONECCT50063.2020.9198472"},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"6801","DOI":"10.1109\/TIT.2014.2357799","article-title":"On the Conditional R\u00e9nyi Entropy","volume":"60","author":"Fehr","year":"2014","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"4273","DOI":"10.1109\/TIT.2012.2192713","article-title":"Conditional R\u00e9nyi Entropies","volume":"58","author":"Teixeira","year":"2012","journal-title":"IEEE Trans. Inf. Theory"}],"container-title":["Entropy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1099-4300\/23\/11\/1427\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T07:23:00Z","timestamp":1760167380000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1099-4300\/23\/11\/1427"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,29]]},"references-count":38,"journal-issue":{"issue":"11","published-online":{"date-parts":[[2021,11]]}},"alternative-id":["e23111427"],"URL":"https:\/\/doi.org\/10.3390\/e23111427","relation":{},"ISSN":["1099-4300"],"issn-type":[{"value":"1099-4300","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,10,29]]}}}