{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T10:19:30Z","timestamp":1770977970897,"version":"3.50.1"},"reference-count":51,"publisher":"MDPI AG","issue":"5","license":[{"start":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T00:00:00Z","timestamp":1652659200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Entropy"],"abstract":"<jats:p>Data science, information theory, probability theory, statistical learning, statistical signal processing, and other related disciplines greatly benefit from non-negative measures of dissimilarity between pairs of probability measures [...]<\/jats:p>","DOI":"10.3390\/e24050712","type":"journal-article","created":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T13:06:23Z","timestamp":1652706383000},"page":"712","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Divergence Measures: Mathematical Foundations and Applications in Information-Theoretic and Statistical Problems"],"prefix":"10.3390","volume":"24","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5681-1273","authenticated-orcid":false,"given":"Igal","family":"Sason","sequence":"first","affiliation":[{"name":"Andrew & Erna Viterbi Faculty of Electrical and Computer Engineering, Technion\u2014Israel Institute of Technology, Haifa 3200003, Israel"},{"name":"Faculty of Mathematics, Technion\u2014Israel Institute of Technology, Haifa 3200003, Israel"}]}],"member":"1968","published-online":{"date-parts":[[2022,5,16]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"621","DOI":"10.1016\/j.sigpro.2012.09.003","article-title":"Divergence measures for statistical data processing\u2014An annotated bibliography","volume":"93","author":"Basseville","year":"2013","journal-title":"Signal Process."},{"key":"ref_2","unstructured":"Nielsen, F. (2019). Some universal insights on divergences for statistics, machine learning and artificial intelligence. Geometric Structures of Information, Springer."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"417","DOI":"10.1561\/0100000004","article-title":"Information Theory and Statistics: A Tutorial","volume":"1","author":"Shields","year":"2004","journal-title":"Found. Trends Commun. Inf. Theory"},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"4986","DOI":"10.1109\/TIT.2021.3085190","article-title":"Generalization error bounds via R\u00e9nyi-, f-divergences and maximal leakage","volume":"67","author":"Esposito","year":"2021","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_5","unstructured":"Liese, F., and Vajda, I. (1987). Convex Statistical Distances, Teubner-Texte Zur Mathematik."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"4394","DOI":"10.1109\/TIT.2006.881731","article-title":"On divergences and informations in statistics and information theory","volume":"52","author":"Liese","year":"2006","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_7","unstructured":"Pardo, L. (2006). Statistical Inference Based on Divergence Measures, Chapman and Hall\/CRC."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"169","DOI":"10.1080\/02331880902986919","article-title":"On divergences of finite measures and their applicability in statistics and information theory","volume":"44","author":"Stummer","year":"2010","journal-title":"Statistics"},{"key":"ref_9","unstructured":"Vajda, I. (1989). Theory of Statistical Inference and Information, Kluwer Academic Publishers."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Longo, G. (1975). A generalization of the rate-distortion theory and applications. Information Theory\u2014New Trends and Open Problems, Springer.","DOI":"10.1007\/978-3-7091-2730-8"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"79","DOI":"10.1214\/aoms\/1177729694","article-title":"On information and sufficiency","volume":"22","author":"Kullback","year":"1951","journal-title":"Ann. Math. Stat."},{"key":"ref_12","unstructured":"R\u00e9nyi, A. On measures of entropy and information. Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics, Available online: https:\/\/digitalassets.lib.berkeley.edu\/math\/ucb\/text\/math_s4_v1_article-27.pdf."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"3797","DOI":"10.1109\/TIT.2014.2320500","article-title":"R\u00e9nyi divergence and Kullback\u2013Leibler divergence","volume":"60","year":"2014","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"131","DOI":"10.1111\/j.2517-6161.1966.tb00626.x","article-title":"A general class of coefficients of divergence of one distribution from another","volume":"28","author":"Ali","year":"1966","journal-title":"J. R. Stat. Soc."},{"key":"ref_15","first-page":"85","article-title":"Eine Informationstheoretische Ungleichung und ihre Anwendung auf den Bewis der Ergodizit\u00e4t von Markhoffschen Ketten","volume":"8","year":"1963","journal-title":"Publ. Math. Inst. Hungar. Acad. Sci."},{"key":"ref_16","first-page":"299","article-title":"Information-type measures of difference of probability distributions and indirect observations","volume":"2","year":"1967","journal-title":"Stud. Sci. Math. Hung."},{"key":"ref_17","first-page":"329","article-title":"On topological properties of f-divergences","volume":"2","year":"1967","journal-title":"Stud. Sci. Math. Hung."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"328","DOI":"10.1143\/JPSJ.18.328","article-title":"Markov processes and the H-theorem","volume":"18","author":"Morimoto","year":"1963","journal-title":"J. Phys. Soc. Jpn."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"145","DOI":"10.1109\/18.61115","article-title":"Divergence measures based on the Shannon entropy","volume":"37","author":"Lin","year":"1991","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"307","DOI":"10.1016\/S0016-0032(96)00063-4","article-title":"The Jensen\u2013Shannon divergence","volume":"334","author":"Pardo","year":"1997","journal-title":"J. Frankl. Inst."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"1602","DOI":"10.1109\/18.850703","article-title":"Some inequalities for information divergence and related measures of discrimination","volume":"46","year":"2000","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Nielsen, F. (2020). On a generalization of the Jensen\u2013Shannon divergence and the Jensen\u2013Shannon centroid. Entropy, 22.","DOI":"10.3390\/e22020221"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Bleuler, C., Lapidoth, A., and Pfister, C. (2020). Conditional R\u00e9nyi divergences and horse betting. Entropy, 22.","DOI":"10.3390\/e22030316"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"157","DOI":"10.1080\/14786440009463897","article-title":"On the criterion that a given system of deviations from the probable in the case of a correlated system of variables is such that it can be reasonably supposed to have arisen from random sampling","volume":"50","author":"Pearson","year":"1900","journal-title":"Lond. Edinb. Dublin Philos. Mag. J. Sci."},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Nishiyama, T., and Sason, I. (2020). On relations between the relative entropy and \u03c72-divergence, generalizations and applications. Entropy, 22.","DOI":"10.3390\/e22050563"},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"1501","DOI":"10.1109\/18.104312","article-title":"Information theoretic inequalities","volume":"37","author":"Dembo","year":"1991","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Carlen, E., Madiman, M., and Werner, E.M. (2017). Forward and reverse entropy power inequalities in convex geometry. Convexity and Concentration, Springer. IMA Volumes in Mathematics and Its Applications.","DOI":"10.1007\/978-1-4939-7005-6"},{"key":"ref_28","first-page":"9","article-title":"Distance-divergence inequalities","volume":"64","author":"Marton","year":"2014","journal-title":"IEEE Inf. Theory Soc. Newsl."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Boucheron, S., Lugosi, G., and Massart, P. (2013). Concentration Inequalities\u2014A Nonasymptotic Theory of Independence, Oxford University Press.","DOI":"10.1093\/acprof:oso\/9780199535255.001.0001"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Raginsky, M., and Sason, I. (2018). Concentration of Measure Inequalities in Information Theory, Communications and Coding, Now Publishers. [3rd ed.].","DOI":"10.1561\/9781680835359"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Reeves, G. (2020). A two-moment inequality with applications to R\u00e9nyi entropy and mutual information. Entropy, 22.","DOI":"10.3390\/e22111244"},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"191","DOI":"10.1007\/BF02018661","article-title":"A class of measures of informativity of observation channels","volume":"2","year":"1972","journal-title":"Period. Mat. Hung."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"104","DOI":"10.1109\/TIT.2013.2288674","article-title":"Sharp inequalities for f-divergences","volume":"60","author":"Guntuboyina","year":"2014","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"3230","DOI":"10.1109\/TIT.2011.2137353","article-title":"On pairs of f-divergences and their joint range","volume":"57","author":"Vajda","year":"2011","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"4926","DOI":"10.1109\/TIT.2011.2159052","article-title":"Data processing theorems and the second law of thermodynamics","volume":"57","author":"Merhav","year":"2011","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"5973","DOI":"10.1109\/TIT.2016.2603151","article-title":"f-divergence inequalities","volume":"62","author":"Sason","year":"2016","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Sason, I. (2018). On f-divergences: Integral representations, local behavior, and inequalities. Entropy, 20.","DOI":"10.3390\/e20050383"},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"1879","DOI":"10.1109\/TIT.2017.2782359","article-title":"Strong data processing inequalities for input constrained additive noise channels","volume":"64","author":"Calmon","year":"2018","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"3355","DOI":"10.1109\/TIT.2016.2549542","article-title":"Strong data processing inequalities and \u03a6-Sobolev inequalities for discrete channels","volume":"62","author":"Raginsky","year":"2016","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_40","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1007\/978-1-4939-7005-6_7","article-title":"Strong data processing inequalities for channels and Bayesian networks","volume":"Volume 161","author":"Carlen","year":"2017","journal-title":"Convexity and Concentration"},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"103","DOI":"10.1134\/S0032946020020015","article-title":"Linear bounds between contraction coefficients for f-divergences","volume":"56","author":"Makur","year":"2020","journal-title":"Probl. Inf. Transm."},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Sason, I. (2019). On data-processing and majorization inequalities for f-divergences with applications. Entropy, 21.","DOI":"10.3390\/e21101022"},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Melbourne, J. (2020). Strongly convex divergences. Entropy, 22.","DOI":"10.3390\/e22111327"},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Broniatowski, M. (2021). Minimum divergence estimators, maximum likelihood and the generalized bootstrap. Entropy, 23.","DOI":"10.3390\/e23020185"},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"440","DOI":"10.1111\/j.2517-6161.1984.tb01318.x","article-title":"Multinomial Goodness-of-Fit Tests","volume":"46","author":"Cressie","year":"1984","journal-title":"J. R. Stat. Soc. Ser. (Methodol.)"},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"Verd\u00fa, S. (2021). Error Exponents and \u03b1-Mutual Information. Entropy, 23.","DOI":"10.3390\/e23020199"},{"key":"ref_47","doi-asserted-by":"crossref","first-page":"379","DOI":"10.1002\/j.1538-7305.1948.tb01338.x","article-title":"A Mathematical Theory of Communication","volume":"27","author":"Shannon","year":"1948","journal-title":"Bell Syst. Tech. J."},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"2307","DOI":"10.1109\/TIT.2010.2043769","article-title":"Channel coding rate in the finite blocklength regime","volume":"56","author":"Polyanskiy","year":"2010","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_49","unstructured":"Augustin, U. (1978). Noisy Channels. [Ph.D. Thesis, Universit\u00e4t Erlangen-N\u00fcrnberg]. Available online: http:\/\/libgen.rs\/book\/index.php?md5=4431004A794657A85AA356F111AA52FE."},{"key":"ref_50","doi-asserted-by":"crossref","first-page":"4","DOI":"10.1109\/TIT.2017.2757496","article-title":"Arimoto-R\u00e9nyi conditional entropy and Bayesian M-ary hypothesis testing","volume":"64","author":"Sason","year":"2018","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"Dwivedi, A., Wang, S., and Tajer, A. (2022). Discriminant analysis under f-divergence measures. Entropy, 24.","DOI":"10.3390\/e24020188"}],"container-title":["Entropy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1099-4300\/24\/5\/712\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T23:11:25Z","timestamp":1760137885000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1099-4300\/24\/5\/712"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,16]]},"references-count":51,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2022,5]]}},"alternative-id":["e24050712"],"URL":"https:\/\/doi.org\/10.3390\/e24050712","relation":{},"ISSN":["1099-4300"],"issn-type":[{"value":"1099-4300","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,5,16]]}}}