{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,31]],"date-time":"2025-12-31T05:01:59Z","timestamp":1767157319469,"version":"build-2238731810"},"reference-count":38,"publisher":"Proceedings of the National Academy of Sciences","issue":"9","content-domain":{"domain":["www.pnas.org"],"crossmark-restriction":true},"short-container-title":["Proc. Natl. Acad. Sci. U.S.A."],"published-print":{"date-parts":[[2009,3,3]]},"abstract":"<jats:p>The study of complex information processing systems requires appropriate theoretical tools to help unravel their underlying design principles. Information theory is one such tool, and has been utilized extensively in the study of the neural code. Although much progress has been made in information theoretic methodology, there is still no satisfying answer to the question: \u201cWhat is the information that a given property of the neural population activity (e.g., the responses of single cells within the population) carries about a set of stimuli?\u201d Here, we answer such questions via the minimum mutual information (MinMI) principle. We quantify the information in any statistical property of the neural response by considering all hypothetical neuronal populations that have the given property and finding the one that contains the minimum information about the stimuli. All systems with higher information values necessarily contain additional information processing mechanisms and, thus, the minimum captures the information related to the given property alone. MinMI may be used to measure information in properties of the neural response, such as that conveyed by responses of small subsets of cells (e.g., singles or pairs) in a large population and cooperative effects between subunits in networks. We show how the framework can be used to study neural coding in large populations and to reveal properties that are not discovered by other information theoretic methods.<\/jats:p>","DOI":"10.1073\/pnas.0806782106","type":"journal-article","created":{"date-parts":[[2009,2,14]],"date-time":"2009-02-14T07:46:39Z","timestamp":1234597599000},"page":"3490-3495","update-policy":"https:\/\/doi.org\/10.1073\/pnas.cm10313","source":"Crossref","is-referenced-by-count":30,"title":["The minimum information principle and its application to neural code analysis"],"prefix":"10.1073","volume":"106","author":[{"given":"Amir","family":"Globerson","sequence":"first","affiliation":[{"name":"School of Computer Science and Engineering, The Hebrew University, Jerusalem 91904, Israel;"}]},{"given":"Eran","family":"Stark","sequence":"additional","affiliation":[{"name":"Department of Physiology, Hadassah Medical School, The Hebrew University, Jerusalem 91120, Israel; and"}]},{"given":"Eilon","family":"Vaadia","sequence":"additional","affiliation":[{"name":"Department of Physiology, Hadassah Medical School, The Hebrew University, Jerusalem 91120, Israel; and"},{"name":"The Interdisciplinary Center for Neural Computation, The Hebrew University, Jerusalem 91904, Israel"}]},{"given":"Naftali","family":"Tishby","sequence":"additional","affiliation":[{"name":"School of Computer Science and Engineering, The Hebrew University, Jerusalem 91904, Israel;"},{"name":"The Interdisciplinary Center for Neural Computation, The Hebrew University, Jerusalem 91904, Israel"}]}],"member":"341","published-online":{"date-parts":[[2009,3,3]]},"reference":[{"key":"e_1_3_3_1_2","doi-asserted-by":"publisher","DOI":"10.1002\/j.1538-7305.1948.tb01338.x"},{"key":"e_1_3_3_2_2","doi-asserted-by":"publisher","DOI":"10.1037\/h0054663"},{"key":"e_1_3_3_3_2","doi-asserted-by":"publisher","DOI":"10.1037\/h0043158"},{"key":"e_1_3_3_4_2","first-page":"331","volume-title":"Current Problems in Animal Behaviour","author":"Barlow H","year":"1960","unstructured":"H Barlow, The coding of sensory messages. Current Problems in Animal Behaviour, eds W Thorpe, OL Zangwill (Cambridge Univ Press, Cambridge, UK), pp. 331\u2013360 (1960)."},{"key":"e_1_3_3_5_2","doi-asserted-by":"publisher","DOI":"10.1109\/2.36"},{"key":"e_1_3_3_6_2","volume-title":"Spikes","author":"Rieke F","year":"1997","unstructured":"F Rieke, D Warland, R de Ruyter van Steveninck, W Bialek Spikes (MIT Press, Cambridge, MA, 1997)."},{"key":"e_1_3_3_7_2","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.95.26.15706"},{"key":"e_1_3_3_8_2","doi-asserted-by":"publisher","DOI":"10.1126\/science.2063199"},{"key":"e_1_3_3_9_2","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.13-07-02758.1993"},{"key":"e_1_3_3_10_2","doi-asserted-by":"publisher","DOI":"10.1038\/2217"},{"key":"e_1_3_3_11_2","doi-asserted-by":"publisher","DOI":"10.1103\/PhysRev.106.620"},{"key":"e_1_3_3_12_2","doi-asserted-by":"publisher","DOI":"10.1162\/089976600300014872"},{"key":"e_1_3_3_13_2","doi-asserted-by":"publisher","DOI":"10.1038\/nature04701"},{"key":"e_1_3_3_14_2","doi-asserted-by":"publisher","DOI":"10.1002\/0471200611"},{"key":"e_1_3_3_15_2","first-page":"142","article-title":"Coding theorems for a discrete source with a fidelity criterion","volume":"4","author":"Shannon C","year":"1959","unstructured":"C Shannon, Coding theorems for a discrete source with a fidelity criterion. IRE Natl Conv Rec 4, 142\u2013163 (1959).","journal-title":"IRE Natl Conv Rec"},{"key":"e_1_3_3_16_2","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1131895100"},{"key":"e_1_3_3_17_2","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.23-37-11539.2003"},{"key":"e_1_3_3_18_2","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.96.6.3257"},{"key":"e_1_3_3_19_2","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.3418-08.2008"},{"key":"e_1_3_3_20_2","doi-asserted-by":"publisher","DOI":"10.1103\/PhysRevLett.91.238701"},{"key":"e_1_3_3_21_2","doi-asserted-by":"publisher","DOI":"10.1038\/373515a0"},{"key":"e_1_3_3_22_2","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.4697-04.2005"},{"key":"e_1_3_3_23_2","doi-asserted-by":"publisher","DOI":"10.1152\/jn.1987.57.1.162"},{"key":"e_1_3_3_24_2","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.5305-03.2004"},{"key":"e_1_3_3_25_2","doi-asserted-by":"publisher","DOI":"10.1152\/jn.1996.76.2.1310"},{"key":"e_1_3_3_26_2","doi-asserted-by":"publisher","DOI":"10.1088\/0954-898X\/14\/1\/303"},{"key":"e_1_3_3_27_2","volume-title":"Coding for a Discrete Information Source with a Distortion Measure","author":"Goblick T","year":"1962","unstructured":"T Goblick Coding for a Discrete Information Source with a Distortion Measure (Massachusetts Institute of Technology, Cambridge, MA, PhD thesis. (1962)."},{"key":"e_1_3_3_28_2","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1995.7.6.1129"},{"key":"e_1_3_3_29_2","volume-title":"Neural Networks: A Comprehensive Foundation","author":"Haykin S","year":"2007","unstructured":"S Haykin Neural Networks: A Comprehensive Foundation (Prentice\u2013Hall, Upper Saddle River, NJ, 2007)."},{"key":"e_1_3_3_30_2","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.1970.1054466"},{"key":"e_1_3_3_31_2","first-page":"193","volume-title":"Proceedings of the UAI","author":"Globerson A","year":"2004","unstructured":"A Globerson, N Tishby, The minimum information principle in discriminative learning. Proceedings of the UAI, eds M Chickering, J Halpern (Assoc for Uncertainty in Artificial Intelligence), pp. 193\u2013200 (2004)."},{"key":"e_1_3_3_32_2","doi-asserted-by":"publisher","DOI":"10.1038\/14731"},{"key":"e_1_3_3_33_2","first-page":"472","volume-title":"Proceedings of COLT","author":"Dud\u00edk M","year":"2004","unstructured":"M Dud\u00edk, S Phillips, RE Schapire, Performance guarantees for regularized maximum entropy density estimation. Proceedings of COLT, eds J Shawe-Taylor, Y Singer (Springer, New York), pp. 472\u2013486 (2004)."},{"key":"e_1_3_3_34_2","doi-asserted-by":"publisher","DOI":"10.1091\/mbc.11.12.4241"},{"key":"e_1_3_3_35_2","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511804441"},{"key":"e_1_3_3_36_2","doi-asserted-by":"publisher","DOI":"10.1214\/aop\/1176996454"},{"key":"e_1_3_3_37_2","first-page":"134","volume-title":"North American Chapter of the Association for Computational Linguistics\u2014Human Language Technologies (NAACL HLT)","author":"Sha F","year":"2003","unstructured":"F Sha, F Pereira, Shallow parsing with conditional random fields. North American Chapter of the Association for Computational Linguistics\u2014Human Language Technologies (NAACL HLT) (Assoc for Comput Linguistics, Boulder, CO), pp. 134\u2013141 (2003)."},{"key":"e_1_3_3_38_2","volume-title":"Pattern Classification","author":"Duda RO","year":"2000","unstructured":"RO Duda, PE Hart, DG Stork Pattern Classification (Wiley Interscience, New York, 2000)."}],"updated-by":[{"DOI":"10.1073\/pnas.0901850106","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2009,3,10]],"date-time":"2009-03-10T00:00:00Z","timestamp":1236643200000}}],"container-title":["Proceedings of the National Academy of Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/pnas.org\/doi\/pdf\/10.1073\/pnas.0806782106","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,4,12]],"date-time":"2022-04-12T18:18:28Z","timestamp":1649787508000},"score":1,"resource":{"primary":{"URL":"https:\/\/pnas.org\/doi\/full\/10.1073\/pnas.0806782106"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2009,3,3]]},"references-count":38,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2009,3,3]]}},"alternative-id":["10.1073\/pnas.0806782106"],"URL":"https:\/\/doi.org\/10.1073\/pnas.0806782106","relation":{},"ISSN":["0027-8424","1091-6490"],"issn-type":[{"value":"0027-8424","type":"print"},{"value":"1091-6490","type":"electronic"}],"subject":[],"published":{"date-parts":[[2009,3,3]]},"assertion":[{"value":"2008-08-05","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2009-03-03","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}