{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,9]],"date-time":"2026-01-09T00:33:24Z","timestamp":1767918804030,"version":"3.49.0"},"reference-count":35,"publisher":"MIT Press","issue":"3","content-domain":{"domain":["direct.mit.edu"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2014,3,1]]},"abstract":"<jats:title>Abstract<\/jats:title>\n               <jats:p>Our ability to make decisions is predicated upon our knowledge of the outcomes of the actions available to us. Reinforcement learning theory posits that actions followed by a reward or punishment acquire value through the computation of prediction errors\u2014discrepancies between the predicted and the actual reward. A multitude of neuroimaging studies have demonstrated that rewards and punishments evoke neural responses that appear to reflect reinforcement learning prediction errors [e.g., Krigolson, O. E., Pierce, L. J., Holroyd, C. B., &amp; Tanaka, J. W. Learning to become an expert: Reinforcement learning and the acquisition of perceptual expertise. Journal of Cognitive Neuroscience, 21, 1833\u20131840, 2009; Bayer, H. M., &amp; Glimcher, P. W. Midbrain dopamine neurons encode a quantitative reward prediction error signal. Neuron, 47, 129\u2013141, 2005; O'Doherty, J. P. Reward representations and reward-related learning in the human brain: Insights from neuroimaging. Current Opinion in Neurobiology, 14, 769\u2013776, 2004; Holroyd, C. B., &amp; Coles, M. G. H. The neural basis of human error processing: Reinforcement learning, dopamine, and the error-related negativity. Psychological Review, 109, 679\u2013709, 2002]. Here, we used the brain ERP technique to demonstrate that not only do rewards elicit a neural response akin to a prediction error but also that this signal rapidly diminished and propagated to the time of choice presentation with learning. Specifically, in a simple, learnable gambling task, we show that novel rewards elicited a feedback error-related negativity that rapidly decreased in amplitude with learning. Furthermore, we demonstrate the existence of a reward positivity at choice presentation, a previously unreported ERP component that has a similar timing and topography as the feedback error-related negativity that increased in amplitude with learning. The pattern of results we observed mirrored the output of a computational model that we implemented to compute reward prediction errors and the changes in amplitude of these prediction errors at the time of choice presentation and reward delivery. Our results provide further support that the computations that underlie human learning and decision-making follow reinforcement learning principles.<\/jats:p>","DOI":"10.1162\/jocn_a_00509","type":"journal-article","created":{"date-parts":[[2013,10,29]],"date-time":"2013-10-29T17:40:09Z","timestamp":1383068409000},"page":"635-644","update-policy":"https:\/\/doi.org\/10.1162\/mitpressjournals.corrections.policy","source":"Crossref","is-referenced-by-count":54,"title":["How We Learn to Make Decisions: Rapid Propagation of Reinforcement Learning Prediction Errors in Humans"],"prefix":"10.1162","volume":"26","author":[{"given":"Olav E.","family":"Krigolson","sequence":"first","affiliation":[{"name":"1Dalhousie University"}]},{"given":"Cameron D.","family":"Hassall","sequence":"additional","affiliation":[{"name":"1Dalhousie University"}]},{"given":"Todd C.","family":"Handy","sequence":"additional","affiliation":[{"name":"2University of British Columbia"}]}],"member":"281","published-online":{"date-parts":[[2014,3,1]]},"reference":[{"key":"2021073000372846500_R1","doi-asserted-by":"crossref","first-page":"129","DOI":"10.1016\/j.neuron.2005.05.020","article-title":"Midbrain dopamine neurons encode a quantitative reward prediction error signal.","volume":"47","author":"Bayer","year":"2005","journal-title":"Neuron"},{"key":"2021073000372846500_R2","doi-asserted-by":"crossref","first-page":"1823","DOI":"10.1111\/j.1460-9568.2008.06138.x","article-title":"Learning-related changes in reward expectancy are reflected in the feedback-related negativity.","volume":"27","author":"Bellebaum","year":"2008","journal-title":"European Journal of Neuroscience"},{"key":"2021073000372846500_R3","volume-title":"Ars conjectandi.","author":"Bernoulli","year":"1713"},{"key":"2021073000372846500_R4","doi-asserted-by":"crossref","first-page":"936","DOI":"10.1162\/jocn.2010.21456","article-title":"Feedback-related negativity codes prediction error but not behavioral adjustment during probabilistic reversal learning.","volume":"23","author":"Chase","year":"2010","journal-title":"Journal of Cognitive Neuroscience"},{"key":"2021073000372846500_R5","doi-asserted-by":"crossref","first-page":"415","DOI":"10.1016\/j.neuroimage.2009.06.076","article-title":"Right frontal cortex generates reward-related theta-band oscillatory activity.","volume":"48","author":"Christie","year":"2009","journal-title":"Neuroimage"},{"key":"2021073000372846500_R6","doi-asserted-by":"crossref","first-page":"371","DOI":"10.1523\/JNEUROSCI.4421-06.2007","article-title":"Reinforcement learning signals predict future decisions.","volume":"27","author":"Cohen","year":"2007","journal-title":"Journal of Neuroscience"},{"key":"2021073000372846500_R7","doi-asserted-by":"crossref","first-page":"9","DOI":"10.1016\/j.jneumeth.2003.10.009","article-title":"EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis.","volume":"134","author":"Delorme","year":"2004","journal-title":"Journal of Neuroscience Methods"},{"key":"2021073000372846500_R9","doi-asserted-by":"crossref","first-page":"12087","DOI":"10.1523\/JNEUROSCI.1410-12.2012","article-title":"The processing of unexpected positive response outcomes in the mediofrontal cortex.","volume":"32","author":"Ferdinand","year":"2012","journal-title":"Journal of Neuroscience"},{"key":"2021073000372846500_R10","doi-asserted-by":"crossref","first-page":"2279","DOI":"10.1126\/science.1066893","article-title":"The medial frontal cortex and the rapid processing of monetary gains and losses.","volume":"295","author":"Gehring","year":"2002","journal-title":"Science"},{"key":"2021073000372846500_R11","doi-asserted-by":"crossref","first-page":"468","DOI":"10.1016\/0013-4694(83)90135-9","article-title":"A new method for off-line removal of ocular artifact.","volume":"55","author":"Gratton","year":"1983","journal-title":"Electroencephalography and Clinical Neurophysiology"},{"key":"2021073000372846500_R12","doi-asserted-by":"crossref","first-page":"148","DOI":"10.1016\/j.biopsycho.2005.04.001","article-title":"The feedback-related negativity reflects the binary evaluation of good versus bad outcomes.","volume":"71","author":"Hajcak","year":"2006","journal-title":"Biological Psychology"},{"key":"2021073000372846500_R13","doi-asserted-by":"crossref","first-page":"151","DOI":"10.1111\/j.1469-8986.2005.00270.x","article-title":"On the ERN and the significance of errors.","volume":"42","author":"Hajcak","year":"2005","journal-title":"Psychophysiology"},{"key":"2021073000372846500_R14","doi-asserted-by":"crossref","first-page":"550","DOI":"10.1111\/psyp.12040","article-title":"Frontal midline theta and N200 amplitude reflect complementary information about expectancy and outcome evaluation.","volume":"50","author":"Hajihosseini","year":"2013","journal-title":"Psychophysiology"},{"key":"2021073000372846500_R15","doi-asserted-by":"crossref","first-page":"679","DOI":"10.1037\/0033-295X.109.4.679","article-title":"The neural basis of human error processing: Reinforcement learning, dopamine, and the error-related negativity.","volume":"109","author":"Holroyd","year":"2002","journal-title":"Psychological Review"},{"key":"2021073000372846500_R16","doi-asserted-by":"crossref","first-page":"548","DOI":"10.1016\/j.cortex.2007.08.013","article-title":"Dorsal anterior cingulate cortex integrates reinforcement history to guide voluntary behavior.","volume":"44","author":"Holroyd","year":"2008","journal-title":"Cortex"},{"key":"2021073000372846500_R17","doi-asserted-by":"crossref","first-page":"913","DOI":"10.1111\/j.1469-8986.2007.00561.x","article-title":"Reward prediction error signals associated with a modified time estimation task.","volume":"44","author":"Holroyd","year":"2007","journal-title":"Psychophysiology"},{"key":"2021073000372846500_R18","doi-asserted-by":"crossref","first-page":"2481","DOI":"10.1097\/00001756-200312190-00037","article-title":"Errors in reward prediction are reflected in the event-related brain potential.","volume":"14","author":"Holroyd","year":"2003","journal-title":"NeuroReport"},{"key":"2021073000372846500_R19","doi-asserted-by":"crossref","first-page":"688","DOI":"10.1111\/j.1469-8986.2008.00668.x","article-title":"The feedback correct-related positivity: Sensitivity of the event-related brain potential to unexpected positive feedback.","volume":"45","author":"Holroyd","year":"2008","journal-title":"Psychophysiology"},{"key":"2021073000372846500_R20","doi-asserted-by":"crossref","first-page":"163","DOI":"10.1037\/0096-3445.134.2.163","article-title":"A mechanism for error detection in speeded response time tasks.","volume":"134","author":"Holroyd","year":"2005","journal-title":"Journal of Experimental Psychology: General"},{"key":"2021073000372846500_R21","doi-asserted-by":"crossref","first-page":"62","DOI":"10.1016\/j.brainres.2011.10.028","article-title":"Cognitive load impacts error evaluation within medial-frontal cortex.","volume":"1430","author":"Krigolson","year":"2012","journal-title":"Brain Research"},{"key":"2021073000372846500_R36","doi-asserted-by":"crossref","first-page":"401","DOI":"10.1007\/s00221-008-1482-x","article-title":"Electroencephalographic correlates of target and outcome errors.","volume":"190","author":"Krigolson","year":"2008","journal-title":"Experimental Brain Research"},{"key":"2021073000372846500_R22","doi-asserted-by":"crossref","first-page":"1833","DOI":"10.1162\/jocn.2009.21128","article-title":"Learning to become an expert: Reinforcement learning and the acquisition of perceptual expertise.","volume":"21","author":"Krigolson","year":"2009","journal-title":"Journal of Cognitive Neuroscience"},{"key":"2021073000372846500_R23","doi-asserted-by":"crossref","first-page":"476","DOI":"10.3758\/BF03210951","article-title":"Using confidence intervals in within-subject designs.","volume":"1","author":"Loftus","year":"1994","journal-title":"Psychonomic Bulletin and Review"},{"key":"2021073000372846500_R24","doi-asserted-by":"crossref","first-page":"203","DOI":"10.1037\/h0087426","article-title":"Using confidence intervals for graphically based data interpretation.","volume":"57","author":"Masson","year":"2003","journal-title":"Canadian Journal of Experimental Psychology"},{"key":"2021073000372846500_R25","volume-title":"Socialism.","author":"Mill","year":"1879"},{"key":"2021073000372846500_R26","doi-asserted-by":"crossref","first-page":"788","DOI":"10.1162\/jocn.1997.9.6.788","article-title":"Event-related brain potentials following incorrect feedback in a time-estimation task: Evidence for a \u201cgeneric\u201d neural system for error detection.","volume":"9","author":"Miltner","year":"1997","journal-title":"Journal of Cognitive Neuroscience"},{"key":"2021073000372846500_R28","doi-asserted-by":"crossref","first-page":"3768","DOI":"10.1016\/j.neuropsychologia.2011.09.037","article-title":"Problem gamblers exhibit reward hypersensitivity in medial-frontal cortex during gambling.","volume":"49","author":"Oberg","year":"2011","journal-title":"Neuropsychologia"},{"key":"2021073000372846500_R29","doi-asserted-by":"crossref","first-page":"769","DOI":"10.1016\/j.conb.2004.10.016","article-title":"Reward representations and reward-related learning in the human brain: Insights from neuroimaging.","volume":"14","author":"O'Doherty","year":"2004","journal-title":"Current Opinion in Neurobiology"},{"key":"2021073000372846500_R37","volume-title":"A Theory of Pavlovian Conditioning: Variations in the Effectiveness of Reinforcement and Nonreinforcement.","author":"Rescorla","year":"1972"},{"key":"2021073000372846500_R30","doi-asserted-by":"crossref","first-page":"1593","DOI":"10.1126\/science.275.5306.1593","article-title":"A neural substrate of prediction and reward.","volume":"275","author":"Schultz","year":"1997","journal-title":"Science"},{"key":"2021073000372846500_R31","volume-title":"Reinforcement learning: An introduction.","author":"Sutton","year":"1998"},{"key":"2021073000372846500_R32","doi-asserted-by":"crossref","first-page":"1642","DOI":"10.1126\/science.1105370","article-title":"Adaptive coding of reward value by dopamine neurons.","volume":"307","author":"Tobler","year":"2005","journal-title":"Science"},{"key":"2021073000372846500_R33","doi-asserted-by":"crossref","first-page":"43","DOI":"10.1038\/35083500","article-title":"Dopamine responses comply with basic assumptions of formal learning theory.","volume":"412","author":"Waelti","year":"2001","journal-title":"Nature"},{"key":"2021073000372846500_R34","doi-asserted-by":"crossref","first-page":"114","DOI":"10.1016\/j.brainres.2009.06.032","article-title":"The P300 and reward valence, magnitude, and expectancy in outcome evaluation.","volume":"1286","author":"Wu","year":"2009","journal-title":"Brain Research"},{"key":"2021073000372846500_R35","doi-asserted-by":"crossref","first-page":"6258","DOI":"10.1523\/JNEUROSCI.4537-03.2004","article-title":"Independent coding of reward magnitude and valence in the human brain.","volume":"24","author":"Yeung","year":"2004","journal-title":"Journal of Neuroscience"}],"container-title":["Journal of Cognitive Neuroscience"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/direct.mit.edu\/jocn\/article-pdf\/26\/3\/635\/1946756\/jocn_a_00509.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"http:\/\/direct.mit.edu\/jocn\/article-pdf\/26\/3\/635\/1946756\/jocn_a_00509.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,7,30]],"date-time":"2021-07-30T00:40:27Z","timestamp":1627605627000},"score":1,"resource":{"primary":{"URL":"https:\/\/direct.mit.edu\/jocn\/article\/26\/3\/635\/28080\/How-We-Learn-to-Make-Decisions-Rapid-Propagation"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2014,3,1]]},"references-count":35,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2014,3,1]]},"published-print":{"date-parts":[[2014,3,1]]}},"URL":"https:\/\/doi.org\/10.1162\/jocn_a_00509","relation":{},"ISSN":["0898-929X","1530-8898"],"issn-type":[{"value":"0898-929X","type":"print"},{"value":"1530-8898","type":"electronic"}],"subject":[],"published-other":{"date-parts":[[2014,3]]},"published":{"date-parts":[[2014,3,1]]}}}