Keyword search (4,163 papers available)

"reinforcement learning" Keyword-tagged Publications:

Title Authors PubMed ID
1 Disentangling prediction error and value in a formal test of dopamine s role in reinforcement learning Usypchuk AA; Maes EJP; Lozzi M; Avramidis DK; Schoenbaum G; Esber GR; Gardner MPH; Iordanova MD; 40738112
CSBN
2 Comprehensive review of reinforcement learning for medical ultrasound imaging Elmekki H; Islam S; Alagha A; Sami H; Spilkin A; Zakeri E; Zanuttini AM; Bentahar J; Kadem L; Xie WF; Pibarot P; Mizouni R; Otrok H; Singh S; Mourad A; 40567264
ENCS
3 Machine learning innovations in CPR: a comprehensive survey on enhanced resuscitation techniques Islam S; Rjoub G; Elmekki H; Bentahar J; Pedrycz W; Cohen R; 40336660
ENCS
4 Computational neuroscience across the lifespan: Promises and pitfalls van den Bos W; Bruckner R; Nassar MR; Mata R; Eppinger B; 29066078
PSYCHOLOGY
5 Does phasic dopamine release cause policy updates? Carter F; Cossette MP; Trujillo-Pisanty I; Pallikaras V; Breton YA; Conover K; Caplan J; Solis P; Voisard J; Yaksich A; Shizgal P; 38039083
PSYCHOLOGY
6 Nonlinear dynamic modeling and model-based AI-driven control of a magnetoactive soft continuum robot in a fluidic environment Moezi SA; Sedaghati R; Rakheja S; 37932207
ENCS
7 Sub-hourly measurement datasets from 6 real buildings: Energy use and indoor climate Sartori I; Walnum HT; Skeie KS; Georges L; Knudsen MD; Bacher P; Candanedo J; Sigounis AM; Prakash AK; Pritoni M; Granderson J; Yang S; Wan MP; 37153123
ENCS
8 Reinforcement learning for automatic quadrilateral mesh generation: A soft actor-critic approach Pan J; Huang J; Cheng G; Zeng Y; 36375347
ENCS
9 Trust-Augmented Deep Reinforcement Learning for Federated Learning Client Selection Rjoub G; Wahab OA; Bentahar J; Cohen R; Bataineh AS; 35875592
ENCS
10 Designing a hybrid reinforcement learning based algorithm with application in prediction of the COVID-19 pandemic in Quebec. Khalilpourazari S, Hashemi Doulabi H 33424076
ENCS
11 Cue-Evoked Dopamine Neuron Activity Helps Maintain but Does Not Encode Expected Value. Mendoza JA, Lafferty CK, Yang AK, Britt JP 31693885
CSBN
12 Metacontrol of decision-making strategies in human aging. Bolenz F, Kool W, Reiter AM, Eppinger B 31397670
PERFORM
13 Developmental Changes in Learning: Computational Mechanisms and Social Influences. Bolenz F, Reiter AMF, Eppinger B 29250006
PERFORM

 

Title:Disentangling prediction error and value in a formal test of dopamine s role in reinforcement learning
Authors:Usypchuk AAMaes EJPLozzi MAvramidis DKSchoenbaum GEsber GRGardner MPHIordanova MD
Link:https://pubmed.ncbi.nlm.nih.gov/40738112/
DOI:10.1016/j.cub.2025.06.076
Publication:Current biology : CB
Keywords:Rescorla-Wagner modelchannelrhodopsinerror correctionmesolimbicoptogeneticsrodentscalar valuetemporal difference reinforcement learningtyrosine hydrohylase
PMID:40738112 Category: Date Added:2025-07-31
Dept Affiliation: CSBN
1 Department of Psychology, Centre for Studies in Behavioural Neurobiology, Concordia University, Montreal, QC H4B 1R6, Canada.
2 NIDA Intramural Research Program, Baltimore, MD 21224, USA; Departments of Anatomy & Neurobiology and Psychiatry, University of Maryland School of Medicine, Baltimore, MD 21201, USA; Solomon H. Snyder Department of Neuroscience, the Johns Hopkins University, Baltimore, MD 21287, USA.
3 Department of Psychology, Centre for Studies in Behavioural Neurobiology, Concordia University, Montreal, QC H4B 1R6, Canada. Electronic address: mihaela.iordanova@concordia.ca.

Description:

The discovery that midbrain dopamine (DA) transients can be mapped onto reward prediction errors (RPEs), the critical signal that drives learning, is a landmark in neuroscience. Causal support for the RPE hypothesis comes from studies showing that stimulating DA neurons can drive learning under conditions where it would not otherwise occur.1,2,3 However, such stimulation might also promote learning by adding reward value and indirectly inducing an RPE. This added value could support new learning even when it is insufficient to support instrumental behavior.4,5 Thus, these competing interpretations are challenging to disentangle and require direct comparison under matched conditions. We developed two computational models grounded in temporal difference reinforcement learning (TDRL)6,7,8 that dissociate the role of DA as an RPE versus a value signal. We validated our models by showing that they both predict learning (unblocking) when ventral tegmental area (VTA) DA stimulation occurs during expected reward delivery in a behavioral blocking design and confirmed this behaviorally. We then contrasted the models by delivering constant optogenetic stimulation during reward across both learning phases of blocking. The value model predicted blocking; the RPE model predicted unblocking. Behavioral results aligned with the latter. Moreover, the RPE model uniquely predicted that constant stimulation would unblock learning at higher frequencies (>20 Hz) when the artificial error alone drives learning. This, too, was confirmed experimentally. We demonstrate a principled computational and empirical dissociation between DA as an RPE versus a value signal. Our results advance understanding of how DA neuron stimulation drives learning.





BookR developed by Sriram Narayanan
for the Concordia University School of Health
Copyright © 2011-2026
Cookie settings
Concordia University