4.5 Review

Reconciling reinforcement learning models with behavioral extinction and renewal: Implications for addiction, relapse, and problem gambling

Journal

PSYCHOLOGICAL REVIEW
Volume 114, Issue 3, Pages 784-805

Publisher

AMER PSYCHOLOGICAL ASSOC
DOI: 10.1037/0033-295X.114.3.784

Keywords

temporal difference reinforcement learning (TDRL); dopamine; reinstantiation; problem gambling

Funding

  1. NIDA NIH HHS [P50 DA01333] Funding Source: Medline

Ask authors/readers for more resources

Because learned associations are quickly renewed following extinction, the extinction process must include processes other than unlearning. However, reinforcement learning models, such as the temporal difference reinforcement learning (TDRL) model, treat extinction as an unlearning of associated value and are thus unable to capture renewal. TDRL models are based on the hypothesis that dopamine carries a reward prediction error signal; these models predict reward by driving that reward error to zero. The authors construct a TDRL model that can accommodate extinction and renewal through two simple processes: (a) a TDRL process that learns the value of situation-action pairs and (b) a situation recognition process that categorizes the observed cues into situations. This model has implications for dysfunctional states, including relapse after addiction and problem gambling.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available