4.7 Article

Rethinking dopamine as generalized prediction error

出版社

ROYAL SOC
DOI: 10.1098/rspb.2018.1645

关键词

reinforcement learning; successor representation; temporal difference learning

资金

  1. National Institutes of Health [CRCNS 1R01MH109177]
  2. Intramural Research Programat NIDA [ZIA-DA000587]
  3. NATIONAL INSTITUTE OF MENTAL HEALTH [R01MH109177] Funding Source: NIH RePORTER
  4. NATIONAL INSTITUTE ON DRUG ABUSE [ZIADA000587] Funding Source: NIH RePORTER

向作者/读者索取更多资源

Midbra in dopamine neurons are commonly thought to report a reward prediction error (RPE), as hypothesized by reinforcement learning (RL) theory. While this theory has been highly successful, several lines of evidence suggest that dopamine activity also encodes sensory prediction errors unrelated to reward. Here, we develop a new theory of dopamine function that embraces a broader conceptualization of prediction errors. By signalling errors in both sensory and reward predictions, dopamine supports a form of RL that lies between model-based and model-free algorithms. This account remains consistent with current canon regarding the correspondence between dopamine transients and RPEs, while also accounting for new data suggesting a role for those signals in phenomena such as sensory preconditioning and identity unblocking, which ostensibly draw upon knowledge beyond reward predictions.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据