4.8 Article

A Neural Signature of Hierarchical Reinforcement Learning

期刊

NEURON
卷 71, 期 2, 页码 370-379

出版社

CELL PRESS
DOI: 10.1016/j.neuron.2011.05.042

关键词

-

资金

  1. Fundacao para a Ciencia e Tecnologia [SFRH/BD/33273/2007]
  2. INRSA Training Grant in Quantitative Neuroscience [2 T32 MH065214]
  3. AFOSR [FA9550-08-1-041]
  4. Sloan Research Fellowship
  5. National Institute of Mental Health [P50 MH062196]
  6. James S. McDonnell Foundation
  7. Fundação para a Ciência e a Tecnologia [SFRH/BD/33273/2007] Funding Source: FCT

向作者/读者索取更多资源

Human behavior displays hierarchical structure: simple actions cohere into subtask sequences, which work together to accomplish overall task goals. Although the neural substrates of such hierarchy have been the target of increasing research, they remain poorly understood. We propose that the computations supporting hierarchical behavior may relate to those in hierarchical reinforcement learning (HRL), a machine-learning framework that extends reinforcement-learning mechanisms into hierarchical domains. To test this, we leveraged a distinctive prediction arising from HRL. In ordinary reinforcement learning, reward prediction errors are computed when there is an unanticipated change in the prospects for accomplishing overall task goals. HRL entails that prediction errors should also occur in relation to task subgoals. In three neuroimaging studies we observed neural responses consistent with such subgoal-related reward prediction errors, within structures previously implicated in reinforcement learning. The results reported support the relevance of HRL to the neural processes underlying hierarchical behavior.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据