4.6 Article

Intrinsic rewards explain context-sensitive valuation in reinforcement learning

期刊

PLOS BIOLOGY
卷 21, 期 7, 页码 -

出版社

PUBLIC LIBRARY SCIENCE
DOI: 10.1371/journal.pbio.3002201

关键词

-

向作者/读者索取更多资源

When observing the outcome of a choice, people are sensitive to the choice's context, such that the experienced value of an option depends on the alternatives. Traditionally, range adaptation has been proposed as the mechanism to explain this phenomenon in reinforcement learning tasks. However, the authors propose that internally defined goals also play a crucial role in shaping the subjective value attributed to any given option. Through multiple studies, they show that a new intrinsically enhanced reinforcement learning model can explain context-sensitive valuation better than range adaptation.
When observing the outcome of a choice, people are sensitive to the choice's context, such that the experienced value of an option depends on the alternatives: getting $1 when the possibilities were 0 or 1 feels much better than when the possibilities were 1 or 10. Context-sensitive valuation has been documented within reinforcement learning (RL) tasks, in which values are learned from experience through trial and error. Range adaptation, wherein options are rescaled according to the range of values yielded by available options, has been proposed to account for this phenomenon. However, we propose that other mechanisms-reflecting a different theoretical viewpoint-may also explain this phenomenon. Specifically, we theorize that internally defined goals play a crucial role in shaping the subjective value attributed to any given option. Motivated by this theory, we develop a new intrinsically enhanced RL model, which combines extrinsically provided rewards with internally generated signals of goal achievement as a teaching signal. Across 7 different studies (including previously published data sets as well as a novel, preregistered experiment with replication and control studies), we show that the intrinsically enhanced model can explain context-sensitive valuation as well as, or better than, range adaptation. Our findings indicate a more prominent role of intrinsic, goal-dependent rewards than previously recognized within formal models of human RL. By integrating internally generated signals of reward, standard RL theories should better account for human behavior, including context-sensitive valuation and beyond.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据