4.5 Article

Computational Modeling of Emotion-Motivated Decisions for Continuous Control of Mobile Robots

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCDS.2019.2963545

Keywords

Decision making; Brain modeling; Computational modeling; Robots; Hippocampus; Neuroscience; Brain-inspired computing; decision making; emotion– memory interactions; emotion-motivated learning; reinforcement learning

Funding

  1. National Key Research and Development Program of China [2017YFB1300200, 2017YFB1300203]
  2. National Natural Science Foundation of China [91648205, 61627808, 61702516]
  3. Chinese Academy of Science [XDB32050100]
  4. Development of Science and Technology of Guangdong Province Special Fund Project [2016B090910001]

Ask authors/readers for more resources

This article proposes an emotion-driven decision-making framework inspired by the impact of emotional reactions on subjective value in human decision-making. By constructing a brain-inspired computational model and a model-based decision-making approach, the continuous control problem is solved, leading to improved learning efficiency and exploration.
Immediate rewards are usually very sparse in the real world, which brings a great challenge to plain learning methods. Inspired by the fact that emotional reactions are incorporated into the computation of subjective value during decision-making in humans, an emotion-motivated decision-making framework is proposed in this article. Specifically, we first build a brain-inspired computational model of amygdala-hippocampus interaction to generate emotional reactions. The intrinsic emotion derives from the external reward and episodic memory and represents three psychological states: 1) valence; 2) novelty; and 3) motivational relevance. Then, a model-based (MB) decision-making approach with emotional intrinsic rewards is proposed to solve the continuous control problem of mobile robots. This method can execute online MB control with the constraint of the model-free policy and global value function, which is conducive to getting a better solution with a faster policy search. The simulation results demonstrate that the proposed approach has higher learning efficiency and maintains a higher level of exploration, especially, in some very sparse-reward environments.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available