4.3 Article

Developing crossmodal expression recognition based on a deep neural model

期刊

ADAPTIVE BEHAVIOR
卷 24, 期 5, 页码 373-396

出版社

SAGE PUBLICATIONS LTD
DOI: 10.1177/1059712316664017

关键词

Crossmodal learning; emotion expression recognition; convolution neural network; self-organizing maps

资金

  1. CAPES Brazilian Federal Agency for the Support and Evaluation of Graduate Education [p.n.5951-13-5]
  2. German Research Foundation DFG under project CML [TRR 169]
  3. Hamburg Landesforschungsforderungsprojekt CROSS

向作者/读者索取更多资源

A robot capable of understanding emotion expressions can increase its own capability of solving problems by using emotion expressions as part of its own decision-making, in a similar way to humans. Evidence shows that the perception of human interaction starts with an innate perception mechanism, where the interaction between different entities is perceived and categorized into two very clear directions: positive or negative. While the person is developing during childhood, the perception evolves and is shaped based on the observation of human interaction, creating the capability to learn different categories of expressions. In the context of human-robot interaction, we propose a model that simulates the innate perception of audio-visual emotion expressions with deep neural networks, that learns new expressions by categorizing them into emotional clusters with a self-organizing layer. The proposed model is evaluated with three different corpora: The Surrey Audio-Visual Expressed Emotion (SAVEE) database, the visual Bi-modal Face and Body benchmark (FABO) database, and the multimodal corpus of the Emotion Recognition in the Wild (EmotiW) challenge. We use these corpora to evaluate the performance of the model to recognize emotional expressions, and compare it to state-of-the-art research.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.3
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据