4.6 Article

Survey on multimodal approaches to emotion recognition

期刊

NEUROCOMPUTING
卷 556, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.neucom.2023.126693

关键词

Emotion recognition; Affect sensing; Multimodal emotion recognition; Computer vision

向作者/读者索取更多资源

Emotion is a natural state of mind created by physiological changes in response to internal or external stimuli. Emotions influence decision-making, which in turn shapes behavior and character. Recognizing and managing emotions is an essential part of emotional intelligence, but it requires expertise in psychology and can be time-consuming. Computational recognition of emotional intelligence offers new possibilities for exploring this field.
Emotion is an instinctive state of mind created by the neurophysiological changes occurring in the human body as reactions to various internal or external stimuli. Emotions play a vital role in decision-making. The choices one makes in day-to-day life determine their behaviour and thus their character. Emotion and behaviour recognition are the key processes in ascertaining Emotional Intelligence (EQ) which is the inherent human potential to understand and manage one's own emotions in positive ways. But the process requires high expertise in the field of psychology and is exhaustive and time-consuming. This has opened a new horizon for exploring the computational recognition of EQ. Emotion Recognition (ER) is one of its sub-processes that identifies various human emotional states. Emotions are detected from physiological signals and also through non-invasive, visionbased algorithms by exploiting video and audio modalities. With the emergence of big data and state-of-art deep learning architectures combined with the vast availability of emotion-rich video content from various streaming platforms, Multimodal Emotion Recognition (MER) which detects emotions through multiple and complementary input modalities from video has gathered momentum in recent years. This survey paper elaborately discusses the unimodal ER through visual, auditory, and linguistic modalities and reviews MER with combined features from these modalities. It also discusses the joint representations and fusion mechanisms used to acquire the intermodal correlations. Finally, we put forward the limitations and gaps identified in the literature along with a few suggestions for future work.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据