4.8 Article

K-Means Clustering-Based Kernel Canonical Correlation Analysis for Multimodal Emotion Recognition in Human-Robot Interaction

期刊

IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS
卷 70, 期 1, 页码 1016-1024

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIE.2022.3150097

关键词

Feature fusion; K-means clustering; Kernel canonical correlation analysis (KCCA); multimodal emotion recognition

向作者/读者索取更多资源

This article proposes a K-means clustering-based kernel canonical correlation analysis algorithm for multimodal emotion recognition in human-robot interaction. By fusing multimodal features from different modalities, the proposed method improves heterogeneity among modalities and enhances the accuracy of emotion recognition.
In this article, K-means clustering-based Kernel canonical correlation analysis algorithm is proposed for multimodal emotion recognition in human-robot interaction (HRI). The multimodal features (gray pixels; time and frequency domain) extracted from facial expression and speech are fused based on Kernel canonical correlation analysis. K-means clustering is used to select features from multiple modalities and reduce dimensionality. The proposed approach can improve the heterogenicity among different modalities and make multiple modalities complementary to promote multimodal emotion recognition. Experiments on two datasets, namely SAVEE and eNTER-FACE'05, are conducted to evaluate the accuracy of the proposed method. The results show that the proposed method produces good recognition rates that are higher than the ones produced by the methods without K-means clustering; more specifically, they are 2.77% higher in SAVEE and 4.7% higher in eNTERFACE'05.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据