4.7 Article

Orthogonal channel attention-based multi-task learning for multi-view facial expression recognition

期刊

PATTERN RECOGNITION
卷 129, 期 -, 页码 -

出版社

ELSEVIER SCI LTD
DOI: 10.1016/j.patcog.2022.108753

关键词

Multi-view facial expression recognition; Orthogonal channel attention; Multi-task learning; Siamese convolutional neural network; Separated channel attention module

资金

  1. National Natural Science Foun-dation of China [61977027]
  2. Hubei Province Technological Innovation Major Project [2019AAA044]

向作者/读者索取更多资源

This paper proposes a novel orthogonal channel attention-based multi-task learning approach for multi-view facial expression recognition. By utilizing a Siamese CNN and a multi-task learning framework, as well as designing a separated channel attention module and an orthogonal channel attention loss, this approach achieves good recognition accuracy on two datasets.
Multi-view facial expression recognition (FER) is a challenging computer vision task due to the large intra-class difference caused by viewpoint variations. This paper presents a novel orthogonal channel attention-based multi-task learning (OCA-MTL) approach for FER. The proposed OCA-MTL approach adopts a Siamese convolutional neural network (CNN) to force the multi-view expression recognition model to learn the same features as the frontal expression recognition model. To further enhance the recognition accuracy of non-frontal expression, the multi-view expression model adopts a multi-task learning framework that regards head pose estimation (HPE) as an auxiliary task. A separated channel attention (SCA) module is embedded in the multi-task learning framework to generate individual attention for FER and HPE. Furthermore, orthogonal channel attention loss is presented to force the model to employ different feature channels to represent the facial expression and head pose, thereby decoupling them. The proposed approach is performed on two public facial expression datasets to evaluate its effectiveness and achieves an average recognition accuracy rate of 88.41 % under 13 viewpoints on Multi-PIE and 89.04% under 5 viewpoints on KDEF, outperforming state-of-the-art methods.(c) 2022 Elsevier Ltd. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据