期刊
PATTERN RECOGNITION
卷 129, 期 -, 页码 -出版社
ELSEVIER SCI LTD
DOI: 10.1016/j.patcog.2022.108753
关键词
Multi-view facial expression recognition; Orthogonal channel attention; Multi-task learning; Siamese convolutional neural network; Separated channel attention module
资金
- National Natural Science Foun-dation of China [61977027]
- Hubei Province Technological Innovation Major Project [2019AAA044]
This paper proposes a novel orthogonal channel attention-based multi-task learning approach for multi-view facial expression recognition. By utilizing a Siamese CNN and a multi-task learning framework, as well as designing a separated channel attention module and an orthogonal channel attention loss, this approach achieves good recognition accuracy on two datasets.
Multi-view facial expression recognition (FER) is a challenging computer vision task due to the large intra-class difference caused by viewpoint variations. This paper presents a novel orthogonal channel attention-based multi-task learning (OCA-MTL) approach for FER. The proposed OCA-MTL approach adopts a Siamese convolutional neural network (CNN) to force the multi-view expression recognition model to learn the same features as the frontal expression recognition model. To further enhance the recognition accuracy of non-frontal expression, the multi-view expression model adopts a multi-task learning framework that regards head pose estimation (HPE) as an auxiliary task. A separated channel attention (SCA) module is embedded in the multi-task learning framework to generate individual attention for FER and HPE. Furthermore, orthogonal channel attention loss is presented to force the model to employ different feature channels to represent the facial expression and head pose, thereby decoupling them. The proposed approach is performed on two public facial expression datasets to evaluate its effectiveness and achieves an average recognition accuracy rate of 88.41 % under 13 viewpoints on Multi-PIE and 89.04% under 5 viewpoints on KDEF, outperforming state-of-the-art methods.(c) 2022 Elsevier Ltd. All rights reserved.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据