4.6 Article

Multi-view facial action unit detection via DenseNets and CapsNets

期刊

MULTIMEDIA TOOLS AND APPLICATIONS
卷 81, 期 14, 页码 19377-19394

出版社

SPRINGER
DOI: 10.1007/s11042-021-11147-w

关键词

Facial action unit; Facial expression recognition; Emotion recognition; CapsNets; DenseNets; Deep learning; Convolutional neural networks

资金

  1. Beijing Nova Program [Z201100006820123]
  2. Beijing Municipal Science and Technology Commission
  3. Natural Science Foundation of China [U1536203, 61972169]
  4. National key research and development program of China [2016QY01W0200]
  5. Major Scientific and Technological Project of Hubei Province [2018AAA068, 2019AAA051]

向作者/读者索取更多资源

Two novel neural network architectures, AUCaps and AUCaps++, are proposed for multi-view and multi-label facial action unit (AU) detection by optimizing the combination of CapsNets and dense blocks. The proposed method outperforms competitors in terms of F1 scores on both within-dataset and cross-dataset evaluations.
Though the standard convolutional neural networks (CNNs) have been proposed to increase the robustness of facial action unit (AU) detection regarding pose variations, it is hard to enhance detection performance because the standard CNNs are not robust enough to affine transformation. To address this issue, two novel architectures termed as AUCaps and AUCaps++ are proposed for multi-view and multi-label facial AU detection in this work. In these two architectures, one or more dense blocks and one capsule networks (CapsNets) are stacked. Specifically, The dense blocks prefixed before CapsNets are used to learn more discriminative high-level AU features, and the CapsNets is exploited to learn more view-invariant AU features. Moreover, the capsule types and digit capsule dimension are optimized to avoid the computation and storage burden caused by the dynamic routing in standard CapsNets. Because the AUCaps and AUCaps++ are trained by jointly optimizing multi-label loss of AU and reconstruction loss of viewpoint image, the proposed method could achieve high F1 score and learn human face roughly in the reconstruction images over different AUs. Numerical results of within-dataset and cross-dataset show that the average F1 scores of the proposed method outperform the competitors using hand-crafted features or deep learning features by a big margin on two public datasets.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据