3.8 Proceedings Paper

ASMNet: a Lightweight Deep Neural Network for Face Alignment and Pose Estimation

出版社

IEEE COMPUTER SOC
DOI: 10.1109/CVPRW53098.2021.00168

关键词

-

向作者/读者索取更多资源

ASM assisted loss function helps improve the performance of face alignment and pose estimation tasks with a lightweight Convolutional Neural Network architecture. By gradually adjusting the regression problem and defining multi-tasks in the loss function, the network is guided to learn a smoother distribution of facial landmark points and achieve better performance in both tasks simultaneously. Comparisons with a larger model, MobileNetV2, show that ASMNet achieves comparable performance in face alignment and significantly better performance in face pose estimation with fewer parameters and operations.
Active Shape Model (ASM) is a statistical model of object shapes that represents a target structure. ASM can guide machine learning algorithms to fit a set of points representing an object (e.g., face) onto an image. This paper presents a lightweight Convolutional Neural Network (CNN) architecture with a loss function being assisted by ASM for face alignment and estimating head pose in the wild. We use ASM to first guide the network towards learning a smoother distribution of the facial landmark points. Inspired by transfer learning, during the training process, we gradually harden the regression problem and guide the network towards learning the original landmark points distribution. We define multi-tasks in our loss function that are responsible for detecting facial landmark points as well as estimating the face pose. Learning multiple correlated tasks simultaneously builds synergy and improves the performance of individual tasks. We compare the performance of our proposed model called ASMNet with MobileNetV2 (which is about 2 times bigger than ASMNet) in both the face alignment and pose estimation tasks. Experimental results on challenging datasets show that by using the proposed ASM assisted loss function, the ASMNet performance is comparable with MobileNetV2 in the face alignment task. In addition, for face pose estimation, ASMNet performs much better than MobileNetV2. ASMNet achieves an acceptable performance for facial landmark points detection and pose estimation while having a significantly smaller number of parameters and floating-point operations compared to many CNN-based models.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据