3.8 Proceedings Paper

Semi-Supervised Monocular 3D Face Reconstruction With End-to-End Shape-Preserved Domain Transfer

Publisher

IEEE
DOI: 10.1109/ICCV.2019.00949

Keywords

-

Funding

  1. SenseTime Group Limited
  2. Research Grants Council of Hong Kong [CUHK 14202217, CUHK 14203118, CUHK 14205615, CUHK 14207814, CUHK 14213616, CUHK 14208417, CUHK 14239816]
  3. CUHK Direct Grant

Ask authors/readers for more resources

Monocular face reconstruction is a challenging task in computer vision, which aims to recover 3D face geometry from a single RGB face image. Recently, deep learning methods have achieved great improvements on monocular face reconstruction. However, for these methods to reach optimal performance, it is paramount to have large-scale training images with ground-truth 3D face geometry, which is generally difficult for human to annotate. To tackle this problem, we propose a semi-supervised monocular reconstruction method, which jointly optimizes a shape-preserved domain-transfer CycleGAN and a shape estimation network. The framework is semi-supervisely trained with 3D rendered images with ground-truth shapes and in-the-wild face images without any extra annotation. The CycleGAN network transforms all realistic images into rendered style and is end-to-end trained in the overall framework. This is the key difference compared with existing CycleGAN-based learning methods, which just used CycleGAN as a separate training sample generator. Novel landmark consistency loss and edge-aware shape estimation loss are proposed for our two networks to jointly solve the challenging face reconstruction problem. Experiments on public face reconstruction datasets demonstrate the effectiveness of our overall method as well as the individual components.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available