4.8 Article

View-Aware Geometry-Structure Joint Learning for Single-View 3D Shape Reconstruction

Journal

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2021.3090917

Keywords

Three-dimensional displays; Shape; Image reconstruction; Geometry; Periodic structures; Solid modeling; Topology; Single-view 3D reconstruction; structure-aware reconstruction; multimodal learning; representation learning

Funding

  1. National Key R&D Program of China [2018YFB1703404]
  2. National Natural Science Funds of China [U1701262, U1801263]

Ask authors/readers for more resources

Reconstructing 3D shape from a single-view image using deep learning has gained popularity, but existing methods suffer from the lack of explicit structure modeling and loss of view information. In this paper, we propose VGSNet, an encoder-decoder architecture that jointly learns the feature representation of 2D image and 3D shape to achieve geometry and structure reconstruction from a single-view image.
Reconstructing a 3D shape from a single-view image using deep learning has become increasingly popular recently. Most existing methods only focus on reconstructing the 3D shape geometry based on image constraints. The lack of explicit modeling of structure relations among shape parts yields low-quality reconstruction results for structure-rich man-made shapes. In addition, conventional 2D-3D joint embedding architecture for image-based 3D shape reconstruction often omits the specific view information from the given image, which may lead to degraded geometry and structure reconstruction. We address these problems by introducing VGSNet, an encoder-decoder architecture for view-aware joint geometry and structure learning. The key idea is to jointly learn a multimodal feature representation of 2D image, 3D shape geometry and structure so that both geometry and structure details can be reconstructed from a single-view image. To this end, we explicitly represent 3D shape structures as part relations and employ image supervision to guide the geometry and structure reconstruction. Trained with pairs of view-aligned images and 3D shapes, the VGSNet implicitly encodes the view-aware shape information in the latent feature space. Qualitative and quantitative comparisons with the state-of-the-art baseline methods as well as ablation studies demonstrate the effectiveness of the VGSNet for structure-aware single-view 3D shape reconstruction.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available