3.8 Article

3D Sketching using Multi-View Deep Volumetric Prediction

Publisher

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3203197

Keywords

sketch-based modeling; deep learning; 3D reconstruction; line drawing

Funding

  1. ERC [ERC-2016-STG 714221]
  2. Intel/NSF VEC award [IIS-1539099]
  3. FBF grant [2018-0017]
  4. ANR project EnHerit [ANR-17-CE23-0008]
  5. Agence Nationale de la Recherche (ANR) [ANR-17-CE23-0008] Funding Source: Agence Nationale de la Recherche (ANR)

Ask authors/readers for more resources

Sketch-based modeling strives to bring the ease and immediacy of drawing to the 3D world. However, while drawings are easy for humans to create, they are very challenging for computers to interpret due to their sparsity and ambiguity. We propose a data-driven approach that tackles this challenge by learning to reconstruct 3D shapes from one or more drawings. At the core of our approach is a deep convolutional neural network (CNN) that predicts occupancy of a voxel grid from a line drawing. This CNN provides an initial 3D reconstruction as soon as the user completes a single drawing of the desired shape. We complement this single-view network with an updater CNN that refines an existing prediction given a new drawing of the shape created from a novel viewpoint. A key advantage of our approach is that we can apply the updater iteratively to fuse information from an arbitrary number of viewpoints, without requiring explicit stroke correspondences between the drawings. We train both CNNs by rendering synthetic contour drawings from hand-modeled shape collections as well as from procedurally-generated abstract shapes. Finally, we integrate our CNNs in an interactive modeling system that allows users to seamlessly draw an object, rotate it to see its 3D reconstruction, and refine it by re-drawing from another vantage point using the 3D reconstruction as guidance.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available