4.6 Article

3D-RVP: A method for 3D object reconstruction from a single depth view using voxel and point

Journal

NEUROCOMPUTING
Volume 430, Issue -, Pages 94-103

Publisher

ELSEVIER
DOI: 10.1016/j.neucom.2020.10.097

Keywords

3D object reconstruction; Encoder-decoder network; Machine learning; Point prediction network

Funding

  1. National Natural Science Foundation of China [61773382, U1909218, U1909204, 61773381, U1811463, 61872365, 61806198]
  2. CAS Key Technology Talent Program
  3. Chinese Guangdong's ST Project [2019B1515120030]

Ask authors/readers for more resources

The paper proposes a method called 3D-RVP, which can reconstruct complete and accurate 3D geometry from a single depth view, addressing the issue of difficulty in improving the resolution of 3D object reconstruction technology. By combining a 3D encoder-decoder network and a point prediction network, the method can output high-resolution 3D models, outperforming existing technologies.
Three-dimensional object reconstruction technology has a wide range of applications such as augment reality, virtual reality, industrial manufacturing and intelligent robotics. Although deep learning-based 3D object reconstruction technology has developed rapidly in recent years, there remain important problems to be solved. One of them is that the resolution of reconstructed 3D models is hard to improve because of the limitation of memory and computational efficiency when deployed on resource-limited devices. In this paper, we propose 3D-RVP to reconstruct a complete and accurate 3D geometry from a single depth view, where R, V and P represent Reconstruction, Voxel and Point, respectively. It is a novel two-stage method that combines a 3D encoder-decoder network with a point prediction network. In the first stage, we propose a 3D encoder-decoder network with residual learning to output coarse prediction results. In the second stage, we propose an iterative subdivision algorithm to predict the labels of adaptively selected points. The proposed method can output high-resolution 3D models by increasing a small number of parameters. Experiments are conducted on widely used benchmarks of a ShapeNet dataset in which four categories of models are selected to test the performance of neural networks. Experimental results show that our proposed method outperforms the state-of-the-arts, and achieves about 2:7% improvement in terms of the intersection-over-union metric. (c) 2020 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available