4.6 Article

HighRes-MVSNet: A Fast Multi-View Stereo Network for Dense 3D Reconstruction From High-Resolution Images

Journal

IEEE ACCESS
Volume 9, Issue -, Pages 11306-11315

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2021.3050556

Keywords

Three-dimensional displays; Image reconstruction; Memory management; Feature extraction; Benchmark testing; Image resolution; Decoding; Convolutional neural network; dense 3D reconstruction; multi-view stereo

Funding

  1. European Institute of Innovation and Technology (EIT) [18004]
  2. RESilient transport InfraSTructure to extreme events (RESIST) Project through the European Union's Horizon 2020 Research and Innovation Program [769066]
  3. H2020 Societal Challenges Programme [769066] Funding Source: H2020 Societal Challenges Programme

Ask authors/readers for more resources

This study introduces an end-to-end deep learning architecture for 3D reconstruction, focusing on reducing memory requirements to utilize information from high-resolution images. By limiting the depth search range and utilizing a pyramid structure to gradually search for depth correspondences, the method can generate highly accurate 3D models using less GPU memory and runtime.
We propose an end-to-end deep learning architecture for 3D reconstruction from high-resolution images. While many approaches focus on improving reconstruction quality alone, we primarily focus on decreasing memory requirements in order to exploit the abundant information provided by modern high-resolution cameras. Towards this end, we present HighRes-MVSNet, a convolutional neural network with a pyramid encoder-decoder structure searching for depth correspondences incrementally over a coarse-to-fine hierarchy. The first stage of our network encodes the image features to a much smaller resolution in order to significantly reduce the memory requirements. Additionally, we limit the depth search range in every hierarchy level to the vicinity of the previous prediction. In this manner, we are able to produce highly accurate 3D models while only using a fraction of the GPU memory and runtime of previous methods. Although our method is aimed at much higher resolution images, we are still able to produce state-of-the-art results on the Tanks and Temples benchmark and achieve outstanding scores on the DTU benchmark.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available