4.7 Article

Efficient 3D Scene Semantic Segmentation via Active Learning on Rendered 2D Images

Journal

IEEE TRANSACTIONS ON IMAGE PROCESSING
Volume 32, Issue -, Pages 3521-3535

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2023.3286708

Keywords

3D semantic segmentation; active learning; rendered multi-view images

Ask authors/readers for more resources

Inspired by Active Learning and 2D-3D semantic fusion, our proposed framework utilizes rendered 2D images to achieve efficient semantic segmentation of large-scale 3D scenes with only a few 2D image annotations. By rendering perspective images in the 3D scene and fine-tuning a pre-trained network for segmentation, we can project and fuse dense predictions onto the 3D model. Through an iterative process of rendering-segmentation-fusion, difficult-to-segment image samples can be generated without complex 3D annotations, resulting in label-efficient 3D scene segmentation. Experimental results on three large-scale datasets demonstrate the effectiveness of our method compared to state-of-the-art approaches.
Inspired by Active Learning and 2D-3D semantic fusion, we proposed a novel framework for 3D scene semantic segmentation based on rendered 2D images, which could efficiently achieve semantic segmentation of any large-scale 3D scene with only a few 2D image annotations. In our framework, we first render perspective images at certain positions in the 3D scene. Then we continuously fine-tune a pre-trained network for image semantic segmentation and project all dense predictions to the 3D model for fusion. In each iteration, we evaluate the 3D semantic model and re-render images in several representative areas where the 3D segmentation is not stable and send them to the network for training after annotation. Through this iterative process of rendering-segmentation-fusion, it can effectively generate difficult-to-segment image samples in the scene, while avoiding complex 3D annotations, so as to achieve label-efficient 3D scene segmentation. Experiments on three large-scale indoor and outdoor 3D datasets demonstrate the effectiveness of the proposed method compared with other state-of-the-art.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available