4.7 Article

Hierarchical multi-view context modelling for 3D object classification and retrieval

Journal

INFORMATION SCIENCES
Volume 547, Issue -, Pages 984-995

Publisher

ELSEVIER SCIENCE INC
DOI: 10.1016/j.ins.2020.09.057

Keywords

Multi-view learning; 3D object retrieval; 3D object classification

Funding

  1. National Natural Science Foundation of China [61772359, 61872267, 61902277]
  2. Tianjin New Generation Artificial Intelligence Major Program [19ZXZNGX00110, 18ZXZNGX00150]
  3. Open Project Program of the State Key Lab of CAD & CG, Zhejiang University [A2005, A2012]
  4. grant of Elite Scholar Program of Tianjin University [2019XRX-0035]
  5. Baidu Pinecone Program

Ask authors/readers for more resources

Recent advances in 3D sensors and software have led to a large amount of 3D data. This paper proposes a hierarchical multi-view context modelling method to address the problem of learning discriminative multi-view visual characteristics in 3D object classification and retrieval. The proposed method demonstrates superiority in experiments on ModelNet10, ModelNet40 and ShapeNetCore55 datasets.
Recent advances in 3D sensors and 3D modelling software have led to big 3D data. 3D object classification and retrieval are becoming important but challenging tasks. One critical problem for them is how to learn the discriminative multi-view visual characteristics. To address it, we proposes a hierarchical multi-view context modelling method (HMVCM). It consists of four key modules. First, the module of view-level context learning is designed to learn visual context features with respect to individual views and their neighbours. This module can imitate the human need to look back and forth to identify and compare the discriminative parts of individual 3D objects based on a joint convolutional neural network (CNN) and bidirectional long short-term memory (Bi-LSTM) network. Then, a multi-view grouping module is introduced to split views into several groups based on their visual appearance. A raw group-level representation can be obtained by the weighted sum of the view-level descriptors. Furthermore, we employ the Bi-LSTM to exploit the context among adjacent groups to generate group-wise context features. Finally, all group-wise context features are fused into a compact 3D object descriptor according to their significance. Extensive experiments on ModelNet10, ModelNet40 and ShapeNetCore55 demonstrate the superiority of the proposed method. (C) 2020 Elsevier Inc. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available