4.7 Article

Hierarchical multi-view context modelling for 3D object classification and retrieval

期刊

INFORMATION SCIENCES
卷 547, 期 -, 页码 984-995

出版社

ELSEVIER SCIENCE INC
DOI: 10.1016/j.ins.2020.09.057

关键词

Multi-view learning; 3D object retrieval; 3D object classification

资金

  1. National Natural Science Foundation of China [61772359, 61872267, 61902277]
  2. Tianjin New Generation Artificial Intelligence Major Program [19ZXZNGX00110, 18ZXZNGX00150]
  3. Open Project Program of the State Key Lab of CAD & CG, Zhejiang University [A2005, A2012]
  4. grant of Elite Scholar Program of Tianjin University [2019XRX-0035]
  5. Baidu Pinecone Program

向作者/读者索取更多资源

Recent advances in 3D sensors and software have led to a large amount of 3D data. This paper proposes a hierarchical multi-view context modelling method to address the problem of learning discriminative multi-view visual characteristics in 3D object classification and retrieval. The proposed method demonstrates superiority in experiments on ModelNet10, ModelNet40 and ShapeNetCore55 datasets.
Recent advances in 3D sensors and 3D modelling software have led to big 3D data. 3D object classification and retrieval are becoming important but challenging tasks. One critical problem for them is how to learn the discriminative multi-view visual characteristics. To address it, we proposes a hierarchical multi-view context modelling method (HMVCM). It consists of four key modules. First, the module of view-level context learning is designed to learn visual context features with respect to individual views and their neighbours. This module can imitate the human need to look back and forth to identify and compare the discriminative parts of individual 3D objects based on a joint convolutional neural network (CNN) and bidirectional long short-term memory (Bi-LSTM) network. Then, a multi-view grouping module is introduced to split views into several groups based on their visual appearance. A raw group-level representation can be obtained by the weighted sum of the view-level descriptors. Furthermore, we employ the Bi-LSTM to exploit the context among adjacent groups to generate group-wise context features. Finally, all group-wise context features are fused into a compact 3D object descriptor according to their significance. Extensive experiments on ModelNet10, ModelNet40 and ShapeNetCore55 demonstrate the superiority of the proposed method. (C) 2020 Elsevier Inc. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据