4.7 Article

QTTNet: Quantized tensor train neural networks for 3D object and video recognition

期刊

NEURAL NETWORKS
卷 141, 期 -, 页码 420-432

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.neunet.2021.05.034

关键词

3DCNN; Tensor train decomposition; Neural network compression; Quantization; 8 bit inference

资金

  1. National Key R&D Program of China [2018AAA0102604, 2018YFE0200200]
  2. National Science Foundation of China [61876215]
  3. Beijing Academy of Artificial Intelligence (BAAI) , China
  4. Institute for Guo Qiang, Tsinghua University
  5. CAAI-Huawei MindSpore Open Fund, China
  6. Beijing Science and Technology Program, China [Z191100007519009]
  7. Open project of Zhejiang laboratory, China
  8. Science and Technology Major Project of Guangzhou, China [202007030006]

向作者/读者索取更多资源

This article introduces a training framework for three-dimensional convolutional neural networks called QTTNet, which combines tensor train decomposition and data quantization to further shrink the model size and reduce memory and time costs. Experimental results demonstrate the effectiveness and competitiveness of this method in compressing 3D object and video recognition models.
Relying on the rapidly increasing capacity of computing clusters and hardware, convolutional neural networks (CNNs) have been successfully applied in various fields and achieved state-of-the-art results. Despite these exciting developments, the huge memory cost is still involved in training and inferring a large-scale CNN model and makes it hard to be widely used in resource-limited portable devices. To address this problem, we establish a training framework for three-dimensional convolutional neural networks (3DCNNs) named QTTNet that combines tensor train (TT) decomposition and data quantization together for further shrinking the model size and decreasing the memory and time cost. Through this framework, we can fully explore the superiority of TT in reducing the number of trainable parameters and the advantage of quantization in decreasing the bit-width of data, particularly compressing 3DCNN model greatly with little accuracy degradation. In addition, due to the low bit quantization to all parameters during the inference process including TT-cores, activations, and batch normalizations, the proposed method naturally takes advantage in memory and time cost. Experimental results of compressing 3DCNNs for 3D object and video recognition on ModelNet40, UCF11, and UCF50 datasets verify the effectiveness of the proposed method. The best compression ratio we have obtained is up to nearly 180x with competitive performance compared with other state-of-the-art researches. Moreover, the total bytes of our QTTNet models on ModelNet40 and UCF11 datasets can be 1000x lower than some typical practices such as MVCNN. (C) 2021 Published by Elsevier Ltd.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据