4.7 Article

Hybrid tensor decomposition in neural network compression

期刊

NEURAL NETWORKS
卷 132, 期 -, 页码 309-320

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.neunet.2020.09.006

关键词

Neural network compression; Hybrid tensor decomposition; Hierarchical Tucker; Tensor-train; Balanced structure

资金

  1. National Key R&D Program of China [2018YFE0200200]
  2. Beijing Academy of Artificial Intelligence (BAAI), China
  3. Institute for Guo Qiang of Tsinghua university, China
  4. Science and Technology Major Project of Guangzhou, China [202007030006]
  5. Ministry of Education, China
  6. open project of Zhejiang laboratory, China

向作者/读者索取更多资源

Deep neural networks (DNNs) have enabled impressive breakthroughs in various artificial intelligence (AI) applications recently due to its capability of learning high-level features from big data. However, the current demand of DNNs for computational resources especially the storage consumption is growing due to that the increasing sizes of models are being required for more and more complicated applications. To address this problem, several tensor decomposition methods including tensor-train (TT) and tensor-ring (TR) have been applied to compress DNNs and shown considerable compression effectiveness. In this work, we introduce the hierarchical Tucker (HT), a classical but rarely-used tensor decomposition method, to investigate its capability in neural network compression. We convert the weight matrices and convolutional kernels to both HT and TT formats for comparative study, since the latter is the most widely used decomposition method and the variant of HT. We further theoretically and experimentally discover that the HT format has better performance on compressing weight matrices, while the TT format is more suited for compressing convolutional kernels. Based on this phenomenon we propose a strategy of hybrid tensor decomposition by combining TT and HT together to compress convolutional and fully connected parts separately and attain better accuracy than only using the TT or HT format on convolutional neural networks (CNNs). Our work illuminates the prospects of hybrid tensor decomposition for neural network compression. (c) 2020 Elsevier Ltd. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据