4.5 Review

Tensor Decomposition for Model Reduction in Neural Networks: A Review

Journal

IEEE CIRCUITS AND SYSTEMS MAGAZINE
Volume 23, Issue 2, Pages 8-28

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/MCAS.2023.3267921

Keywords

Tensor decomposition; convolution neural network acceleration; recurrent neural network acceleration; transformer acceleration; canonical polyadic decomposition; Tucker decomposition; tensor train decomposition; tensor ring decomposition; block-term decomposition; hierarchical Tucker decomposition; model compression

Ask authors/readers for more resources

This article discusses the importance of neural networks in complex CV and NLP tasks, and explores the effectiveness and benefits of using low-rank tensor approximations to compress model parameters.
Modern neural networks have revolutionized the fields of computer vision (CV) and Natural Language Processing (NLP). They are widely used for solving complex CV tasks and NLP tasks such as image classification, image generation, and machine translation. Most state-of-the-art neural networks are over-parameterized and require a high computational cost. One straightforward solution is to replace the layers of the networks with their low-rank tensor approximations using different tensor decomposition methods. This article reviews six tensor decomposition methods and illustrates their ability to compress model parameters of convolutional neural networks (CNNs), recurrent neural networks (RNNs) and Transformers. The accuracy of some compressed models can be higher than the original versions. Evaluations indicate that tensor decompositions can achieve significant reductions in model size, run-time and energy consumption, and are well suited for implementing neural networks on edge devices.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available