4.6 Article

On the Redundancy in the Rank of Neural Network Parameters and Its Controllability

期刊

APPLIED SCIENCES-BASEL
卷 11, 期 2, 页码 -

出版社

MDPI
DOI: 10.3390/app11020725

关键词

matrix rank; neural network; pruning; redundancy; regularization

资金

  1. MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program [IITP-2020-2018-0-01405]
  2. MSIT (Ministry of Science and ICT), Korea, under the ICT Creative Consilience program [IITP-2020-0-01819]

向作者/读者索取更多资源

The paper examines the redundancy in the ranks of neural network parameters and introduces a regularization method to reduce it, resulting in significant improvements in training dynamics and fewer trainable parameters. Experimental results confirm the efficacy of the proposed approach, showcasing a substantial increase in accuracy and speedup in training steps, as well as a reduction in network size by 30.65%.
In this paper, we show that parameters of a neural network can have redundancy in their ranks, both theoretically and empirically. When viewed as a function from one space to another, neural networks can exhibit feature correlation and slower training due to this redundancy. Motivated by this, we propose a novel regularization method to reduce the redundancy in the rank of parameters. It is a combination of an objective function that makes the parameter rank-deficient and a dynamic low-rank factorization algorithm that gradually reduces the size of this parameter by fusing linearly dependent vectors together. This regularization-by-pruning approach leads to a neural network with better training dynamics and fewer trainable parameters. We also present experimental results that verify our claims. When applied to a neural network trained to classify images, this method provides statistically significant improvement in accuracy and 7.1 times speedup in terms of number of steps required for training. Furthermore, this approach has the side benefit of reducing the network size, which led to a model with 30.65% fewer trainable parameters.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据