4.8 Article

Towards understanding residual and dilated dense neural networks via convolutional sparse coding

期刊

NATIONAL SCIENCE REVIEW
卷 8, 期 3, 页码 -

出版社

OXFORD UNIV PRESS
DOI: 10.1093/nsr/nwaa159

关键词

convolutional neural network; convolutional sparse coding; residual neural network; mixed-scale dense neural network; dilated convolution; dense connection

资金

  1. National Key Research and Development Program of China [2019YFA0709501]
  2. National Natural Science Foundation of China [11661141019, 61621003]
  3. National Ten Thousand Talent Program for Young Top-notch Talents
  4. CAS Frontier Science Research Key Project for Top Young Scientists [QYZDB-SSWSYS008]

向作者/读者索取更多资源

This study presents mathematically equivalent forms of advanced deep learning models such as residual neural networks and dilated dense neural networks from the perspective of convolutional sparse coding. By considering factors such as initialization, dictionary design, and number of iterations, the researchers proposed novel multilayer models that have been proven effective through extensive numerical experiments and comparisons with competing methods. The study also provides a clear mathematical understanding of skip connections, dilated convolution, and dense connections in these models.
From the view of convolutional sparse coding, we build mathematically equivalent forms of two advanced deep learning models including residual and dilated dense neural networks with skip connections. Convolutional neural network (CNN) and its variants have led to many state-of-the-art results in various fields. However, a clear theoretical understanding of such networks is still lacking. Recently, a multilayer convolutional sparse coding (ML-CSC) model has been proposed and proved to equal such simply stacked networks (plain networks). Here, we consider the initialization, the dictionary design and the number of iterations to be factors in each layer that greatly affect the performance of the ML-CSC model. Inspired by these considerations, we propose two novel multilayer models: the residual convolutional sparse coding (Res-CSC) model and the mixed-scale dense convolutional sparse coding (MSD-CSC) model. They are closely related to the residual neural network (ResNet) and the mixed-scale (dilated) dense neural network (MSDNet), respectively. Mathematically, we derive the skip connection in the ResNet as a special case of a new forward propagation rule for the ML-CSC model. We also find a theoretical interpretation of dilated convolution and dense connection in the MSDNet by analyzing the MSD-CSC model, which gives a clear mathematical understanding of each. We implement the iterative soft thresholding algorithm and its fast version to solve the Res-CSC and MSD-CSC models. The unfolding operation can be employed for further improvement. Finally, extensive numerical experiments and comparison with competing methods demonstrate their effectiveness.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据