4.6 Article

Least auxiliary loss-functions with impact growth adaptation (Laliga) for convolutional neural networks

期刊

NEUROCOMPUTING
卷 453, 期 -, 页码 413-427

出版社

ELSEVIER
DOI: 10.1016/j.neucom.2021.01.106

关键词

Deep neural network generalization; Features vanishing; Overfitting; Regularization; Model simplification; Auxiliary loss-functions

向作者/读者索取更多资源

The paper introduces a method of Least Auxiliary Loss-functions with Impact Growth Adaptation to address model selection and overfitting issues in Convolutional Neural Networks. Experimental results demonstrate its effectiveness across various datasets, particularly showing significant improvements in Visual Geometry Group networks.
Model selection is a challenge, and a popular Convolutional Neural Networks (CNN) usually takes extra need parameters. It causes overfitting in real applications. Besides, the extracted hidden features would be lost when the number of convolution layers increases. We use the least auxiliary loss-functions to solve both of these problems. To this end, an optimization problem is stated to select a set of layers with the highest contributions in the training process. Also, an impact growth adaptation procedure adjusts the weights of losses. The constructed Least Auxiliary Loss-functions with Impact Growth Adaptation (Laliga) is a professional forum to select the best settings of auxiliary loss functions for CNNs training. Laliga memorizes the hidden features carefully and better represents the space by using non redundant and more relevant features. Also, it uses singular value decomposition to regularize the weights. The theoretical results show that Laliga decreases overfitting substantially. Although this algorithm is useful for all CNN models, its results are auspicious for Visual Geometry Group (VGG) networks. The testing accuracies of Laliga for different VGG models on MNIST, CIFAR-10, and CIFAR-100 datasets are 99.7%, 92.3%, and 73.4%, indicating Laliga overcomes many regularization methods in the dropout family. Besides, on more complicated datasets Caltech-101 and Caltech-256, its accuracies raise than 66.1% and 33.2%, which are better than dropout and close to Adaptive Spectral Regularization (ASR) results, although Laliga converges rapidly than ASR. Finally, we analyze the results of Laliga in a transportation case study. (c) 2021 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据