4.4 Article

SEMI-IMPLICIT BACK PROPAGATION

期刊

出版社

BIEMDAS ACAD PUBLISHERS INC
DOI: 10.23952/jnva.7.2023.4.08

关键词

Back propagation; Neural network; Proximal mapping; SGD

向作者/读者索取更多资源

Deep neural network (DNN) has gained significant attention in various applications, and the effectiveness of DNN heavily relies on network training algorithms. Although stochastic gradient descent (SGD) and other explicit gradient-based methods are commonly used, they face challenges such as gradient vanishing and explosion when training complex and deep neural networks. In this paper, a semi-implicit backpropagation method is proposed, which combines the ideas of error backpropagation (BP) and proximal point methods (PPM), to overcome these challenges and improve performance in terms of loss decreasing and accuracy.
Deep neural network (DNN) has been attracting a great attention in various applications. Network training algorithms play essential roles for the effectiveness of DNN. Although stochastic gra-dient descent (SGD) and other explicit gradient-based methods are the most popular algorithms, there are still many challenges such as gradient vanishing and explosion occurring in training a complex and deep neural networks. Motivated by the idea of error back propagation (BP) and proximal point methods (PPM), we propose a semi-implicit back propagation method for neural network training. Similar to the BP, the update on the neurons are propagated in a backward fashion and the parameters are optimized with proximal mapping. The implicit update for both hidden neurons and parameters allows to choose large step size in the training algorithm. Theoretically, we demonstrate the convergence of the proposed method under some standard assumptions. The experiments on illustrative examples, and two real data sets: MNIST and CIFAR-10, demonstrate that the proposed semi-implicit BP algorithm leads to bet -ter performance in terms of both loss decreasing and training/test accuracy, with a detail comparison to SGD/Adam and a similar algorithm proximal back propagation (ProxBP).

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.4
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据