4.4 Article

A New Likelihood Ratio Method for Training Artificial Neural Networks

期刊

INFORMS JOURNAL ON COMPUTING
卷 34, 期 1, 页码 638-655

出版社

INFORMS
DOI: 10.1287/ijoc.2021.1088

关键词

stochastic gradient estimation; arti; neural network; image identi

资金

  1. National Natural Science Foundation of China [71901003, 72022001]
  2. National Science Foundation [MMI-1834710, IIS-1849280]
  3. CAS Pioneer Hundred Talents Program [2017-074]

向作者/读者索取更多资源

A new method for computing the gradients of artificial neural networks was proposed, which bypasses the continuity requirement of traditional methods by injecting artificial noises into signals. The method shows similar computational complexity and more transparent formulas compared to traditional methods. Additionally, a likelihood ratio-based method was developed to train more general ANNs and improve their robustness.
We investigate a new approach to compute the gradients of artificial neural networks (ANNs), based on the so-called push-out likelihood ratio method. Unlike the widely used backpropagation (BP) method that requires continuity of the loss function and the activation function, our approach bypasses this requirement by injecting artificial noises into the signals passed along the neurons. We show how this approach has a similar computational complexity as BP, and moreover is more advantageous in terms of removing the backward recursion and eliciting transparent formulas. We also formalize the connection between BP, a pivotal technique for training ANNs, and infinitesimal perturbation analysis, a classic path-wise derivative estimation approach, so that both our new proposed methods and BP can be better understood in the context of stochastic gradient estimation. Our approach allows efficient training for ANNs with more flexibility on the loss and activation functions, and shows empirical improvements on the robustness of ANNs under adversarial attacks and corruptions of natural noises. Summary of Contribution: Stochastic gradient estimation has been studied actively in simulation for decades and becomes more important in the era of machine learning and artificial intelligence. The stochastic gradient descent is a standard technique for training the artificial neural networks (ANNs), a pivotal problem in deep learning. The most popular stochastic gradient estimation technique is the backpropagation method. We find that the backpropagation method lies in the family of infinitesimal perturbation analysis, a path wise gradient estimation technique in simulation. Moreover, we develop a new likelihood ratio-based method, another popular family of gradient estimation technique in simulation, for training more general ANNs, and demonstrate that the new training method can improve the robustness of the ANN.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.4
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据