4.7 Article

Guided parallelized stochastic gradient descent for delay compensation

期刊

APPLIED SOFT COMPUTING
卷 102, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.asoc.2021.107084

关键词

Asynchronous/synchronous stochastic; gradient descent; Classification; Deep learning; Gradient Methods; Stochastic gradient descent

向作者/读者索取更多资源

The proposed guided SGD algorithm compensates for the deviation caused by delay and encourages consistent examples to steer the convergence of SGD, reducing the impact of delay on neural network models.
Stochastic gradient descent (SGD) algorithm and its variations have been effectively used to optimize neural network models. However, with the rapid growth of big data and deep learning, SGD is no longer the most suitable choice due to its natural behavior of sequential optimization of the error function. This has led to the development of parallel SGD algorithms, such as asynchronous SGD (ASGD) and synchronous SGD (SSGD) to train deep neural networks. However, it introduces a high variance due to the delay in parameter (weight) update. We address this delay in our proposed algorithm and try to minimize its impact. We employed guided SGD (gSGD) that encourages consistent examples to steer the convergence by compensating the unpredictable deviation caused by the delay. Its convergence rate is also similar to A/SSGD, however, some additional (parallel) processing is required to compensate for the delay. The experimental results demonstrate that our proposed approach has been able to mitigate the impact of delay for the quality of classification accuracy. The guided approach with SSGD clearly outperforms sequential SGD and even achieves an accuracy close to sequential SGD for some benchmark datasets. (C) 2021 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据