4.7 Article

On the Decentralized Stochastic Gradient Descent With Markov Chain Sampling

期刊

IEEE TRANSACTIONS ON SIGNAL PROCESSING
卷 71, 期 -, 页码 2895-2909

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TSP.2023.3297053

关键词

Markov chain sampling; gradient descent; decentralization; distributed machine learning; convergence

向作者/读者索取更多资源

This paper studies the decentralized Markov chain gradient descent (DMGD), a variant of the decentralized stochastic gradient method. It analyzes the convergence rates of DMGD on a connected graph, highlighting the critical dependencies on the graph topology and the mixing time of the Markov chain. The numerical experiments also verify the sample efficiency of DMGD.
The decentralized stochastic gradient method emerges as a promising solution for solving large-scale machine learning problems. This paper studies the decentralized Markov chain gradient descent (DMGD), a variant of the decentralized stochastic gradient method, which draws random samples along the trajectory of a Markov chain. DMGD arises when obtaining independent samples is costly or impossible, excluding the use of the traditional stochastic gradient algorithms. Specifically, we consider the DMGD over a connected graph, where each node only communicates with its neighbors by sending and receiving the intermediate results. We establish both ergodic and nonergodic convergence rates of DMGD, which elucidate the critical dependencies on the topology of the graph that connects all nodes and the mixing time of the Markov chain. We further numerically verify the sample efficiency of DMGD.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据