4.7 Article

Buffered Asynchronous SGD for Byzantine Learning

Journal

JOURNAL OF MACHINE LEARNING RESEARCH
Volume 24, Issue -, Pages -

Publisher

MICROTOME PUBL

Keywords

distributed machine learning; momentum; asynchronous Byzantine learning; buffer; stochastic gradient descent

Ask authors/readers for more resources

Distributed learning is widely used in various fields such as cluster-based large-scale learning, federated learning, and edge computing. Byzantine learning, which deals with failure or attack in distributed learning, has gained attention recently. This paper proposes a novel method called buffered asynchronous stochastic gradient descent (BASGD) for asynchronous Byzantine learning (ABL). BASGD is the first ABL method that can resist non-omniscient attacks without storing any instances on the server. An improved variant of BASGD, called BASGD with momentum (BASGDm), is also introduced. Both BASGD and BASGDm have a wider scope of application and are proven to be convergent and able to resist failure or attack. Empirical results demonstrate the superior performance of our methods compared to existing ABL baselines in the presence of failure or attack.
Distributed learning has become a hot research topic due to its wide application in clusterbased large-scale learning, federated learning, edge computing, and so on. Most traditional distributed learning methods typically assume no failure or attack. However, many unexpected cases, such as communication failure and even malicious attack, may happen in real applications. Hence, Byzantine learning (BL), which refers to distributed learning with failure or attack, has recently attracted much attention. Most existing BL methods are synchronous, which are impractical in some applications due to heterogeneous or offline workers. In these cases, asynchronous BL (ABL) is usually preferred. In this paper, we propose a novel method, called buffered asynchronous stochastic gradient descent (BASGD), for ABL. To the best of our knowledge, BASGD is the first ABL method that can resist non-omniscient attacks without storing any instances on the server. Furthermore, we also propose an improved variant of BASGD, called BASGD with momentum (BASGDm), by introducing local momentum into BASGD. Compared with those methods which need to store instances on server, BASGD and BASGDm have a wider scope of application. Both BASGD and BASGDm are compatible with various aggregation rules. Moreover, both BASGD and BASGDm are proved to be convergent and able to resist failure or attack. Empirical results show that our methods significantly outperform existing ABL baselines when there exists failure or attack on workers.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available