4.7 Article

Towards Efficient Scheduling of Federated Mobile Devices Under Computational and Statistical Heterogeneity

期刊

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TPDS.2020.3023905

关键词

Mobile handsets; Computational modeling; Training; Task analysis; Distributed databases; Convergence; Servers; Federated learning; on-device deep learning; scheduling optimization; non-IID data

资金

  1. US NSF [CCF-1850045, IIS-2007386]
  2. State of Virginia Commonwealth Cyber Initiative

向作者/读者索取更多资源

Federated learning allows for privacy-preserved collaboration by sharing only model parameters, but faces challenges on mobile devices due to computational heterogeneity and non-identically distributed data. To address these issues, two efficient algorithms are proposed for scheduling workloads based on data distribution, achieving significant performance improvements in empirical experiments.
Originated from distributed learning, federated learning enables privacy-preserved collaboration on a new abstracted level by sharing the model parameters only. While the current research mainly focuses on optimizing learning algorithms and minimizing communication overhead left by distributed learning, there is still a considerable gap when it comes to the real implementation on mobile devices. In this article, we start with an empirical experiment to demonstrate computation heterogeneity is a more pronounced bottleneck than communication on the current generation of battery-powered mobile devices, and the existing methods are haunted by mobile stragglers. Further, non-identically distributed data across the mobile users makes the selection of participants critical to the accuracy and convergence. To tackle the computational and statistical heterogeneity, we utilize data as a tuning knob and propose two efficient polynomial-time algorithms to schedule different workloads on various mobile devices, when data is identically or non-identically distributed. For identically distributed data, we combine partitioning and linear bottleneck assignment to achieve near-optimal training time without accuracy loss. For non-identically distributed data, we convert it into an average cost minimization problem and propose a greedy algorithm to find a reasonable balance between computation time and accuracy. We also establish an offline profiler to quantify the runtime behavior of different devices, which serves as the input to the scheduling algorithms. We conduct extensive experiments on a mobile testbed with two datasets and up to 20 devices. Compared with the common benchmarks, the proposed algorithms achieve 2-100x speedup epoch-wise, 2-7 percent accuracy gain and boost the convergence rate by more than 100 percent on CIFAR10.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据