3.8 Proceedings Paper

Learning while Respecting Privacy and Robustness to Adversarial Distributed Datasets

出版社

IEEE

关键词

-

资金

  1. University of Minnesota Doctoral Dissertation Fellowship (DDF)
  2. NSF [1901134, 2126052]
  3. ARO STIR [W911NF2110297]
  4. U.S. Department of Defense (DOD) [W911NF2110297] Funding Source: U.S. Department of Defense (DOD)
  5. Directorate For Engineering
  6. Div Of Electrical, Commun & Cyber Sys [2126052] Funding Source: National Science Foundation

向作者/读者索取更多资源

The use of federated learning framework allows for the distributed training of models on massive datasets without the need to upload data to a central server. However, maintaining performance and robustness in the face of varying data distributions across workers is a challenge. This study proposes a distributionally robust optimization framework and develops a primal-dual algorithm to ensure the trained model's robustness against adversarial attacks and distributional uncertainties.
Massive datasets are typically distributed geographically across multiple sites, where scalability, data privacy and integrity, as well as bandwidth scarcity typically discourage uploading these data to a central server. This has propelled the so-called federated learning framework where multiple workers exchange information with a server to learn a centralized model using data locally generated and/or stored across workers. This learning framework necessitates workers to communicate iteratively with the server. Although appealing for its scalability, one needs to carefully address the various data distribution shifts across workers, which degrades the performance of the learnt model. In this context, the distributionally robust optimization framework is considered here. The objective is to endow the trained model with robustness against adversarially manipulated input data, or, distributional uncertainties, such as mismatches between training and testing data distributions, or among datasets stored at different workers. To this aim, the data distribution is assumed unknown, and to land within a Wasserstein ball centered around the empirical data distribution. This robust learning task entails an infinite-dimensional optimization problem, which is challenging. Leveraging a strong duality result, a surrogate is obtained, for which a primal-dual algorithm is developed. Compared to classical methods, the proposed algorithm offers robustness with little computational overhead. Numerical tests using image datasets showcase the merits of the proposed algorithm under several existing adversarial attacks and distributional uncertainties.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据