3.8 Proceedings Paper

Learning while Respecting Privacy and Robustness to Adversarial Distributed Datasets

Publisher

IEEE

Keywords

-

Funding

  1. University of Minnesota Doctoral Dissertation Fellowship (DDF)
  2. NSF [1901134, 2126052]
  3. ARO STIR [W911NF2110297]
  4. U.S. Department of Defense (DOD) [W911NF2110297] Funding Source: U.S. Department of Defense (DOD)
  5. Directorate For Engineering
  6. Div Of Electrical, Commun & Cyber Sys [2126052] Funding Source: National Science Foundation

Ask authors/readers for more resources

The use of federated learning framework allows for the distributed training of models on massive datasets without the need to upload data to a central server. However, maintaining performance and robustness in the face of varying data distributions across workers is a challenge. This study proposes a distributionally robust optimization framework and develops a primal-dual algorithm to ensure the trained model's robustness against adversarial attacks and distributional uncertainties.
Massive datasets are typically distributed geographically across multiple sites, where scalability, data privacy and integrity, as well as bandwidth scarcity typically discourage uploading these data to a central server. This has propelled the so-called federated learning framework where multiple workers exchange information with a server to learn a centralized model using data locally generated and/or stored across workers. This learning framework necessitates workers to communicate iteratively with the server. Although appealing for its scalability, one needs to carefully address the various data distribution shifts across workers, which degrades the performance of the learnt model. In this context, the distributionally robust optimization framework is considered here. The objective is to endow the trained model with robustness against adversarially manipulated input data, or, distributional uncertainties, such as mismatches between training and testing data distributions, or among datasets stored at different workers. To this aim, the data distribution is assumed unknown, and to land within a Wasserstein ball centered around the empirical data distribution. This robust learning task entails an infinite-dimensional optimization problem, which is challenging. Leveraging a strong duality result, a surrogate is obtained, for which a primal-dual algorithm is developed. Compared to classical methods, the proposed algorithm offers robustness with little computational overhead. Numerical tests using image datasets showcase the merits of the proposed algorithm under several existing adversarial attacks and distributional uncertainties.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available