4.6 Article

Shielding Collaborative Learning: Mitigating Poisoning Attacks Through Client-Side Detection

Journal

IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING
Volume 18, Issue 5, Pages 2029-2041

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TDSC.2020.2986205

Keywords

Collaborative work; Data models; Servers; Computational modeling; Training; Task analysis; Training data; Poisoning attack; collaborative learning; deep learning; privacy

Funding

  1. NSFC [61822207, U1636219, U1804263]
  2. Equipment Pre-Research Joint Fund of Ministry of Education of China (Youth Talent) [6141A02033327]
  3. Outstanding Youth Foundation of Hubei Province [2017CFA047]
  4. Fundamental Research Funds for the Central Universities [2042019kf0210]
  5. Plan for Scientific Innovation Talent of Henan Province [184200510018]

Ask authors/readers for more resources

The article presents a novel defense scheme to detect anomalous updates in both IID and non-IID settings, while protecting client-level privacy to prevent poisoning attacks. Experimental evaluations show that the scheme is significantly robust against representative attacks.
Collaborative learning allows multiple clients to train a joint model without sharing their data with each other. Each client performs training locally and then submits the model updates to a central server for aggregation. Since the server has no visibility into the process of generating the updates, collaborative learning is vulnerable to poisoning attacks where a malicious client can generate a poisoned update to introduce backdoor functionality to the joint model. The existing solutions for detecting poisoned updates, however, fail to defend against the recently proposed attacks, especially in the non-IID (independent and identically distributed) setting. In this article, we present a novel defense scheme to detect anomalous updates in both IID and non-IID settings. Our key idea is to realize client-side cross-validation, where each update is evaluated over other clients' local data. The server will adjust the weights of the updates based on the evaluation results when performing aggregation. To adapt to the unbalanced distribution of data in the non-IID setting, a dynamic client allocation mechanism is designed to assign detection tasks to the most suitable clients. During the detection process, we also protect the client-level privacy to prevent malicious clients from knowing the participations of other clients, by integrating differential privacy with our design without degrading the detection performance. Our experimental evaluations on three real-world datasets show that our scheme is significantly robust to two representative poisoning attacks.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available