4.7 Article

Differential Privacy Preserving of Training Model in Wireless Big Data with Edge Computing

Journal

IEEE TRANSACTIONS ON BIG DATA
Volume 6, Issue 2, Pages 283-295

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TBDATA.2018.2829886

Keywords

Data privacy; Big Data; Training; Machine learning; Privacy; Wireless communication; Wireless sensor networks; Wireless big data; smart edges; differential privacy; training data privacy; machine learning; correlated datasets; laplacian mechanism; tensorflow

Funding

  1. NSFC [61572262]
  2. China Postdoctoral Science Foundation [2017M610252]
  3. China Postdoctoral Science Special Foundation [2017T100297]
  4. Research Council of Norway [240079/F20]
  5. project Security in IoT for Smart Grids part of the IKTPLUSS program - Norwegian Research Council [248113/O70]

Ask authors/readers for more resources

With the popularity of smart devices and the widespread use of machine learning methods, smart edges have become the mainstream of dealing with wireless big data. When smart edges use machine learning models to analyze wireless big data, nevertheless, some models may unintentionally store a small portion of the training data with sensitive records. Thus, intruders can expose sensitive information by careful analysis of this model. To solve this privacy issue, in this paper, we propose and implement a machine learning strategy for smart edges using differential privacy. We focus our attention on privacy protection in training datasets in wireless big data scenario. Moreover, we guarantee privacy protection by adding Laplace mechanisms, and design two different algorithms Output Perturbation (OPP) and Objective Perturbation (OJP), which satisfy differential privacy. In addition, we consider the privacy preserving issues presented in the existing literatures for differential privacy in the correlated datasets, and further provided differential privacy preserving methods for correlated datasets, guaranteeing privacy by theoretical deduction. Finally, we implement the experiments on the TensorFlow, and evaluate our strategy on four datasets, i.e., MNIST, SVHN, CIFAR-10 and STL-10. The experiment results show that our methods can efficiently protect the privacy of training datasets and guarantee the accuracy on benchmark datasets.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available