4.7 Article

Differential Privacy Preserving of Training Model in Wireless Big Data with Edge Computing

期刊

IEEE TRANSACTIONS ON BIG DATA
卷 6, 期 2, 页码 283-295

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TBDATA.2018.2829886

关键词

Data privacy; Big Data; Training; Machine learning; Privacy; Wireless communication; Wireless sensor networks; Wireless big data; smart edges; differential privacy; training data privacy; machine learning; correlated datasets; laplacian mechanism; tensorflow

资金

  1. NSFC [61572262]
  2. China Postdoctoral Science Foundation [2017M610252]
  3. China Postdoctoral Science Special Foundation [2017T100297]
  4. Research Council of Norway [240079/F20]
  5. project Security in IoT for Smart Grids part of the IKTPLUSS program - Norwegian Research Council [248113/O70]

向作者/读者索取更多资源

With the popularity of smart devices and the widespread use of machine learning methods, smart edges have become the mainstream of dealing with wireless big data. When smart edges use machine learning models to analyze wireless big data, nevertheless, some models may unintentionally store a small portion of the training data with sensitive records. Thus, intruders can expose sensitive information by careful analysis of this model. To solve this privacy issue, in this paper, we propose and implement a machine learning strategy for smart edges using differential privacy. We focus our attention on privacy protection in training datasets in wireless big data scenario. Moreover, we guarantee privacy protection by adding Laplace mechanisms, and design two different algorithms Output Perturbation (OPP) and Objective Perturbation (OJP), which satisfy differential privacy. In addition, we consider the privacy preserving issues presented in the existing literatures for differential privacy in the correlated datasets, and further provided differential privacy preserving methods for correlated datasets, guaranteeing privacy by theoretical deduction. Finally, we implement the experiments on the TensorFlow, and evaluate our strategy on four datasets, i.e., MNIST, SVHN, CIFAR-10 and STL-10. The experiment results show that our methods can efficiently protect the privacy of training datasets and guarantee the accuracy on benchmark datasets.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据