4.7 Article

Differentially Private Unknown Worker Recruitment for Mobile Crowdsensing Using Multi-Armed Bandits

期刊

IEEE TRANSACTIONS ON MOBILE COMPUTING
卷 20, 期 9, 页码 2779-2794

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TMC.2020.2990221

关键词

Differential privacy; mobile crowdsensing; multi-armed bandit; worker recruitment

资金

  1. National Key R&D Program of China [2018AAA0101204]
  2. National Natural Science Foundation of China (NSFC) [61872330, 61572457, 61379132, U1709217, 61572342]
  3. NSF [CNS 1757533, CNS 1629746, CNS 1564128, CNS 1449860, CNS 1461932, CNS 1460971, IIP 1439672]
  4. NSF of Jiangsu Province in China [BK20191194, BK20131174, BK2009150]
  5. Anhui Initiative in Quantum Information Technologies [AHY150300]

向作者/读者索取更多资源

This paper focuses on the privacy-preserving recruitment of unknown workers in mobile crowdsensing, proposing algorithms based on differentially private multi-armed bandit games to achieve near-optimal expected accumulative rewards within a given budget. Extensive simulations validate the significant performances of the proposed algorithms based on both real-trace and synthetic datasets.
Mobile crowdsensing is a new paradigm by which a platform can recruit mobile workers to perform some sensing tasks by using their smart mobile devices. In this paper, we focus on a privacy-preserving unknown worker recruitment issue. The platform needs to recruit some workers without knowing the qualities of them completing tasks. Meanwhile, these quality information also needs to be protected from disclosure. To tackle these challenges, we model the unknown worker recruitment as a Differentially Private Multi-Armed Bandit (DP-MAB) game by seeing each worker as an arm of DP-MAB and the task completion quality contributed by each worker as the reward of pulling arm. Then, recruiting workers is equivalent to designing a bandit policy of pulling DP-MAB arms. Under this model, we propose a Differentially Private epsilon-First-based arm-pulling (DPF) algorithm and a Differentially Private UCB-based arm-pulling (DPU) algorithm, which can achieve the nearly optimal expected accumulative rewards under a given budget. We also analyze the regrets of the DPF and DPU algorithms and prove that both of them are delta-differentially private on the task completion qualities (delta > 0). Finally, we conduct extensive simulations to verify the significant performances of DPF and DPU based on both the real-trace and synthetic datasets.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据