4.7 Article

Distributed online bandit linear regressions with differential privacy

向作者/读者索取更多资源

This paper addresses the distributed online bandit linear regression problems with privacy protection. Two efficient differentially private distributed online regression algorithms are developed, achieving ε-differential privacy and establishing the regret K is the time horizon in the cases of one-point and two-point bandit feedback. Furthermore, a tradeoff between the algorithms' privacy level and convergence is identified.
This paper addresses the distributed online bandit linear regression problems with privacy protection, in which the training data are spread in a multi-agent network. Each node identifies a linear predictor to fit the training data and experiences a square loss on each round. The purpose is to minimize the regret that assesses the difference of the accumulated loss between the online linear predictor and the optimal offline linear predictor. Moreover, the differential privacy strategy is adopted to prevent the adversary from inferring the parameter vector of any node. Two efficient differentially private distributed online regression algorithms are developed in the cases of one-point and two-point bandit feedback. Our upper bounds in O(K3 / 4 ) and O(root K) for one-point and two-point bandit feedback, respectively, where analysis suggests that the developed algorithms achieve e-differential privacy and establish the regret K is the time horizon. We also show that there exists a tradeoff between our algorithms' privacy level and convergence. Finally, the performance of the proposed algorithms is validated by a numerical example.(c) 2023 The Franklin Institute. Published by Elsevier Inc. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据