TY - GEN
T1 - Global and Local Differential Privacy for Collaborative Bandits
AU - Wang, Huazheng
AU - Zhao, Qian
AU - Wu, Qingyun
AU - Chopra, Shubham
AU - Khaitan, Abhinav
AU - Wang, Hongning
N1 - Funding Information:
We thank the anonymous reviewers for their insightful comments. This work was supported in part by Bloomberg Data Science Ph.D. Fellowship and National Science Foundation Award IIS-1553568 and IIS-1618948.
Publisher Copyright:
© 2020 ACM.
PY - 2020/9/22
Y1 - 2020/9/22
N2 - Collaborative bandit learning has become an emerging focus for personalized recommendation. It leverages user dependence for joint model estimation and recommendation. As such online learning solutions directly learn from users, e.g., result clicks, they bring in new challenges in privacy protection. Despite the existence of recent studies about privacy in contextual bandit algorithms, how to efficiently protect user privacy in a collaborative bandit learning environment remains unknown. In this paper, we develop a general solution framework to achieve differential privacy in collaborative bandit algorithms, under the notion of global differential privacy and local differential privacy. The key idea is to inject noise in a bandit model's sufficient statistics (either on server side to achieve global differential privacy or client side to achieve local differential privacy) and calibrate the noise scale with respect to the structure of collaboration among users. We study two popularly used collaborative bandit algorithms to illustrate the application of our solution framework. Theoretical analysis proves our derived private algorithms reduce the added regret caused by privacy-preserving mechanism compared to its linear bandits counterparts, i.e., collaboration actually helps to achieve stronger privacy with the same amount of injected noise. We also empirically evaluate the algorithms on both synthetic and real-world datasets to demonstrate the trade-off between privacy and utility.
AB - Collaborative bandit learning has become an emerging focus for personalized recommendation. It leverages user dependence for joint model estimation and recommendation. As such online learning solutions directly learn from users, e.g., result clicks, they bring in new challenges in privacy protection. Despite the existence of recent studies about privacy in contextual bandit algorithms, how to efficiently protect user privacy in a collaborative bandit learning environment remains unknown. In this paper, we develop a general solution framework to achieve differential privacy in collaborative bandit algorithms, under the notion of global differential privacy and local differential privacy. The key idea is to inject noise in a bandit model's sufficient statistics (either on server side to achieve global differential privacy or client side to achieve local differential privacy) and calibrate the noise scale with respect to the structure of collaboration among users. We study two popularly used collaborative bandit algorithms to illustrate the application of our solution framework. Theoretical analysis proves our derived private algorithms reduce the added regret caused by privacy-preserving mechanism compared to its linear bandits counterparts, i.e., collaboration actually helps to achieve stronger privacy with the same amount of injected noise. We also empirically evaluate the algorithms on both synthetic and real-world datasets to demonstrate the trade-off between privacy and utility.
UR - http://www.scopus.com/inward/record.url?scp=85092707750&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85092707750&partnerID=8YFLogxK
U2 - 10.1145/3383313.3412254
DO - 10.1145/3383313.3412254
M3 - Conference contribution
AN - SCOPUS:85092707750
T3 - RecSys 2020 - 14th ACM Conference on Recommender Systems
SP - 150
EP - 159
BT - RecSys 2020 - 14th ACM Conference on Recommender Systems
PB - Association for Computing Machinery, Inc
T2 - 14th ACM Conference on Recommender Systems, RecSys 2020
Y2 - 22 September 2020 through 26 September 2020
ER -