Federated Online Clustering of Bandits

08/31/2022
by   Xutong Liu, et al.
0

Contextual multi-armed bandit (MAB) is an important sequential decision-making problem in recommendation systems. A line of works, called the clustering of bandits (CLUB), utilize the collaborative effect over users and dramatically improve the recommendation quality. Owing to the increasing application scale and public concerns about privacy, there is a growing demand to keep user data decentralized and push bandit learning to the local server side. Existing CLUB algorithms, however, are designed under the centralized setting where data are available at a central server. We focus on studying the federated online clustering of bandit (FCLUB) problem, which aims to minimize the total regret while satisfying privacy and communication considerations. We design a new phase-based scheme for cluster detection and a novel asynchronous communication protocol for cooperative bandit learning for this problem. To protect users' privacy, previous differential privacy (DP) definitions are not very suitable, and we propose a new DP notion that acts on the user cluster level. We provide rigorous proofs to show that our algorithm simultaneously achieves (clustered) DP, sublinear communication complexity and sublinear regret. Finally, experimental evaluations show our superior performance compared with benchmark algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/02/2021

Privacy-Preserving Communication-Efficient Federated Multi-Armed Bandits

Communication bottleneck and data privacy are two critical concerns in f...
research
06/08/2023

Federated Linear Contextual Bandits with User-level Differential Privacy

This paper studies federated linear contextual bandits under the notion ...
research
02/27/2023

On Differentially Private Federated Linear Contextual Bandits

We consider cross-silo federated linear contextual bandit (LCB) problem ...
research
06/12/2022

Distributed Differential Privacy in Multi-Armed Bandits

We consider the standard K-armed bandit problem under a distributed trus...
research
05/29/2019

Differential Privacy for Multi-armed Bandits: What Is It and What Is Its Cost?

We introduce a number of privacy definitions for the multi-armed bandit ...
research
12/11/2021

Privacy Amplification via Shuffling for Linear Contextual Bandits

Contextual bandit algorithms are widely used in domains where it is desi...
research
01/28/2023

(Private) Kernelized Bandits with Distributed Biased Feedback

In this paper, we study kernelized bandits with distributed biased feedb...

Please sign up or login with your details

Forgot password? Click here to reset