Improved Algorithm on Online Clustering of Bandits

02/25/2019
by   Wei Chen, et al.
10

We generalize the setting of online clustering of bandits by allowing non-uniform distribution over user frequencies. A more efficient algorithm is proposed with simple set structures to represent clusters. We prove a regret bound for the new algorithm which is free of the minimal frequency over users. The experiments on both synthetic and real datasets consistently show the advantage of the new algorithm over existing methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/14/2018

KL-UCB-switch: optimal regret bounds for stochastic bandits from both a distribution-dependent and a distribution-free viewpoints

In the context of K-armed stochastic bandits with distribution only assu...
research
10/20/2020

POND: Pessimistic-Optimistic oNline Dispatch

This paper considers constrained online dispatch with unknown arrival, r...
research
10/21/2022

Anonymous Bandits for Multi-User Systems

In this work, we present and study a new framework for online learning i...
research
06/21/2019

Randomized Exploration in Generalized Linear Bandits

We study two randomized algorithms for generalized linear bandits, GLM-T...
research
09/15/2019

Online k-means Clustering

We study the problem of online clustering where a clustering algorithm h...
research
03/09/2019

Linear Bandits with Feature Feedback

This paper explores a new form of the linear bandit problem in which the...
research
06/01/2017

Scalable Generalized Linear Bandits: Online Computation and Hashing

Generalized Linear Bandits (GLBs), a natural extension of the stochastic...

Please sign up or login with your details

Forgot password? Click here to reset