Online Clustering of Bandits

01/31/2014
by   Claudio Gentile, et al.
0

We introduce a novel algorithmic approach to content recommendation based on adaptive clustering of exploration-exploitation ("bandit") strategies. We provide a sharp regret analysis of this algorithm in a standard stochastic noise setting, demonstrate its scalability properties, and prove its effectiveness on a number of artificial and real-world datasets. Our experiments show a significant increase in prediction performance over state-of-the-art methods for bandit problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/04/2013

A Gang of Bandits

Multi-armed bandit problems are receiving a great deal of attention beca...
research
08/06/2016

On Context-Dependent Clustering of Bandits

We investigate a novel cluster-of-bandit algorithm CAB for collaborative...
research
04/26/2016

Distributed Clustering of Linear Bandits in Peer to Peer Networks

We provide two distributed confidence ball algorithms for solving linear...
research
10/12/2015

Context-Aware Bandits

We propose an efficient Context-Aware clustering of Bandits (CAB) algori...
research
10/30/2021

Efficient Inference Without Trading-off Regret in Bandits: An Allocation Probability Test for Thompson Sampling

Using bandit algorithms to conduct adaptive randomised experiments can m...
research
02/25/2021

Batched Neural Bandits

In many sequential decision-making problems, the individuals are split i...
research
02/23/2018

An Algorithmic Framework to Control Bias in Bandit-based Personalization

Personalization is pervasive in the online space as it leads to higher e...

Please sign up or login with your details

Forgot password? Click here to reset