Towards Optimal Algorithms for Multi-Player Bandits without Collision Sensing Information

03/24/2021
by   Wei Huang, et al.
35

We propose a novel algorithm for multi-player multi-armed bandits without collision sensing information. Our algorithm circumvents two problems shared by all state-of-the-art algorithms: it does not need as an input a lower bound on the minimal expected reward of an arm, and its performance does not scale inversely proportionally to the minimal expected reward. We prove a theoretical regret upper bound to justify these claims. We complement our theoretical results with numerical experiments, showing that the proposed algorithm outperforms state-of-the-art in practice as well.

READ FULL TEXT
research
02/19/2021

A High Performance, Low Complexity Algorithm for Multi-Player Bandits Without Collision Sensing Information

Motivated by applications in cognitive radio networks, we consider the d...
research
11/07/2017

Multi-Player Bandits Models Revisited

Multi-player Multi-Armed Bandits (MAB) have been extensively studied in ...
research
06/25/2021

Multi-player Multi-armed Bandits with Collision-Dependent Reward Distributions

We study a new stochastic multi-player multi-armed bandits (MP-MAB) prob...
research
11/15/2022

Multi-Player Bandits Robust to Adversarial Collisions

Motivated by cognitive radios, stochastic Multi-Player Multi-Armed Bandi...
research
09/06/2022

Multi-Armed Bandits with Self-Information Rewards

This paper introduces the informational multi-armed bandit (IMAB) model ...
research
07/14/2020

Optimal Learning for Structured Bandits

We study structured multi-armed bandits, which is the problem of online ...
research
10/27/2021

Heterogeneous Multi-player Multi-armed Bandits: Closing the Gap and Generalization

Despite the significant interests and many progresses in decentralized m...

Please sign up or login with your details

Forgot password? Click here to reset