Pure-Exploration for Infinite-Armed Bandits with General Arm Reservoirs

11/15/2018
by   Maryam Aziz, et al.
0

This paper considers a multi-armed bandit game where the number of arms is much larger than the maximum budget and is effectively infinite. We characterize necessary and sufficient conditions on the total budget for an algorithm to return an ϵ-good arm with probability at least 1 - δ. In such situations, the sample complexity depends on ϵ, δ and the so-called reservoir distribution ν from which the means of the arms are drawn iid. While a substantial literature has developed around analyzing specific cases of ν such as the beta distribution, our analysis makes no assumption about the form of ν. Our algorithm is based on successive halving with the surprising exception that arms start to be discarded after just a single pull, requiring an analysis that goes beyond concentration alone. The provable correctness of this algorithm also provides an explanation for the empirical observation that the most aggressive bracket of the Hyperband algorithm of Li et al. (2017) for hyperparameter tuning is almost always best.

READ FULL TEXT
research
05/14/2012

Multiple Identifications in Multi-Armed Bandits

We study the problem of identifying the top m arms in a multi-armed band...
research
09/08/2016

On Sequential Elimination Algorithms for Best-Arm Identification in Multi-Armed Bandits

We consider the best-arm identification problem in multi-armed bandits, ...
research
07/09/2017

Nonlinear Sequential Accepts and Rejects for Identification of Top Arms in Stochastic Bandits

We address the M-best-arm identification problem in multi-armed bandits....
research
10/29/2020

Learning to Actively Learn: A Robust Approach

This work proposes a procedure for designing algorithms for specific ada...
research
06/03/2023

Asymptotically Optimal Pure Exploration for Infinite-Armed Bandits

We study pure exploration with infinitely many bandit arms generated i.i...
research
11/27/2022

Constrained Pure Exploration Multi-Armed Bandits with a Fixed Budget

We consider a constrained, pure exploration, stochastic multi-armed band...
research
06/16/2020

Finding All ε-Good Arms in Stochastic Bandits

The pure-exploration problem in stochastic multi-armed bandits aims to f...

Please sign up or login with your details

Forgot password? Click here to reset