Optimal Order Simple Regret for Gaussian Process Bandits

08/20/2021
∙
by   Sattar Vakili, et al.
∙
0
∙

Consider the sequential optimization of a continuous, possibly non-convex, and expensive to evaluate objective function f. The problem can be cast as a Gaussian Process (GP) bandit where f lives in a reproducing kernel Hilbert space (RKHS). The state of the art analysis of several learning algorithms shows a significant gap between the lower and upper bounds on the simple regret performance. When N is the number of exploration trials and Îģ_N is the maximal information gain, we prove an 𝒊Ėƒ(√(Îģ_N/N)) bound on the simple regret performance of a pure exploration algorithm that is significantly tighter than the existing bounds. We show that this bound is order optimal up to logarithmic factors for the cases where a lower bound on regret is known. To establish these results, we prove novel and sharp confidence intervals for GP models applicable to RKHS elements which may be of broader interest.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
∙ 09/15/2020

On Information Gain and Regret Bounds in Gaussian Process Bandits

Consider the sequential optimization of an expensive to evaluate and pos...
research
∙ 06/09/2020

Scalable Thompson Sampling using Sparse Gaussian Process Models

Thompson Sampling (TS) with Gaussian Process (GP) models is a powerful t...
research
∙ 10/28/2021

Open Problem: Tight Online Confidence Intervals for RKHS Elements

Confidence intervals are a crucial building block in the analysis of var...
research
∙ 04/19/2013

Parallel Gaussian Process Optimization with Upper Confidence Bound and Pure Exploration

In this paper, we consider the challenge of maximizing an unknown functi...
research
∙ 11/03/2019

Zeroth Order Non-convex optimization with Dueling-Choice Bandits

We consider a novel setting of zeroth order non-convex optimization, whe...
research
∙ 07/06/2021

Weighted Gaussian Process Bandits for Non-stationary Environments

In this paper, we consider the Gaussian process (GP) bandit optimization...
research
∙ 01/25/2016

Time-Varying Gaussian Process Bandit Optimization

We consider the sequential Bayesian optimization problem with bandit fee...

Please sign up or login with your details

Forgot password? Click here to reset