Efficient Marginal Likelihood Computation for Gaussian Process Regression

10/29/2011
by   Andrea Schirru, et al.
0

In a Bayesian learning setting, the posterior distribution of a predictive model arises from a trade-off between its prior distribution and the conditional likelihood of observed data. Such distribution functions usually rely on additional hyperparameters which need to be tuned in order to achieve optimum predictive performance; this operation can be efficiently performed in an Empirical Bayes fashion by maximizing the posterior marginal likelihood of the observed data. Since the score function of this optimization problem is in general characterized by the presence of local optima, it is necessary to resort to global optimization strategies, which require a large number of function evaluations. Given that the evaluation is usually computationally intensive and badly scaled with respect to the dataset size, the maximum number of observations that can be treated simultaneously is quite limited. In this paper, we consider the case of hyperparameter tuning in Gaussian process regression. A straightforward implementation of the posterior log-likelihood for this model requires O(N^3) operations for every iteration of the optimization procedure, where N is the number of examples in the input dataset. We derive a novel set of identities that allow, after an initial overhead of O(N^3), the evaluation of the score function, as well as the Jacobian and Hessian matrices, in O(N) operations. We prove how the proposed identities, that follow from the eigendecomposition of the kernel matrix, yield a reduction of several orders of magnitude in the computation time for the hyperparameter optimization problem. Notably, the proposed solution provides computational advantages even with respect to state of the art approximations that rely on sparse kernel matrices.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/20/2022

Noise Estimation in Gaussian Process Regression

We develop a computational procedure to estimate the covariance hyperpar...
research
04/21/2020

Bayesian Optimization of Hyperparameters when the Marginal Likelihood is Estimated by MCMC

Bayesian models often involve a small set of hyperparameters determined ...
research
11/14/2019

Conjugate Gradients for Kernel Machines

Regularized least-squares (kernel-ridge / Gaussian process) regression i...
research
09/20/2021

Barely Biased Learning for Gaussian Process Regression

Recent work in scalable approximate Gaussian process regression has disc...
research
10/09/2019

Optimal Training of Fair Predictive Models

Recently there has been sustained interest in modifying prediction algor...
research
07/16/2022

A Singular Woodbury and Pseudo-Determinant Matrix Identities and Application to Gaussian Process Regression

We study a matrix that arises in a singular formulation of the Woodbury ...
research
09/24/2019

Numerical evaluation of the transition probability of the simple birth-and-death process

The simple (linear) birth-and-death process is a widely used stochastic ...

Please sign up or login with your details

Forgot password? Click here to reset