Consistent Online Gaussian Process Regression Without the Sample Complexity Bottleneck

04/23/2020
by   Alec Koppel, et al.
2

Gaussian processes provide a framework for nonlinear nonparametric Bayesian inference widely applicable across science and engineering. Unfortunately, their computational burden scales cubically with the training sample size, which in the case that samples arrive in perpetuity, approaches infinity. This issue necessitates approximations for use with streaming data, which to date mostly lack convergence guarantees. Thus, we develop the first online Gaussian process approximation that preserves convergence to the population posterior, i.e., asymptotic posterior consistency, while ameliorating its intractable complexity growth with the sample size. We propose an online compression scheme that, following each a posteriori update, fixes an error neighborhood with respect to the Hellinger metric centered at the current posterior, and greedily tosses out past kernel dictionary elements until its boundary is hit. We call the resulting method Parsimonious Online Gaussian Processes (POG). For diminishing error radius, exact asymptotic consistency is preserved (Theorem 1(i)) at the cost of unbounded memory in the limit. On the other hand, for constant error radius, POG converges to a neighborhood of the population posterior (Theorem 1(ii))but with finite memory at-worst determined by the metric entropy of the feature space (Theorem 2). Experimental results are presented on several nonlinear regression problems which illuminates the merits of this approach as compared with alternatives that fix the subspace dimension defining the history of past points.

READ FULL TEXT

page 9

page 10

research
08/14/2023

Improved Scaling with Dimension in the Bernstein-von Mises Theorem for Two Statistical Models

Past works have shown that the Bernstein-von Mises theorem, on the asymp...
research
10/05/2020

Bayesian Fixed-domain Asymptotics: Bernstein-von Mises Theorem for Covariance Parameters in a Gaussian Process Model

Gaussian process models typically contain finite dimensional parameters ...
research
03/02/2021

Kernel Interpolation for Scalable Online Gaussian Processes

Gaussian processes (GPs) provide a gold standard for performance in onli...
research
11/27/2022

Radial Neighbors for Provably Accurate Scalable Approximations of Gaussian Processes

In geostatistical problems with massive sample size, Gaussian processes ...
research
07/26/2021

Wasserstein-Splitting Gaussian Process Regression for Heterogeneous Online Bayesian Inference

Gaussian processes (GPs) are a well-known nonparametric Bayesian inferen...
research
01/08/2023

Skewed Bernstein-von Mises theorem and skew-modal approximations

Deterministic Gaussian approximations of intractable posterior distribut...
research
09/25/2019

Optimally Compressed Nonparametric Online Learning

Batch training of machine learning models based on neural networks is no...

Please sign up or login with your details

Forgot password? Click here to reset