Statistical Inference for Polyak-Ruppert Averaged Zeroth-order Stochastic Gradient Algorithm

02/10/2021
by   Yanhao Jin, et al.
0

As machine learning models are deployed in critical applications, it becomes important to not just provide point estimators of the model parameters (or subsequent predictions), but also quantify the uncertainty associated with estimating the model parameters via confidence sets. In the last decade, estimating or training in several machine learning models has become synonymous with running stochastic gradient algorithms. However, computing the stochastic gradients in several settings is highly expensive or even impossible at times. An important question which has thus far not been addressed sufficiently in the statistical machine learning literature is that of equipping zeroth-order stochastic gradient algorithms with practical yet rigorous inferential capabilities. Towards this, in this work, we first establish a central limit theorem for Polyak-Ruppert averaged stochastic gradient algorithm in the zeroth-order setting. We then provide online estimators of the asymptotic covariance matrix appearing in the central limit theorem, thereby providing a practical procedure for constructing asymptotically valid confidence sets (or intervals) for parameter estimation (or prediction) in the zeroth-order setting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/04/2019

Statistical Inference for Model Parameters in Stochastic Gradient Descent via Batch Means

Statistical inference of true model parameters based on stochastic gradi...
research
02/05/2021

Online Statistical Inference for Gradient-free Stochastic Optimization

As gradient-free stochastic optimization gains emerging attention for a ...
research
02/10/2020

A Fully Online Approach for Covariance Matrices Estimation of Stochastic Gradient Descent Solutions

Stochastic gradient descent (SGD) algorithm is widely used for parameter...
research
06/25/2022

Statistical inference with implicit SGD: proximal Robbins-Monro vs. Polyak-Ruppert

The implicit stochastic gradient descent (ISGD), a proximal version of S...
research
07/11/2020

Solving Bayesian Risk Optimization via Nested Stochastic Gradient Estimation

In this paper, we aim to solve Bayesian Risk Optimization (BRO), which i...
research
06/09/2023

A Central Limit Theorem for Stochastic Saddle Point Optimization

In this work, we study the Uncertainty Quantification (UQ) of an algorit...
research
03/15/2012

Combining Spatial and Telemetric Features for Learning Animal Movement Models

We introduce a new graphical model for tracking radio-tagged animals and...

Please sign up or login with your details

Forgot password? Click here to reset