State and parameter learning with PaRIS particle Gibbs

01/02/2023
by   Gabriel Cardoso, et al.
0

Non-linear state-space models, also known as general hidden Markov models, are ubiquitous in statistical machine learning, being the most classical generative models for serial data and sequences in general. The particle-based, rapid incremental smoother PaRIS is a sequential Monte Carlo (SMC) technique allowing for efficient online approximation of expectations of additive functionals under the smoothing distribution in these models. Such expectations appear naturally in several learning contexts, such as likelihood estimation (MLE) and Markov score climbing (MSC). PARIS has linear computational complexity, limited memory requirements and comes with non-asymptotic bounds, convergence results and stability guarantees. Still, being based on self-normalised importance sampling, the PaRIS estimator is biased. Our first contribution is to design a novel additive smoothing algorithm, the Parisian particle Gibbs PPG sampler, which can be viewed as a PaRIS algorithm driven by conditional SMC moves, resulting in bias-reduced estimates of the targeted quantities. We substantiate the PPG algorithm with theoretical results, including new bounds on bias and variance as well as deviation inequalities. Our second contribution is to apply PPG in a learning framework, covering MLE and MSC as special examples. In this context, we establish, under standard assumptions, non-asymptotic bounds highlighting the value of bias reduction and the implicit Rao–Blackwellization of PPG. These are the first non-asymptotic results of this kind in this setting. We illustrate our theoretical results with numerical experiments supporting our claims.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/21/2022

Particle-based, rapid incremental smoother meets particle Gibbs

The particle-based, rapid incremental smoother (PARIS) is a sequential M...
research
12/28/2018

Particle-based adaptive-lag online marginal smoothing in general state-space models

We present a novel algorithm, an adaptive-lag smoother, approximating ef...
research
04/04/2022

Variance estimation for Sequential Monte Carlo Algorithms: a backward sampling approach

In this paper, we consider the problem of online asymptotic variance est...
research
08/01/2021

Fast and numerically stable particle-based online additive smoothing: the AdaSmooth algorithm

We present a novel sequential Monte Carlo approach to online smoothing o...
research
06/09/2015

Measuring Sample Quality with Stein's Method

To improve the efficiency of Monte Carlo estimation, practitioners are t...
research
08/30/2017

Asymptotic Bias of Stochastic Gradient Search

The asymptotic behavior of the stochastic gradient algorithm with a bias...
research
06/20/2022

Mobility estimation for Langevin dynamics using control variates

The scaling of the mobility of two-dimensional Langevin dynamics in a pe...

Please sign up or login with your details

Forgot password? Click here to reset