Convergence properties of data augmentation algorithms for high-dimensional robit regression

12/20/2021
by   Sourav Mukherjee, et al.
0

The logistic and probit link functions are the most common choices for regression models with a binary response. However, these choices are not robust to the presence of outliers/unexpected observations. The robit link function, which is equal to the inverse CDF of the Student's t-distribution, provides a robust alternative to the probit and logistic link functions. A multivariate normal prior for the regression coefficients is the standard choice for Bayesian inference in robit regression models. The resulting posterior density is intractable and a Data Augmentation (DA) Markov chain is used to generate approximate samples from the desired posterior distribution. Establishing geometric ergodicity for this DA Markov chain is important as it provides theoretical guarantees for asymptotic validity of MCMC standard errors for desired posterior expectations/quantiles. Previous work [Roy(2012)] established geometric ergodicity of this robit DA Markov chain assuming (i) the sample size n dominates the number of predictors p, and (ii) an additional constraint which requires the sample size to be bounded above by a fixed constant which depends on the design matrix X. In particular, modern high-dimensional settings where n < p are not considered. In this work, we show that the robit DA Markov chain is trace-class (i.e., the eigenvalues of the corresponding Markov operator are summable) for arbitrary choices of the sample size n, the number of predictors p, the design matrix X, and the prior mean and variance parameters. The trace-class property implies geometric ergodicity. Moreover, this property allows us to conclude that the sandwich robit chain (obtained by inserting an inexpensive extra step in between the two steps of the DA chain) is strictly better than the robit DA chain in an appropriate sense.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/17/2018

Geometric ergodicity of Polya-Gamma Gibbs sampler for Bayesian logistic regression with a flat prior

Logistic regression model is the most popular model for analyzing binary...
research
01/11/2021

Block Gibbs samplers for logistic mixed models: convergence properties and a comparison with full Gibbs samplers

Logistic linear mixed model (LLMM) is one of the most widely used statis...
research
09/18/2021

Asynchronous and Distributed Data Augmentation for Massive Data Settings

Data augmentation (DA) algorithms are widely used for Bayesian inference...
research
04/16/2018

Trace class Markov chains for the Normal-Gamma Bayesian shrinkage model

High-dimensional data, where the number of variables exceeds or is compa...
research
11/18/2019

A Distributed Algorithm for Polya-Gamma Data Augmentation

The Polya-Gamma data augmentation (PG-DA) algorithm is routinely used fo...
research
09/22/2018

A default prior for regression coefficients

When the sample size is not too small, M-estimators of regression coeffi...
research
10/29/2018

Prior-preconditioned conjugate gradient method for accelerated Gibbs sampling in "large n & large p" sparse Bayesian regression

In a modern observational study based on healthcare databases, the numbe...

Please sign up or login with your details

Forgot password? Click here to reset