DeepAI AI Chat
Log In Sign Up

Optimal Scaling Results for a Wide Class of Proximal MALA Algorithms

by   Francesca R Crucinio, et al.

We consider a recently proposed class of MCMC methods which uses proximity maps instead of gradients to build proposal mechanisms which can be employed for both differentiable and non-differentiable targets. These methods have been shown to be stable for a wide class of targets, making them a valuable alternative to Metropolis-adjusted Langevin algorithms (MALA); and have found wide application in imaging contexts. The wider stability properties are obtained by building the Moreau-Yoshida envelope for the target of interest, which depends on a parameter λ. In this work, we investigate the optimal scaling problem for this class of algorithms, which encompasses MALA, and provide practical guidelines for the implementation of these methods.


page 1

page 2

page 3

page 4


Optimal Scaling of MCMC Beyond Metropolis

The problem of optimally scaling the proposal distribution in a Markov c...

Optimal Scaling for the Proximal Langevin Algorithm in High Dimensions

The Metropolis-adjusted Langevin (MALA) algorithm is a sampling algorith...

Counterexamples for optimal scaling of Metropolis-Hastings chains with rough target densities

For sufficiently smooth targets of product form it is known that the var...

Optimal Scaling of Metropolis Algorithms on General Target Distributions

The main limitation of the existing optimal scaling results for Metropol...

Optimal scaling of random walk Metropolis algorithms using Bayesian large-sample asymptotics

High-dimensional limit theorems have been shown to be useful to derive t...

Asymptotic distribution of certain degenerate V- and U-statistics with estimated parameters

The asymptotic distribution of a wide class of V- and U-statistics with ...

Optimal design of the Barker proposal and other locally-balanced Metropolis-Hastings algorithms

We study the class of first-order locally-balanced Metropolis–Hastings a...