Highly Scalable Maximum Likelihood and Conjugate Bayesian Inference for ERGMs on Graph Sets with Equivalent Vertices

10/26/2021
by   Fan Yin, et al.
0

The exponential family random graph modeling (ERGM) framework provides a flexible approach for the statistical analysis of networks. As ERGMs typically involve normalizing factors that are costly to compute, practical inference relies on a variety of approximations or other workarounds. Markov Chain Monte Carlo maximum likelihood (MCMC MLE) provides a powerful tool to approximate the MLE of ERGM parameters, and is feasible for typical models on single networks with as many as a few thousand nodes. MCMC-based algorithms for Bayesian analysis are more expensive, and high-quality answers are challenging to obtain on large graphs. For both strategies, extension to the pooled case - in which we observe multiple networks from a common generative process - adds further computational cost, with both time and memory scaling linearly in the number of graphs. This becomes prohibitive for large networks, or where large numbers of graph observations are available. Here, we exploit some basic properties of the discrete exponential families to develop an approach for ERGM inference in the pooled case that (where applicable) allows an arbitrarily large number of graph observations to be fit at no additional computational cost beyond preprocessing the data itself. Moreover, a variant of our approach can also be used to perform Bayesian inference under conjugate priors, again with no additional computational cost in the estimation phase. As we show, the conjugate prior is easily specified, and is well-suited to applications such as regularization. Simulation studies show that the pooled method leads to estimates with good frequentist properties, and posterior estimates under the conjugate prior are well-behaved. We demonstrate our approach with applications to pooled analysis of brain functional connectivity networks and to replicated x-ray crystal structures of hen egg-white lysozyme.

READ FULL TEXT
research
10/15/2019

Challenges in Bayesian inference via Markov chain Monte Carlo for neural networks

Markov chain Monte Carlo (MCMC) methods and neural networks are instrume...
research
04/06/2018

Computationally efficient inference for latent position network models

Latent position models are nowadays widely used for the analysis of netw...
research
06/16/2022

Generalised Bayesian Inference for Discrete Intractable Likelihood

Discrete state spaces represent a major computational challenge to stati...
research
04/17/2020

Kernel-based Approximate Bayesian Inference for Exponential Family Random Graph Models

Bayesian inference for exponential family random graph models (ERGMs) is...
research
11/08/2010

Efficient Bayesian Inference for Generalized Bradley-Terry Models

The Bradley-Terry model is a popular approach to describe probabilities ...
research
12/06/2017

Fast spatial inference in the homogeneous Ising model

The Ising model is important in statistical modeling and inference in ma...
research
05/21/2021

Removing the mini-batching error in Bayesian inference using Adaptive Langevin dynamics

The computational cost of usual Monte Carlo methods for sampling a poste...

Please sign up or login with your details

Forgot password? Click here to reset