Graphical model inference: Sequential Monte Carlo meets deterministic approximations

By Fredrik Lindsten, Jouni Helske and Matti Vihola

January 1, 2018

PDF

Abstract

Approximate inference in probabilistic graphical models (PGMs) can be grouped into deterministic methods and Monte-Carlo-based methods. The former can often provide accurate and rapid inferences, but are typically associated with biases that are hard to quantify. The latter enjoy asymptotic consistency, but can suffer from high computational costs. In this paper we present a way of bridging the gap between deterministic and stochastic inference. Specifically, we suggest an efficient sequential Monte Carlo (SMC) algorithm for PGMs which can leverage the output from deterministic inference methods. While generally applicable, we show explicitly how this can be done with loopy belief propagation, expectation propagation, and Laplace approximations. The resulting algorithm can be viewed as a post-correction of the biases associated with these methods and, indeed, numerical results show clear improvements over the baseline deterministic methods as well as over plain SMC.

Posted on:
January 1, 2018
Length:
1 minute read, 140 words
Tags:
Sequential Monte Carlo
See Also:
bssm: Bayesian Inference of Non-linear and Non-Gaussian State Space Models in R
bssm: Bayesian Inference of Non-linear and Non-Gaussian State Space Models in R
Importance sampling type estimators based on approximate marginal Markov chain Monte Carlo