Circularly-Coupled Markov Chain Sampling

I show how to run an N-time-step Markov chain simulation in a circular fashion, so that the state at time 0 follows the state at time N-1 in the same way as states at times t follow those at times t-1 for 0<t<N. This wrap-around of the chain is achieved using a coupling procedure, and produces states that all have close to the equilibrium distribution of the Markov chain, under the assumption that coupled chains are likely to coalesce in less than N/2 iterations. This procedure therefore automatically eliminates the initial portion of the chain that would otherwise need to be discarded to get good estimates of equilibrium averages. The assumption of rapid coalescence can be tested using auxiliary chains started at times spaced between 0 and N. When multiple processors are available, such auxiliary chains can be simulated in parallel, and pieced together to give the circularly-coupled chain, in less time than a sequential simulation would have taken, provided that coalescence is indeed rapid. The practical utility of these procedures is dependent on the development of good coupling schemes. I show how a specialized random-grid Metropolis algorithm can be used to produce the required exact coalescence. On its own, this method is not efficient in high dimensions, but it can be used to produce exact coalescence once other methods have brought the coupled chains close together. I investigate how well this combined scheme works with standard Metropolis, Langevin, and Gibbs sampling updates. Using such strategies, I show that circular coupling can work effectively in a Bayesian logistic regression problem.

[1]  Jeffrey S. Rosenthal,et al.  Convergence Rates for Markov Chains , 1995, SIAM Rev..

[2]  Charles J. Geyer,et al.  Practical Markov Chain Monte Carlo , 1992 .

[3]  Radford M. Neal,et al.  Improving Markov chain Monte Carlo Estimators by Coupling to an Approximating Chain , 2001 .

[4]  Sally Rosenthal,et al.  Parallel computing and Monte Carlo algorithms , 1999 .

[5]  V. Johnson Studying Convergence of Markov Chain Monte Carlo Algorithms Using Coupled Sample Paths , 1996 .

[6]  W. K. Hastings,et al.  Monte Carlo Sampling Methods Using Markov Chains and Their Applications , 1970 .

[7]  James Allen Fill,et al.  An interruptible algorithm for perfect sampling via Markov chains , 1997, STOC '97.

[8]  Alistair Sinclair,et al.  Algorithms for Random Generation and Counting: A Markov Chain Approach , 1993, Progress in Theoretical Computer Science.

[9]  S. Duane,et al.  Hybrid Monte Carlo , 1987 .

[10]  A. Horowitz A generalized guided Monte Carlo algorithm , 1991 .

[11]  Radford M. Neal Sampling from multimodal distributions using tempered transitions , 1996, Stat. Comput..

[12]  J. Rosenthal Minorization Conditions and Convergence Rates for Markov Chain Monte Carlo , 1995 .

[13]  V. Johnson A Coupling-Regeneration Scheme for Diagnosing Convergence in Markov Chain Monte Carlo Algorithms , 1998 .

[14]  J. Rosenthal,et al.  Possible biases induced by MCMC convergence diagnostics , 1999 .

[15]  Benjamin D. Shaw Monte Carlo Methods , 2017 .

[16]  Gareth O. Roberts,et al.  Convergence assessment techniques for Markov chain Monte Carlo , 1998, Stat. Comput..

[17]  Christian P. Robert,et al.  MCMC Convergence Diagnostics : A « Reviewww » , 1998 .

[18]  David Bruce Wilson,et al.  Exact sampling with coupled Markov chains and applications to statistical mechanics , 1996, Random Struct. Algorithms.

[19]  T. Lindvall Lectures on the Coupling Method , 1992 .

[20]  W. Gilks,et al.  Adaptive Rejection Sampling for Gibbs Sampling , 1992 .

[21]  Berend Smit,et al.  Understanding molecular simulation: from algorithms to applications , 1996 .

[22]  A. Gelman,et al.  Weak convergence and optimal scaling of random walk Metropolis algorithms , 1997 .

[23]  J. Doll,et al.  Brownian dynamics as smart Monte Carlo simulation , 1978 .

[24]  D. Rubin,et al.  Inference from Iterative Simulation Using Multiple Sequences , 1992 .

[25]  Peter Green,et al.  Exact sampling for Bayesian inference: towards general purpose algorithms , 1998 .

[26]  M. Tanner,et al.  Facilitating the Gibbs Sampler: The Gibbs Stopper and the Griddy-Gibbs Sampler , 1992 .

[27]  Geoffrey E. Hinton,et al.  Bayesian Learning for Neural Networks , 1995 .