ANALYSIS OF A NONREVERSIBLE MARKOV CHAIN SAMPLER

We analyze the convergence to stationarity of a simple nonreversible Markov chain that serves as a model for several nonreversible Markov chain sampling methods that are used in practice. Our theoretical and numerical results show that nonreversibility can indeed lead to improvements over the diffusive behavior of simple Markov chain sampling schemes. The analysis uses both probabilistic techniques and an explicit diagonalization. 1. Introduction. Markov chain sampling methods are commonly used in statistics [33, 32], computer science [31], statistical mechanics [3] and quantum field theory [34, 23]. In all these fields, distributions are encountered that are difficult to sample from directly, but for which a Markov chain that converges to the distribution can easily be constructed. For many such methods (e.g., the Metropolis algorithm [25, 13], and the Gibbs sampler [17, 16] with a random scan) the Markov chain constructed is reversible. Some of these methods explore the distribution by means of a diffusive random walk. We use the term “diffusive” for processes like the ordinary random walk on a d-dimensional lattice which require time of order T 2 to travel distance T. Some other common methods, such as the Gibbs sampler with a systematic scan, use a Markov chain that is not reversible, but have diffusive behavior resembling that of a related reversible chain [30]. Some Markov chain methods attempt to avoid the inefficiencies of such diffusive exploration. The Hybrid Monte Carlo method [15] uses an elaborate Metropolis proposal that can make large changes to the state. In a variant of this method due to Horowitz [21], a similar effect is produced using a Markov chain that is carefully designed to be nonreversible. (See [34, 23, 27] for reviews of these methods.) The overrelaxation method [1] also employs a nonreversible Markov chain as a way of suppressing diffusive behavior, as discussed in [29]. In this paper, we analyze a nonreversible Markov chain that does a onedimensional walk, as an abstraction of these practical sampling methods, particularly that of Horowitz [21]. Gustafson [19] has also recently tried using adaptations of Horowitz’s method. We find that the nonreversible walk does indeed converge more rapidly than the usual simple random walk. We ana

[1]  N. Metropolis,et al.  Equation of State Calculations by Fast Computing Machines , 1953, Resonance.

[2]  S. Adler Over-relaxation method for the Monte Carlo evaluation of the partition function for multiquadratic actions , 1981 .

[3]  Donald Geman,et al.  Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  S. Duane,et al.  Hybrid Monte Carlo , 1987 .

[5]  Adrian F. M. Smith,et al.  Bayesian computation via the gibbs sampler and related markov chain monte carlo methods (with discus , 1993 .

[6]  Bradley Efron,et al.  PROBABILISTIC-GEOMETRIC THEOREMS ARISING FROM THE ANALYSIS OF CONTINGENCY TABLES , 1987 .

[7]  P. Diaconis Group representations in probability and statistics , 1988 .

[8]  N. J. A. Sloane,et al.  Gray codes for reflection groups , 1989, Graphs Comb..

[9]  李幼升,et al.  Ph , 1989 .

[10]  D. Toussaint Introduction to algorithms for Monte Carlo simulations and their application to QCD , 1989 .

[11]  A. Kennedy The theory of hybrid stochastic algorithms , 1990 .

[12]  Adrian F. M. Smith,et al.  Sampling-Based Approaches to Calculating Marginal Densities , 1990 .

[13]  P. Diaconis,et al.  Strong Stationary Times Via a New Form of Duality , 1990 .

[14]  A. Horowitz A generalized guided Monte Carlo algorithm , 1991 .

[15]  T. Lindvall Lectures on the Coupling Method , 1992 .

[16]  Alistair Sinclair,et al.  Algorithms for Random Generation and Counting: A Markov Chain Approach , 1993, Progress in Theoretical Computer Science.

[17]  M THEORE,et al.  Moderate Growth and Random Walk on Finite Groups , 1994 .

[18]  P. Diaconis,et al.  Gray codes for randomization procedures , 1994 .

[19]  L. Tierney Markov Chains for Exploring Posterior Distributions , 1994 .

[20]  Persi Diaconis,et al.  What do we know about the Metropolis algorithm? , 1995, STOC '95.

[21]  Geoffrey E. Hinton,et al.  Bayesian Learning for Neural Networks , 1995 .

[22]  P. Diaconis,et al.  Rectangular Arrays with Fixed Margins , 1995 .

[23]  Shing-Tung Yau,et al.  On sampling with Markov chains , 1996, Random Struct. Algorithms.

[24]  Jun S. Liu,et al.  Metropolized independent sampling with comparisons to rejection sampling and importance sampling , 1996, Stat. Comput..

[25]  G. Roberts,et al.  Updating Schemes, Correlation Structure, Blocking and Parameterization for the Gibbs Sampler , 1997 .

[26]  P Gustafson,et al.  Large hierarchical Bayesian analysis of multivariate survival data. , 1997, Biometrics.

[27]  Paul Gustafson,et al.  A guided walk Metropolis algorithm , 1998, Stat. Comput..

[28]  Radford M. Neal,et al.  Suppressing Random Walks in Markov Chain Monte Carlo Using Ordered Overrelaxation , 1995, Learning in Graphical Models.

[29]  P. Diaconis,et al.  Algebraic algorithms for sampling from conditional distributions , 1998 .

[30]  Fang Chen,et al.  Lifting Markov chains to speed up mixing , 1999, STOC '99.

[31]  Antonietta Mira,et al.  Ordering Monte Carlo Markov Chains , 1999 .

[32]  Gerard T. Barkema,et al.  Monte Carlo Methods in Statistical Physics , 1999 .