Markov Chain Monte Carlo Methods Based on `Slicing' the Density Function

One way to sample from a distribution is to sample uniformly from the region under the plot of its density function. A Markov chain that converges to this uniform distribution can be constructed by alternating uniform sampling in the vertical direction with uniform sampling from the horizontaìslice' deened by the current vertical position. Variations on such`slice sampling' methods can easily be implemented for univariate distributions , and can be used to sample from a multivariate distribution by updating each variable in turn. This approach is often easier to implement than Gibbs sampling, and may be more eecient than easily-constructed versions of the Metropolis algorithm. Slice sampling is therefore attractive in routine Markov chain Monte Carlo applications, and for use by software that automatically generates a Markov chain sampler from a model specii-cation. One can also easily devise overrelaxed versions of slice sampling, which sometimes greatly improve sampling eeciency by suppressing random walk behaviour. Random walks can also be avoided in some slice sampling schemes that simultaneously update all variables.

[1]  N. Metropolis,et al.  Equation of State Calculations by Fast Computing Machines , 1953, Resonance.

[2]  W. K. Hastings,et al.  Monte Carlo Sampling Methods Using Markov Chains and Their Applications , 1970 .

[3]  S. Adler Over-relaxation method for the Monte Carlo evaluation of the partition function for multiquadratic actions , 1981 .

[4]  Wang,et al.  Nonuniversal critical dynamics in Monte Carlo simulations. , 1987, Physical review letters.

[5]  A. Sokal,et al.  Generalization of the Fortuin-Kasteleyn-Swendsen-Wang representation and Monte Carlo algorithm. , 1988, Physical review. D, Particles and fields.

[6]  Adrian F. M. Smith,et al.  Sampling-Based Approaches to Calculating Marginal Densities , 1990 .

[7]  P. Barone,et al.  Improving Stochastic Relaxation for Gussian Random Fields , 1990, Probability in the Engineering and Informational Sciences.

[8]  A. Horowitz A generalized guided Monte Carlo algorithm , 1991 .

[9]  P. Green,et al.  Metropolis Methods, Gaussian Proposals and Antithetic Variables , 1992 .

[10]  Radford M. Neal An improved acceptance procedure for the hybrid Monte Carlo algorithm , 1992, hep-lat/9208011.

[11]  J. Besag,et al.  Spatial Statistics and Bayesian Computation , 1993 .

[12]  Peter Green,et al.  Spatial statistics and Bayesian computation (with discussion) , 1993 .

[13]  Walter R. Gilks,et al.  Adaptive rejection metropolis sampling , 1995 .

[14]  W. Gilks,et al.  Adaptive Rejection Metropolis Sampling Within Gibbs Sampling , 1995 .

[15]  Geoffrey E. Hinton,et al.  Bayesian Learning for Neural Networks , 1995 .

[16]  Brendan J. Frey,et al.  Continuous Sigmoidal Belief Networks Trained using Slice Sampling , 1996, NIPS.

[17]  Antonietta Mira,et al.  On the use of auxiliary variables in Markov chain Monte Carlo sampling , 1997 .

[18]  Radford M. Neal,et al.  Suppressing Random Walks in Markov Chain Monte Carlo Using Ordered Overrelaxation , 1995, Learning in Graphical Models.

[19]  D. Higdon Auxiliary Variable Methods for Markov Chain Monte Carlo with Applications , 1998 .

[20]  P. Damlen,et al.  Gibbs sampling for Bayesian non‐conjugate and hierarchical models by using auxiliary variables , 1999 .

[21]  J. Rosenthal,et al.  Convergence of Slice Sampler Markov Chains , 1999 .