Honest Exploration of Intractable Probability Distributions via Markov Chain Monte Carlo
Top Cited Papers
Open Access
- 1 November 2001
- journal article
- Published by Institute of Mathematical Statistics in Statistical Science
- Vol. 16 (4) , 312-334
- https://doi.org/10.1214/ss/1015346317
Abstract
Two important questions that must be answered whenever a Markov chain Monte Carlo (MCMC) algorithm is used are (Q1) What is an appropriate burn-in? and (Q2) How long should the sampling continue after burn-in?Developing rigorous answers to these questions presently requires a detailed study of the convergence properties of the underlying Markov chain. Consequently, in most practical applications of MCMC, exact answers to (Q1)and (Q2) are not sought. The goal of this paper is to demystify the analysis that leads to honest answers to (Q1) and (Q2). The authors hope that this article will serve as a bridge between those developing Markov chain theory and practitioners using MCMC to solve practical problems. The ability to address (Q1) and (Q2) formally comes from establishing a drift condition and an associated minorization condition, which together imply that the underlying Markov chain is geometrically ergodic. In this article, we explain exactly what drift and minorization are as well as how and why these conditions can be used to form rigorous answers to (Q1) and (Q2). The basic ideas are as follows. The results of Rosenthal (1995) and Roberts and Tweedie (1999) allow one to use drift and minorization conditions to construct a formula giving an analytic upper bound on the distance to stationarity. A rigorous answer to (Q1) can be calculated using this formula. The desired characteristics of the target distribution are typically estimated using ergodic averages. Geometric ergodicity of the underlying Markov chain implies that there are central limit theorems available for ergodic averages (Chan and Geyer 1994). The regenerative simulation technique (Mykland, Tierney and Yu, 1995; Robert, 1995) can be used to get a consistent estimate of the variance of the asymptotic normal distribution. Hence, an asymptotic standard error can be calculated, which provides an answer to (Q2) in the sense that an appropriate time to stop sampling can be determined. The methods are illustrated using a Gibbs sampler for a Bayesian version of the one-way random effects model and a data set concerning styrene exposure.Keywords
This publication has 53 references indexed in Scilit:
- Gibbs Sampling with Diffuse Proper Priors: A Valid Approach to Data-Driven Inference?Journal of Computational and Graphical Statistics, 1998
- Understanding the Metropolis-Hastings AlgorithmThe American Statistician, 1995
- Annealing Markov Chain Monte Carlo with Applications to Ancestral InferenceJournal of the American Statistical Association, 1995
- Minorization Conditions and Convergence Rates for Markov Chain Monte CarloJournal of the American Statistical Association, 1995
- Markov Chains for Exploring Posterior DistributionsThe Annals of Statistics, 1994
- Covariance structure of the Gibbs sampler with applications to the comparisons of estimators and augmentation schemesBiometrika, 1994
- [Practical Markov Chain Monte Carlo]: RejoinderStatistical Science, 1992
- Geometric Bounds for Eigenvalues of Markov ChainsThe Annals of Applied Probability, 1991
- Illustration of Bayesian Inference in Normal Data Models Using Gibbs SamplingJournal of the American Statistical Association, 1990
- Sampling-Based Approaches to Calculating Marginal DensitiesJournal of the American Statistical Association, 1990