mcmc-gibbs-intro - Markov Chain Monte Carlo and Gibbs...

Info iconThis preview shows pages 1–3. Sign up to view the full content.

View Full Document Right Arrow Icon
Markov Chain Monte Carlo and Gibbs Sampling Lecture Notes for EEB 581, version 26 April 2004 c ± B. Walsh 2004 A major limitation towards more widespread implementation of Bayesian ap- proaches is that obtaining the posterior distribution often requires the integration of high-dimensional functions. This can be computationally very difficult, but several approaches short of direct integration have been proposed (reviewed by Smith 1991, Evans and Swartz 1995, Tanner 1996). We focus here on Markov Chain Monte Carlo ( MCMC ) methods, which attempt to simulate direct draws from some complex distribution of interest. MCMC approaches are so-named be- cause one uses the previous sample values to randomly generate the next sample value, generating a Markov chain (as the transition probabilities between sample values are only a function of the most recent sample value). The realization in the early 1990’s (Gelfand and Smith 1990) that one particu- lar MCMC method, the Gibbs sampler , is very widely applicable to a broad class of Bayesian problems has sparked a major increase in the application of Bayesian analysis, and this interest is likely to continue expanding for sometime to come. MCMC methods have their roots in the Metropolis algorithm (Metropolis and Ulam 1949, Metropolis et al. 1953), an attempt by physicists to compute com- plex integrals by expressing them as expectations for some distribution and then estimate this expectation by drawing samples from that distribution. The Gibbs sampler (Geman and Geman 1984) has its origins in image processing. It is thus somewhat ironic that the powerful machinery of MCMC methods had essentially no impact on the field of statistics until rather recently. Excellent (and detailed) treatments of MCMC methods are found in Tanner (1996) and Chapter two of Draper (2000). Additional references are given in the particular sections below. MONTE CARLO INTEGRATION The original Monte Carlo approach was a method developed by physicists to use random number generation to compute integrals. Suppose we wish to compute a complex integral Z b a h ( x ) dx (1a) If we can decompose h ( x ) into the production of a function f ( x ) and a probability 1
Background image of page 1

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon
2 MCMC AND GIBBS SAMPLING density function p ( x ) defined over the interval ( a, b ) , then note that Z b a h ( x ) dx = Z b a f ( x ) p ( x ) dx = E p ( x ) [ f ( x ) ] (1b) so that the integral can be expressed as an expectation of f ( x ) over the density p ( x ) . Thus, if we draw a large number x 1 , ··· ,x n of random variables from the density p ( x ) , then Z b a h ( x ) dx = E p ( x ) [ f ( x )] 1 n n X i =1 f ( x i ) (1c) This is referred to as Monte Carlo integration . Monte Carlo integration can be used to approximate posterior (or marginal posterior) distributions required for a Bayesian analysis. Consider the integral I ( y )= R f ( y | x ) p ( x ) dx , which we approximate by b I ( y 1 n n X i =1 f ( y | x i ) (2a) where x i are draws from the density p ( x ) . The estimated Monte Carlo standard error is given by SE 2 [ b I ( y )]= 1 n ˆ 1 n - 1 n X i =1 f ( y | x i ) - b I ( y ) · 2 !
Background image of page 2
Image of page 3
This is the end of the preview. Sign up to access the rest of the document.

This note was uploaded on 02/07/2012 for the course CSCI 5512 taught by Professor Staff during the Spring '08 term at Minnesota.

Page1 / 24

mcmc-gibbs-intro - Markov Chain Monte Carlo and Gibbs...

This preview shows document pages 1 - 3. Sign up to view the full document.

View Full Document Right Arrow Icon
Ask a homework question - tutors are online