Lecture Slides 17

Lecture Slides 17 - AMS 210: Applied Linear Algebra...

Info iconThis preview shows pages 1–7. Sign up to view the full content.

View Full Document Right Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon
This is the end of the preview. Sign up to access the rest of the document.

Unformatted text preview: AMS 210: Applied Linear Algebra November 12, 2009 AMS 210 Topics Today Problem Set 9 Markov Chain Stable Distributions Markov Chain Stable Distribution Examples Absorbing States in Markov Chains AMS 210 Problem Set 9 This weeks problem set is due next Thursday, November 19. Read sections 4.2, 4.3 (first two examples), 4.4, 4.5 (first two examples). Exercises: 4.2: 2, 5, 6; 4.3: 2, 3; 4.4: 1, 6, 16, 19. Show work. If multiple sheets, stapling is strongly preferred. Graders might deduct points for insecurely bound/collated problem sets. AMS 210 Solving for Markov Chain Stable Distributions Appeared in section 3.5. If A is the transition matrix and p is set of probabilities of being in certain states, then Ap represents the next state. (This is old.) Solve for when p = Ip = Ap , so when ( I- A ) p = . Alternatively, this happens when there is an eigenvector with an eigenvalue of 1 (this is just a definition of an eigenvalue/vector). Since the set of probabilities involved must sum to 1, require that 1 p = 1. n + 1 equations for n variables; but multiples of eigenvectors are eigenvectors. AMS 210 Markov Chain Stable Distribution Examples: General 2x2 Let A = 1- b a b 1- a . Solve stable state to get- bp 1 + ap 2 = 0, bp 1- ap 2 = 0, and p 1 + p 2 = 1. p 1 = a a + b and p 2 = b a + b . AMS 210 Markov Chain Stable Distributions Aside: matrix in section 3.4, example 1 is wrong. Column 1 sums to .5. Goal: find when distibutions in Markov chains converge to a stable distribution. Definition: A Markov chain with transition matrix A is regular if for some positive integer h , the matrix A h has all positive entries. Theorem: Every regular Markov chain with transition matrix A has a stable probability vector p * to which p ( k ) = A k p converges, for any probability vector p . All the columns of A k p also converge to p * ....
View Full Document

Page1 / 15

Lecture Slides 17 - AMS 210: Applied Linear Algebra...

This preview shows document pages 1 - 7. Sign up to view the full document.

View Full Document Right Arrow Icon
Ask a homework question - tutors are online