18 175 lecture 32 more markov chains
play

18.175: Lecture 32 More Markov chains Scott Sheffield MIT 1 18.175 - PowerPoint PPT Presentation

18.175: Lecture 32 More Markov chains Scott Sheffield MIT 1 18.175 Lecture 32 Outline General setup and basic properties Recurrence and transience 2 18.175 Lecture 32 Outline General setup and basic properties Recurrence and transience 3 18.175


  1. 18.175: Lecture 32 More Markov chains Scott Sheffield MIT 1 18.175 Lecture 32

  2. Outline General setup and basic properties Recurrence and transience 2 18.175 Lecture 32

  3. Outline General setup and basic properties Recurrence and transience 3 18.175 Lecture 32

  4. Markov chains: general definition � Consider a measurable space ( S , S ). � A function p : S × S → R is a transition probability if � For each x ∈ S , A → p ( x , A ) is a probability measure on S , S ). � For each A ∈ S , the map x → p ( x , A ) is a measurable function. � Say that X n is a Markov chain w.r.t. F n with transition probability p if P ( X n +1 ∈ B |F n ) = p ( X n , B ). � How do we construct an infinite Markov chain? Choose p and initial distribution µ on ( S , S ). For each n < ∞ write P ( X j ∈ B j , 0 ≤ j ≤ n ) = µ ( dx 0 ) p ( x 0 , dx 1 ) · · · B 0 B 1 p ( x n − 1 , dx n ) . B n Extend to n = ∞ by Kolmogorov’s extension theorem. 4 18.175 Lecture 32

  5. Markov chains Definition, again: Say X n is a Markov chain w.r.t. F n with � � transition probability p if P ( X n +1 ∈ B |F n ) = p ( X n , B ). Construction, again: Fix initial distribution µ on ( S , S ). For � � each n < ∞ write P ( X j ∈ B j , 0 ≤ j ≤ n ) = µ ( dx 0 ) p ( x 0 , dx 1 ) · · · B 0 B 1 p ( x n − 1 , dx n ) . B n Extend to n = ∞ by Kolmogorov’s extension theorem. Notation: Extension produces probability measure P µ on � � sequence space ( S 0 , 1 ,... , S 0 , 1 ,... ). Theorem: ( X 0 , X 1 , . . . ) chosen from P µ is Markov chain. � � Theorem: If X n is any Markov chain with initial distribution � � µ and transition p , then finite dim. probabilities are as above. 18.175 Lecture 32 5

  6. Markov properties S { 0 , 1 ,... } , S { 0 , 1 ,... } Markov property: Take (Ω 0 , F ) = , and � � let P µ be Markov chain measure and θ n the shift operator on Ω 0 (shifts sequence n units to left, discarding elements shifted off the edge). If Y : Ω 0 → R is bounded and measurable then E µ ( Y ◦ θ n |F n ) = E X n Y . Strong Markov property: Can replace n with a.s. finite � � stopping time N and function Y can vary with time. Suppose that for each n , Y n : Ω n → R is measurable and | Y n | ≤ M for all n . Then E µ ( Y N ◦ θ N |F N ) = E X N Y N , where RHS means E x Y n evaluated at x = X n , n = N . 6 18.175 Lecture 32

  7. Properties Property of infinite opportunities: Suppose X n is Markov � � chain and P ( ∪ ∞ m = n +1 { X m ∈ B m }| X n ) ≥ δ > 0 on { X n ∈ A n } . Then P ( { X n ∈ A n i . o . } − { X n ∈ B n i . o . } ) = 0. Reflection principle: Symmetric random walks on R . Have � � P (sup m ≥ n S m > a ) ≤ 2 P ( S n > a ). Proof idea: Reflection picture. � � 7 18.175 Lecture 32

  8. Reversibility Measure µ called reversible if µ ( x ) p ( x , y ) = µ ( y ) p ( y , x ) for � � all x , y . Reversibility implies stationarity. Implies that amount of mass � � moving from x to y is same as amount moving from y to x . Net flow of zero along each edge. Markov chain called reversible if admits a reversible probability � � measure. Are all random walks on (undirected) graphs reversible? � � What about directed graphs? � � 8 18.175 Lecture 32

  9. Cycle theorem Kolmogorov’s cycle theorem: Suppose p is irreducible. � � Then exists reversible measure if and only if p ( x , y ) > 0 implies p ( y , x ) > 0 � � n for any loop x 0 , x 1 , . . . x n with i =1 p ( x i , x i − 1 ) > 0, we have � � n p ( x i − 1 , x i ) = 1 . � p ( x i , x i − 1 ) i =1 Useful idea to have in mind when constructing Markov chains � � with given reversible distribution, as needed in Monte Carlo Markov Chains (MCMC) applications. 9 18.175 Lecture 32

  10. Outline General setup and basic properties Recurrence and transience 10 18.175 Lecture 32

  11. Outline General setup and basic properties Recurrence and transience 11 18.175 Lecture 32

  12. Query Interesting question: If A is an infinite probability transition � � matrix on a countable state space, what does the (infinite) matrix I + A + A 2 + A 3 + . . . = ( I − A ) − 1 represent (if the sum converges)? Question: Does it describe the expected number of y hits � � when starting at x ? Is there a similar interpretation for other power series? A λ A ? How about e or e � � Related to distribution after a Poisson random number of � � steps? 12 18.175 Lecture 32

  13. Recurrence Consider probability walk from y ever returns to y . � � If it’s 1, return to y infinitely often, else don’t. Call y a � � recurrent state if we return to y infinitely often. 13 18.175 Lecture 32

  14. MIT OpenCourseWare http://ocw.mit.edu 18.175 Theory of Probability Spring 2014 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

Recommend


More recommend