determinism vs randomness
play

Determinism vs randomness There are two main kinds of processes in - PowerPoint PPT Presentation

Determinism vs randomness There are two main kinds of processes in Nature, distinguished by their time evolution. Mathematics for Informatics 4a In a deterministic process, the future state of the system is completely determined by the present


  1. Determinism vs randomness There are two main kinds of processes in Nature, distinguished by their time evolution. Mathematics for Informatics 4a In a deterministic process, the future state of the system is completely determined by the present state. José Figueroa-O’Farrill Physical systems whose time evolution is described by differential equations are deterministic; e.g., classical mechanics (Newton’s equation) quantum mechanics (Schrödinger’s equation) the weather (chaotic but deterministic!) Stochastic (or random ) processes are non-deterministic: the time evolution is subject to a probability distribution. Examples of stochastic processes are Lecture 15 Random walks 14 March 2012 Markov chains Birth-death processes Queues These are the subject of the last part of this course. José Figueroa-O’Farrill mi4a (Probability) Lecture 15 1 / 20 José Figueroa-O’Farrill mi4a (Probability) Lecture 15 2 / 20 Stochastic processes Markov chains Let ( Ω , F , P ) be a probability space. We assume that S is countable so that the X t are discrete random variables. Let S be a set called the state space of the system. The set S can be countable or uncountable. We will also assume that we have a discrete-time process, so that T = { 0, 1, 2, . . . } . Let T be an index set, to be thought of as “time”. It can be continuous or discrete . Definition A stochastic (or random ) process with state space S is a A stochastic process X = { X 0 , X 1 , X 2 , . . . } is a Markov chain if it collection of random variables X t : Ω → S indexed by t ∈ T . satisfies the Markov property : The interpretation is that X t is the state of the system at time t , which for a non-deterministic system is a random P ( X n + 1 = s | X 0 = s 0 , . . . , X n = s n ) = P ( X n + 1 = s | X n = s n ) variable with some probability distribition. for all n � 0 and s 0 , s 1 , . . . , s n , s ∈ S . There are many kinds of stochastic processes, differing in how the probability of X t being in a given state depends on the history of the system; that is, in which state the system “given the present, the future does not depend on the past” was in times t ′ < t . José Figueroa-O’Farrill mi4a (Probability) Lecture 15 3 / 20 José Figueroa-O’Farrill mi4a (Probability) Lecture 15 4 / 20

  2. Random walks Proposition Consider a particle moving on the integer lattice in R : The sequence { X 0 , X 1 , X 2 , . . . } exhibits spatial homogeneity : q p P ( X n = j | X 0 = a ) = P ( X n = j + b | X 0 = a + b ) i − 1 i i + 1 Proof. Therefore S = Z and J i are independent random variables with � n � � P ( J i = 1 ) = p P ( J i = − 1 ) = q = 1 − p P ( X n = j | X 0 = a ) = P J i = j − a i = 1 and also Let X n denote the position of the particle at time n , so that � n � � n P ( X n = j + b | X 0 = a + b ) = P J i = j + b − ( a + b ) = j − a � X n = X 0 + J i i = 1 i = 1 José Figueroa-O’Farrill mi4a (Probability) Lecture 15 5 / 20 José Figueroa-O’Farrill mi4a (Probability) Lecture 15 6 / 20 Proposition Proposition The sequence { X 0 , X 1 , X 2 , . . . } exhibits temporal homogeneity : The sequence { X 0 , X 1 , X 2 , . . . } exhibits the Markov property : P ( X m + n = j | X 0 = i 0 , . . . , X m = i m ) = P ( X m + n = j | X m = i m ) P ( X n = j | X 0 = a ) = P ( X n + m = j | X m = a ) Proof. Proof. � n This follows because � � P ( X n = j | X 0 = a ) = P J i = j − a n � i = 1 X m + n = X m + J i whereas i = m + 1   m + n so X m + n does not depend explicitly on the X j for j < m . � P ( X n + m = j | X m = a ) = P J i = j − a   i = m + 1 but the J i are i.i.d. José Figueroa-O’Farrill mi4a (Probability) Lecture 15 7 / 20 José Figueroa-O’Farrill mi4a (Probability) Lecture 15 8 / 20

  3. Example (Gambler’s ruin) Example (Gambler’s ruin – continued) A gambler starts with £ k and plays a game in which a fair coin Letting p k = P k ( R ) , we have the following difference equation: is tossed repeatedly: winning £1 if heads and − £1 if tails. The p k = 1 p 0 = 1 p N = 0 2 ( p k + 1 + p k − 1 ) game stops when the gambler’s fortune is either £ N ( N > k ) or £0. What is the probability that the gambler is ultimately ruined? Let a k = p k − p k − 1 . Then This is an example of a random walk on a finite set { 0, 1, 2, . . . , N } . Let R denote the event that the gambler is a k − a k − 1 = p k − p k − 1 − ( p k − 1 − p k − 2 ) eventually ruined and let H and T denote the events that the = p k − 2 p k − 1 + p k − 2 first toss is heads and tails, respectively. Let P k ( R ) denote the = p k − ( p k + p k − 2 ) + p k − 2 = 0 probability that gambler is eventually ruined starting with £ k . Then Therefore a k = a 1 for all k and hence P k ( R ) = P k ( R | H ) P ( H ) + P k ( R | T ) P ( T ) but clearly P k ( R | H ) = P k + 1 ( R ) and P k ( R | T ) = P k − 1 ( R ) , whence p k = a 1 + p k − 1 = 2 a 1 + p k − 2 = · · · = ka 1 + p 0 P k ( R ) = 1 2 P k + 1 ( R ) + 1 Since p 0 = 1 and p N = 0, we find a 1 = − 1 N , whence p k = 1 − k N . 2 P k − 1 ( R ) José Figueroa-O’Farrill mi4a (Probability) Lecture 15 9 / 20 José Figueroa-O’Farrill mi4a (Probability) Lecture 15 10 / 20 Example (Gambler’s ruin – continued) Example (Gambler’s ruin – continued) What about if the coin is not fair? Imposing the boundary conditions Let P ( H ) = p and P ( T ) = q = 1 − p , with p � = q . Now � N � q 1 = p 0 = c 1 + c 2 0 = p N = c 1 + c 2 p p k = pp k + 1 + qp k − 1 1 � k � N − 1 whence with the same boundary conditions p 0 = 1 and p N = 0. Try a 1 � N � q c 1 = − c 2 c 2 = solution p k = θ k for some θ . Then p � N � q 1 − p θ k = pθ k + 1 + qθ k − 1 = ⇒ pθ 2 − θ + q = 0 and hence with roots θ 1 = 1 and θ 2 = q p . The general solution is then � N � k � k � N � � � � q q q q − p p p p p k = − � N + � N = p k = c 1 θ k 1 + c 2 θ k � N � � � q q q 2 1 − 1 − 1 − p p p for some c 1 , c 2 which are determined by p 0 = 1 and p N = 0. José Figueroa-O’Farrill mi4a (Probability) Lecture 15 11 / 20 José Figueroa-O’Farrill mi4a (Probability) Lecture 15 12 / 20

  4. Transition matrix Theorem Let’s go back to the case of a general Markov chain The transition matrix of a Markov chain is stochastic ; that is, { X 0 , X 1 , X 2 , . . . } . p ij � 0 1 Since S is countable we will assume it is a subset of Z . � j p ij = 1 for all i (i.e., rows sum to 1 ) 2 The evolution of a Markov chain is described by its transition probabilities Proof. P ( X n + 1 = j | X n = i ) This is obvious since the p ij are probabilities. 1 We will make the additional assumption of temporal 2 � � homogeneity: p ij = P ( X n + 1 = j | X n = i ) = 1 j j P ( X n + 1 = j | X n = i ) = P ( X 1 = j | X 0 = i ) since X n + 1 must take some value. Therefore the transition probabilities are encoded in a transition matrix P = ( p ij ) , where p ij = P ( X n + 1 = j | X n = i ) José Figueroa-O’Farrill mi4a (Probability) Lecture 15 13 / 20 José Figueroa-O’Farrill mi4a (Probability) Lecture 15 14 / 20 Example Example (Continued) Let X n denote the state of a computer at the start of the n th We often represent Markov chains graphically; e.g., day. The computer can be in either of two states: X n = 0 if it is q broken or X n = 1 if in working order. Let π n ( 0 ) = P ( X n = 0 ) and π n ( 1 ) = P ( X n = 1 ) = 1 − π n ( 0 ) . 1 − p 0 1 1 − q Let the transition probabilities be p P ( X n + 1 = 1 | X n = 0 ) = p P ( X n + 1 = 0 | X n = 0 ) = 1 − p which allows us to read the transition probabilities at a glance P ( X n + 1 = 0 | X n = 1 ) = q P ( X n + 1 = 1 | X n = 1 ) = 1 − q and write down the transition matrix: (Notice that p + q need not equal 1!) � 0 → 0 � � 1 − p � 0 → 1 p Therefore the transition matrix is P = = 1 → 0 1 → 1 1 − q q � 1 − p � p P = 1 − q q A typical question is: What is P ( X n + 1 = 0 ) ? We will answer this naively at first. José Figueroa-O’Farrill mi4a (Probability) Lecture 15 15 / 20 José Figueroa-O’Farrill mi4a (Probability) Lecture 15 16 / 20

Recommend


More recommend