markov models and hidden markov models
play

Markov Models and Hidden Markov Models Robert Platt Northeastern - PowerPoint PPT Presentation

Markov Models and Hidden Markov Models Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. RN, AIMA Markov Models We have already seen that an MDP provides a useful framework for modeling


  1. Markov Models and Hidden Markov Models Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. RN, AIMA

  2. Markov Models We have already seen that an MDP provides a useful framework for modeling stochastic control problems. Markov Models: model any kind of temporally dynamic system.

  3. Probability again: Independence Two random variables, x and y , are independent when: The outcomes of two different coin flips are usually independent of each other Image: Berkeley CS188 course notes (downloaded Summer 2015)

  4. Probability again: Independence If: Then: Why?

  5. Are T and W independent? T W P hot sun 0.4 hot rain 0.1 cold sun 0.2 cold rain 0.3 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  6. Are T and W independent? T P hot 0.5 cold 0.5 T W P hot sun 0.4 hot rain 0.1 cold sun 0.2 cold rain 0.3 W P sun 0.6 rain 0.4 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  7. Are T and W independent? T P hot 0.5 cold 0.5 T W P T W P hot sun 0.4 hot sun 0.3 hot rain 0.1 hot rain 0.2 cold sun 0.2 cold sun 0.3 cold rain 0.3 cold rain 0.2 W P sun 0.6 rain 0.4 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  8. Conditional independence Independence: Conditional independence: Equivalent statements of conditional independence: Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  9. Conditional independence: example cavity toothache catch P(toothache, catch | cavity) = P(toothache | cavity) = P(catch | cavity) or... P(toothache | cavity) = P(toothache | cavity, catch) P(catch | cavity) = P(catch | cavity, toothache)

  10. Conditional independence: example  What about this domain:  T raffjc  Umbrella  Raining Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  11. Conditional independence: example  What about this domain:  Fire  Smoke  Alarm Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  12. Markov Processes State at time=1 State at time=2 transitions

  13. Markov Processes State at time=1 State at time=2 transitions Since this is a Markov process, we assume transitions are Markov: Process model: Markov assumption:

  14. Markov Processes How do we calculate:

  15. Markov Processes How do we calculate:

  16. Markov Processes How do we calculate:

  17. Markov Processes How do we calculate: Can we simplify this expression?

  18. Markov Processes How do we calculate:

  19. Markov Processes How do we calculate:

  20. Markov Processes How do we calculate: In general:

  21. Markov Processes How do we calculate: Process model In general:

  22. Markov Processes: example  States: X = {rain, sun}  Initial distribution: 1.0 sun  Process model: P(X t | X t-1 ): T wo new ways of representing the same CPT X t-1 X t P(X t |X t-1 ) sun sun 0.9 0.9 sun sun 0.3 0.1 sun rain 0.1 0.9 0.3 rain sun rain rain rain sun 0.3 0.7 0.7 rain rain 0.7 0.1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  23. Simulating dynamics forward Joint distribution: But, suppose we want to predict the state at time T, given a prior distribution at time 1? ...

  24. Markov Processes: example 0.3 0.9  Initial distribution: 1.0 sun rain sun 0.7 0.1  What is the probability distribution after one step? Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  25. Simulating dynamics forward  From initial observation of sun P( X 1 ) P( X 2 ) P( X 3 ) P( X 4 ) P( X ∞ )  From initial observation of rain P( X 1 ) P( X 2 ) P( X 3 ) P( X 4 ) P( X ∞ )  From yet another initial distribution P(X 1 ): … P( X 1 ) P( X ∞ ) Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  26. Simulating dynamics forward  From initial observation of sun P( X 1 ) P( X 2 ) P( X 3 ) P( X 4 ) P( X ∞ )  From initial observation of rain This is called the stationary distribution P( X 1 ) P( X 2 ) P( X 3 ) P( X 4 ) P( X ∞ )  From yet another initial distribution P(X 1 ): … P( X 1 ) P( X ∞ ) Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  27. Hidden Markov Models (HMMs) Hidden Markov Models: markov models applied to estimation problems – speech to text – tracking in computer vision – robot localization

  28. Hidden Markov Models (HMMs) Called an “emission” State, , is assumed to be unobserved However, you get to make one observation, , on each timestep.

  29. Hidden Markov Models (HMMs) Sensor Markov Assumption: the current observation depends only on current state:

  30. HMM example Rain t-1 Rain t Rain t+1 Umbrella Umbrella Umbrella R t R t+1 P(R t+1 |R t ) t-1 t t+1 +r +r 0.7 +r -r 0.3 -r +r 0.3 -r -r 0.7  An HMM is defjned by: R t U t P(U t |R t )  Initial distribution: +r +u 0.9  T ransitions: +r -u 0.1  Emissions: -r +u 0.2 -r -u 0.8 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  31. Real world HMM applications  Speech recognition HMMs:  Observations are acoustic signals (continuous valued)  States are specifjc positions in specifjc words (so, tens of thousands)  Machine translation HMMs:  Observations are words (tens of thousands)  States are translation options  Robot tracking:  Observations are range readings (continuous)  States are positions on a map (continuous) Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  32. HMM Filtering  Filtering, or monitoring, is the task of tracking the distribution B t (X) = P t (X t | e 1 , …, e t ) (the belief state) over time  We start with B 1 (X) in an initial setting, usually uniform  As time passes, or we get observations, we update B(X)  The Kalman fjlter was invented in the 60’s and fjrst implemented as a method of trajectory estimation for the Apollo program Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  33. HMM Filtering Given a prior distribution, , and a series of observations, , calculate the posterior distribution: Two steps: Process update Observation update

  34. HMM Filtering Given a prior distribution, , and a series of observations, , calculate the posterior distribution: Two steps: Process update Observation update

  35. HMM Filtering Given a prior distribution, , and a series of observations, , calculate the posterior distribution: “Beliefs” Two steps: Process update Observation update

  36. Process update This is just forward simulation of the Markov Model

  37. Process update: example  As time passes, uncertainty “accumulates” (T ransition model: ghosts usually go clockwise) T = 1 T = 2 T = 5 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  38. Observation update Where is a normalization factor

  39. Observation update  As we get observations, beliefs get reweighted, uncertainty “decreases” Before observation After observation Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  40. Robot localization example Prob 0 1 Observation model: can read in which directions there is a wall, never more than 1 mistake Process model: may not execute action with small prob. Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  41. Robot localization example Prob 0 1 Lighter grey: was possible to get the reading, but less likely b/c required 1 mistake Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  42. Robot localization example Prob 0 1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  43. Robot localization example Prob 0 1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  44. Robot localization example Prob 0 1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  45. Robot localization example Prob 0 1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  46. Weather HMM example R t R t+1 P(R t+1 |R t ) +r +r 0.7 +r -r 0.3 -r +r 0.3 B(+r) = 0.5 -r -r 0.7 B(-r) = 0.5 Rain 0 Rain 1 Rain 2 R t U t P(U t |R t ) +r +u 0.9 +r -u 0.1 -r +u 0.2 Umbrella 1 Umbrella 2 -r -u 0.8 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  47. Weather HMM example R t R t+1 P(R t+1 |R t ) +r +r 0.7 B’(+r) = 0.5 B’(-r) = 0.5 +r -r 0.3 -r +r 0.3 -r -r 0.7 B(+r) = 0.5 B(-r) = 0.5 Rain 0 Rain 1 Rain 2 R t U t P(U t |R t ) +r +u 0.9 +r -u 0.1 -r +u 0.2 Umbrella 1 Umbrella 2 -r -u 0.8 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  48. Weather HMM example R t R t+1 P(R t+1 |R t ) +r +r 0.7 B’(+r) = 0.5 B’(-r) = 0.5 +r -r 0.3 -r +r 0.3 -r -r 0.7 B(+r) = 0.818 B(+r) = 0.5 B(-r) = 0.182 B(-r) = 0.5 Rain 0 Rain 1 Rain 2 R t U t P(U t |R t ) +r +u 0.9 +r -u 0.1 -r +u 0.2 Umbrella 1 Umbrella 2 -r -u 0.8 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

Recommend


More recommend