i ntroduction to mobile robotics probabilistic robotics
play

I ntroduction to Mobile Robotics Probabilistic Robotics Wolfram - PowerPoint PPT Presentation

I ntroduction to Mobile Robotics Probabilistic Robotics Wolfram Burgard 1 Probabilistic Robotics Key idea: Explicit representation of uncertainty (using the calculus of probability theory) Perception = state estimation Action


  1. I ntroduction to Mobile Robotics Probabilistic Robotics Wolfram Burgard 1

  2. Probabilistic Robotics Key idea: Explicit representation of uncertainty (using the calculus of probability theory)  Perception = state estimation  Action = utility optimization 2

  3. Axiom s of Probability Theory P (A) denotes probability that proposition A is true.    3

  4. A Closer Look at Axiom 3 True A A ∧ B B B 4

  5. Using the Axiom s 5

  6. Discrete Random Variables  X denotes a random variable  X can take on a countable number of values in {x 1 , x 2 , …, x n }  P ( X=x i ) or P ( x i ) is the probability that the random variable X takes on value x i  P ( ) is called probability mass function .  E.g. 6

  7. Continuous Random Variables  X takes on values in the continuum.  p ( X=x ) or p ( x ) is a probability density function  E.g. p ( x ) x 7

  8. “ Probability Sum s up to One ” Discrete case Continuous case 8

  9. Joint and Conditional Probability  P ( X=x and Y=y ) = P ( x,y )  If X and Y are independent then P ( x,y ) = P ( x ) P ( y )  P ( x | y ) is the probability of x given y P ( x | y ) = P ( x,y ) / P ( y ) P ( x,y ) = P ( x | y ) P ( y )  If X and Y are independent then P ( x | y ) = P ( x ) 9

  10. Law of Total Probability Discrete case Continuous case 10

  11. Marginalization Discrete case Continuous case 11

  12. Bayes Form ula 12

  13. Norm alization Algorithm : 13

  14. Bayes Rule w ith Background Know ledge 14

  15. Conditional I ndependence = P ( x , y z ) P ( x | z ) P ( y | z ) = P ( x z ) P ( x | z , y )  Equivalent to = P ( y z ) P ( y | z , x ) and  But this does not necessarily mean = P ( x , y ) P ( x ) P ( y ) (independence/ marginal independence) 15

  16. Sim ple Exam ple of State Estim ation  Suppose a robot obtains measurement z  What is P ( open | z ) ? 16

  17. Causal vs. Diagnostic Reasoning  P ( open|z ) is diagnostic  P ( z|open ) is causal  In some situations, causal knowledge is easier to obtain count frequencies!  Bayes rule allows us to use causal knowledge: P ( z | open ) P ( open ) = P ( open | z ) P ( z ) 17

  18. Exam ple P ( z| ¬ open ) = 0.3  P ( z|open ) = 0.6  P ( open ) = P ( ¬ open ) = 0.5 P ( z | open ) P ( open ) = P ( open | z ) + ¬ ¬ P ( z | open ) p ( open ) P ( z | open ) p ( open ) ⋅ 0 . 6 0 . 5 0 . 3 = = = P ( open | z ) 0 . 67 ⋅ + ⋅ + 0 . 6 0 . 5 0 . 3 0 . 5 0 . 3 0 . 15  z raises the probability that the door is open 18

  19. Com bining Evidence  Suppose our robot obtains another observation z 2  How can we integrate this new information?  More generally, how can we estimate P ( x | z 1 , ..., z n ) ? 19

  20. Recursive Bayesian Updating Markov assum ption: z n is independent of z 1 ,...,z n-1 if we know x 20

  21. Exam ple: Second Measurem ent P ( z 2 | ¬ open ) = 0.3  P ( z 2 |open ) = 0.25  P ( open|z 1 ) =2/3 • z 2 lowers the probability that the door is open 21

  22. Actions  Often the world is dynam ic since  actions carried out by the robot ,  actions carried out by other agents ,  or just the tim e passing by change the world  How can we incorporate such actions ? 23

  23. Typical Actions  The robot turns its w heels to move  The robot uses its m anipulator to grasp an object  Plants grow over tim e …  Actions are never carried out w ith absolute certainty  In contrast to measurements, actions generally increase the uncertainty 24

  24. Modeling Actions  To incorporate the outcome of an action u into the current “ belief ” , we use the conditional pdf P ( x | u, x’ )  This term specifies the pdf that executing u changes the state from x’ to x . 25

  25. Exam ple: Closing the door 26

  26. State Transitions P ( x | u, x’ ) for u = “ close door ” : 0.9 0.1 open closed 1 0 If the door is open, the action “close door ” succeeds in 90% of all cases 27

  27. I ntegrating the Outcom e of Actions Continuous case: Discrete case: We will make an independence assumption to get rid of the u in the second factor in the sum. 28

  28. Exam ple: The Resulting Belief 29

  29. Bayes Filters: Fram ew ork  Given:  Stream of observations z and action data u:  Sensor model P ( z | x )  Action model P ( x | u, x’ )  Prior probability of the system state P ( x )  W anted:  Estimate of the state X of a dynamical system  The posterior of the state is also called Belief : 30

  30. Markov Assum ption Underlying Assum ptions  Static world  Independent noise  Perfect model, no approximation errors 31

  31. z = observation u = action Bayes Filters x = state Bayes Markov Total prob. Markov Markov 32

  32. ∫ = η Bayes Filter Algorithm Bel ( x ) P ( z | x ) P ( x | u , x ) Bel ( x ) dx − − − t t t t t t 1 t 1 t 1 Algorithm Bayes_ filter ( Bel(x) , d ): 1. η = 0 2. 3. If d is a perceptual data item z then 4. For all x do 5. 6. 7. For all x do 8. 9. Else if d is an action data item u then For all x do 10. 11. Return Bel ’ (x) 12. 33

  33. Bayes Filters are Fam iliar! ∫ = η Bel ( x ) P ( z | x ) P ( x | u , x ) Bel ( x ) dx − − − t t t t t t 1 t 1 t 1  Kalman filters  Particle filters  Hidden Markov models  Dynamic Bayesian networks  Partially Observable Markov Decision Processes (POMDPs) 34

  34. Probabilistic Localization

  35. Probabilistic Localization

  36. Sum m ary  Bayes rule allows us to compute probabilities that are hard to assess otherwise.  Under the Markov assumption, recursive Bayesian updating can be used to efficiently combine evidence.  Bayes filters are a probabilistic tool for estimating the state of dynamic systems. 37

Recommend


More recommend