course overview and introduction
play

Course Overview and Introduction Probabilistic Graphical Models - PowerPoint PPT Presentation

Course Overview and Introduction Probabilistic Graphical Models Sharif University of Technology Soleymani Spring 2018 Some slides have been adopted from Eric Zing, CMU Course info Instructor: Mahdieh Soleymani Email:


  1. Course Overview and Introduction Probabilistic Graphical Models Sharif University of Technology Soleymani Spring 2018 Some slides have been adopted from Eric Zing, CMU

  2. Course info  Instructor: Mahdieh Soleymani  Email: soleymani@sharif.edu  Teacher assistants:  Fatemeh Seyyedsalehi  Ehsan Montahaei  Amirshayan Haghipour  Seyed Mohammad Chavosian  Sajad Behfar 2

  3. Text book  D. Koller and N. Friedman, “ Probabilistic Graphical Models: Principles and Techniques ” , MIT Press, 2009.  M.I. Jordan, “ An Introduction to Probabilistic Graphical Models ” , Preprint .  Other C.M. Bishop, “ Pattern Recognition and Machine Learning ” , Springer, 2006.  Chapters 8-11,13.  K.P. Murphy, “ Machine Learning:A Probabilistic Perspective ” , MIT Press, 2012.  3

  4. Evaluation policy  Mid-term: 20%  Final: 30%  Mini-exams: 10%  Home works & project: 40% 4

  5. Why using probabilistic models?  Partial knowledge of the state of the world  Noisy or incomplete observations  We may not know or cover all the involved phenomena in our model  Partial knowledge can cause the world seems to be stochastic  To deal with partial knowledge and/or stochastic worlds we need reasoning under uncertainty 5

  6. Reasoning under uncertainty 6

  7. Representation, inference, and learning  We will cover three aspects of the probabilistic models:  Representation of probabilistic knowledge  Inference algorithms on these models  Learning: Using the data to acquire the distribution 7

  8. Representation, inference, and learning  Representation: When variables tends to interact directly with few other variables (local structure)  Inference: answering queries using the model  algorithms for answering questions/queries according to the model and/or based given observation.  Learning of both the parameters and the structure of the graphical models 8

  9. Main problems  Representation : what is the joint probability dist. on multiple variables? 𝑄(𝑌 1 , … , 𝑌 𝑒 )  How many state configurations in total?  Are they all needed to be represented?  Do we get any scientific/medical insight?  Inference : If not all variables are observable, how to compute the conditional distribution of latent variables given evidence?  Learning : where do we get all this probabilities?  Maximal-likelihood estimation? but how many data do we need?  Are there other est. principles?  Where do we put domain knowledge in terms of plausible relationships between variables, and plausible values of the probabilities? 9

  10. Probability review  Marginal probabilities  𝑄 𝑌 = 𝑧 𝑄(𝑌, 𝑍 = 𝑧)  Conditional probabilities 𝑄(𝑌,𝑍)  𝑄 𝑌|𝑍 = 𝑄(𝑍)  Bayes rule: 𝑄 𝑍|𝑌 𝑄(𝑌)  𝑄 𝑌|𝑍 = 𝑄(𝑍)  Chain rule: 𝑜  𝑄 𝑌 1 , … , 𝑌 𝑜 = 𝑗=1 𝑄(𝑌 𝑗 |𝑌 1 , … , 𝑌 𝑗−1 ) 10

  11. Medical diagnosis example  Representation 𝑒 1 𝑒 2 𝑒 3 𝑒 4 diseases Findings 𝑔 𝑔 𝑔 𝑔 𝑔 5 1 2 3 4 (symptoms & tests) 𝑄(𝑔 1 |𝑒 1 ) 𝑄(𝑔 2 |𝑒 1 , 𝑒 2 , 𝑒 3 ) 𝑄(𝑔 3 |𝑒 3 ) … 11

  12. Example Graphical Model 12

  13. Representation: summary of advantages  Representing large multivariate distributions directly and exhaustively is hopeless:  The number of parameters is exponential in the number of random variables  Inference can be exponential in the number of variables  PGM representation  Compact representation of the joint distribution  Transparent  We can combine expert knowledge and accumulated data to learn the model  Effective for inference and learning 13

  14. Why using a graph for representation?  Intuitively appealing interface by which we can models highly interacting sets of variables  It allows us to design efficient general purpose inference algorithms 14

  15. Graph structure  Denotes conditional dependence structure between random variables  One view: Graph represents a set of independencies  Another view: Graph shows a skeleton for factorizing a joint distribution 15

  16. PGMs as a framework  General-purpose framework for representing uncertain knowledge and learning and inference in uncertain conditions .  A graph-based representation as the basis of encoding a complex distribution compactly  allows declarative representation (with clear semantics) of the probabilistic knowledge 16

  17. PGMs as a framework  Intuitive & compact data structure for representation  Efficient reasoning using general-purpose algorithms  Sparse parameterization (enables us to elicit or learn from data) 17

  18. PGM: declarative representation  Separation of knowledge and reasoning  We need to specify our model for a specific application that represents our probabilistic knowledge  There is a general suite of reasoning algorithms that can be used. 18

  19. Data Integration  Due to the local structure of the graph, we can handle different source of information  Modular combination of heterogeneous parts – data fusion  Combining different data modality  Example:T ext + Image + Network => Holistic Social Media 19

  20. History  Wright 1921, 1934 and before  Bayesian networks are independently developed by Spiegelhalter and Lauritzen in statistics and Pearl in computer science in the late 1980 ’ s  First applications (1990 ’ s): expert systems and information retrieval 20

  21. PGMs: some application areas  Machine Learning and computational statistics  Computer vision: e.g., segmenting and denoising images  Robotics: e.g., robot localization and mapping  Natural Language Processing  Speech recognition  Information Retrieval  AI: game playing, planning  Computational Biology  Networks: decoding messages (sent over a noisy channel)  Medical diagnosis and prognosis  … 21

  22. Graphical models: directed & undirected  Two kinds of graphical models:  Directed: Bayesian Networks (BNs)  Undirected: Markov Random Fields (MRFs) B A A B C C D D Causality relations Correlation of variables 22

  23. Graphical models: directed & undirected [Pathfinder Project, 1992] 23

  24. Medical diagnosis example  Representation 𝑒 1 𝑒 2 𝑒 3 𝑒 4 diseases Findings 𝑔 𝑔 𝑔 𝑔 𝑔 5 1 2 3 4 (symptoms & tests)  Inference: Given symptoms, what disease is likely?  Eliciting or learning the required probabilities from the data 24

  25. Image denoising example 25 [Bishop]

  26. Genetic pedigree example A0 B0 A B Bg Ag A1 B1 D C C0 D0 Cg Dg C1 D1 E E0 Eg E1 26

  27. Plan in our course  Fundamentals of Graphical Models:  Representation  Bayesian Network  Markov Random Fields  Exact inference  Basics of learning  Case studies: Popular graphical models  Multivariate Gaussian Models  FA, PPCA  HMM, CRF, Kalman filter  Approximate inference  Variational methods  Monte Carlo algorithms 27

Recommend


More recommend