csci 8980 advanced topics in graphical models mixture
play

CSci 8980: Advanced Topics in Graphical Models Mixture Models, EM, - PowerPoint PPT Presentation

Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net CSci 8980: Advanced Topics in Graphical Models Mixture Models, EM, Exponential Families Instructor: Arindam Banerjee September


  1. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net CSci 8980: Advanced Topics in Graphical Models Mixture Models, EM, Exponential Families Instructor: Arindam Banerjee September 11, 2007

  2. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM p ( Z ) = � Since z i are independent, optimal ˜ i ˜ p ( z i )

  3. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM p ( Z ) = � Since z i are independent, optimal ˜ i ˜ p ( z i ) Sufficient to work with such ˜ p in F (˜ p , θ )

  4. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM p ( Z ) = � Since z i are independent, optimal ˜ i ˜ p ( z i ) Sufficient to work with such ˜ p in F (˜ p , θ ) Then F (˜ p , θ ) = � i F i (˜ p i , θ ) where F i (˜ p i , θ ) = E ˜ p i [log p ( x i , z i | θ )] + H (˜ p i )

  5. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM p ( Z ) = � Since z i are independent, optimal ˜ i ˜ p ( z i ) Sufficient to work with such ˜ p in F (˜ p , θ ) Then F (˜ p , θ ) = � i F i (˜ p i , θ ) where F i (˜ p i , θ ) = E ˜ p i [log p ( x i , z i | θ )] + H (˜ p i ) Incremental algorithm that works one point at a time

  6. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM (Contd.) Basic Incremental EM

  7. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM (Contd.) Basic Incremental EM E-step: Choose a data item i to be updated p ( t ) p ( t − 1) Set ˜ = ˜ for j � = i j j p ( t ) = p ( z i | x i , θ ( t ) ) Set ˜ i

  8. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM (Contd.) Basic Incremental EM E-step: Choose a data item i to be updated p ( t ) p ( t − 1) Set ˜ = ˜ for j � = i j j p ( t ) = p ( z i | x i , θ ( t ) ) Set ˜ i M-step: Set θ ( t ) to argmax θ E ˜ p ( t ) [log p ( x , z | θ )]

  9. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM (Contd.) Basic Incremental EM E-step: Choose a data item i to be updated p ( t ) p ( t − 1) Set ˜ = ˜ for j � = i j j p ( t ) = p ( z i | x i , θ ( t ) ) Set ˜ i M-step: Set θ ( t ) to argmax θ E ˜ p ( t ) [log p ( x , z | θ )] M-step needs to look at all components of ˜ p

  10. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM (Contd.) Basic Incremental EM E-step: Choose a data item i to be updated p ( t ) p ( t − 1) Set ˜ = ˜ for j � = i j j p ( t ) = p ( z i | x i , θ ( t ) ) Set ˜ i M-step: Set θ ( t ) to argmax θ E ˜ p ( t ) [log p ( x , z | θ )] M-step needs to look at all components of ˜ p Can be simplified by using sufficient statistics

  11. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM (Contd.) Basic Incremental EM E-step: Choose a data item i to be updated p ( t ) p ( t − 1) Set ˜ = ˜ for j � = i j j p ( t ) = p ( z i | x i , θ ( t ) ) Set ˜ i M-step: Set θ ( t ) to argmax θ E ˜ p ( t ) [log p ( x , z | θ )] M-step needs to look at all components of ˜ p Can be simplified by using sufficient statistics For a distribution p ( x | θ ), s ( x ) is a sufficient statistic if p ( x | s ( x ) , θ ) = p ( x | s ( x )) = ⇒ p ( x | θ ) = h ( x ) q ( s ( x ) | θ )

  12. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM with Sufficient Statistics EM with sufficient statistics

  13. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM with Sufficient Statistics EM with sufficient statistics s ( t ) = E ˜ p ( z ) = p ( z | x , θ ( t − 1) ) E-step: Set ˜ p [ s ( x , z )] where ˜

  14. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM with Sufficient Statistics EM with sufficient statistics s ( t ) = E ˜ p ( z ) = p ( z | x , θ ( t − 1) ) E-step: Set ˜ p [ s ( x , z )] where ˜ M-step: Set θ ( t ) to θ , the max likelihood given ˜ s ( t )

  15. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM with Sufficient Statistics EM with sufficient statistics s ( t ) = E ˜ p ( z ) = p ( z | x , θ ( t − 1) ) E-step: Set ˜ p [ s ( x , z )] where ˜ M-step: Set θ ( t ) to θ , the max likelihood given ˜ s ( t ) Incremental EM with sufficient statistics

  16. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM with Sufficient Statistics EM with sufficient statistics s ( t ) = E ˜ p ( z ) = p ( z | x , θ ( t − 1) ) E-step: Set ˜ p [ s ( x , z )] where ˜ M-step: Set θ ( t ) to θ , the max likelihood given ˜ s ( t ) Incremental EM with sufficient statistics E-step: Choose a data item i to be updated s ( t ) s ( t − 1) Set ˜ = ˜ , for j � = i j j s ( t ) p i ( z i ) = p ( z i | x i , θ ( t − 1) ) Set ˜ = E ˜ p i [ s i ( x i , z i )], where ˜ i s ( t ) = ˜ s ( t − 1) − ˜ s ( t − 1) s ( t ) Set ˜ + ˜ i i

  17. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Incremental EM with Sufficient Statistics EM with sufficient statistics s ( t ) = E ˜ p ( z ) = p ( z | x , θ ( t − 1) ) E-step: Set ˜ p [ s ( x , z )] where ˜ M-step: Set θ ( t ) to θ , the max likelihood given ˜ s ( t ) Incremental EM with sufficient statistics E-step: Choose a data item i to be updated s ( t ) s ( t − 1) Set ˜ = ˜ , for j � = i j j s ( t ) p i ( z i ) = p ( z i | x i , θ ( t − 1) ) Set ˜ = E ˜ p i [ s i ( x i , z i )], where ˜ i s ( t ) = ˜ s ( t − 1) − ˜ s ( t − 1) s ( t ) Set ˜ + ˜ i i M-step: Set θ ( t ) to θ , the max likelihood given ˜ s ( t )

  18. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Example Consider a mixture of 2 univariate Gaussians

  19. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Example Consider a mixture of 2 univariate Gaussians Parameter set θ = ( α, µ 1 , σ 1 , µ 2 , σ 2 )

  20. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Example Consider a mixture of 2 univariate Gaussians Parameter set θ = ( α, µ 1 , σ 1 , µ 2 , σ 2 ) Sufficient statistics s i ( x i , z i ) = [ z i (1 − z i ) z i x i (1 − z i ) x i z i x 2 i (1 − z i ) x 2 i ]

  21. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Example Consider a mixture of 2 univariate Gaussians Parameter set θ = ( α, µ 1 , σ 1 , µ 2 , σ 2 ) Sufficient statistics s i ( x i , z i ) = [ z i (1 − z i ) z i x i (1 − z i ) x i z i x 2 i (1 − z i ) x 2 i ] Given s ( x , z ) = � i s ( x i , z i ) = ( n 1 , n 2 , m 1 , m 2 , q 1 , q 2 ) � 2 n 1 µ h = m h h = q h � m h σ 2 α = , , − n 1 + n 2 n h n h n h

  22. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Sparse EM Consider a mixture model with many components

  23. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Sparse EM Consider a mixture model with many components Most p ( z | x , θ ) will be negligibly small

  24. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Sparse EM Consider a mixture model with many components Most p ( z | x , θ ) will be negligibly small Computation can be saved by freezing these

  25. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Sparse EM Consider a mixture model with many components Most p ( z | x , θ ) will be negligibly small Computation can be saved by freezing these Only a small set of component posteriors need to be updated � q ( t ) z , if z �∈ S t p ( t ) ( z ) = ˜ Q ( t ) r ( t ) if z ∈ S t z

  26. Variants Auxiliary Functions Mixture of Gaussians Exponential Family Mixtures of Exponential Families Bayes Net Sparse EM Consider a mixture model with many components Most p ( z | x , θ ) will be negligibly small Computation can be saved by freezing these Only a small set of component posteriors need to be updated � q ( t ) z , if z �∈ S t p ( t ) ( z ) = ˜ Q ( t ) r ( t ) if z ∈ S t z S t = set of plausible values

Recommend


More recommend