variational hamiltonian monte carlo via score matching
play

Variational Hamiltonian Monte Carlo via Score Matching Cheng Zhang - PowerPoint PPT Presentation

Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Rfrences Variational Hamiltonian Monte Carlo via Score Matching Cheng Zhang (Joint work with Prof. Shahbaba and Prof. Zhao)


  1. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Variational Hamiltonian Monte Carlo via Score Matching Cheng Zhang (Joint work with Prof. Shahbaba and Prof. Zhao) Department of Mathematics University of California, Irvine Jan 6, 2017 Cheng Zhang UCI Variational HMC Jan 6, 2017 1 / 27

  2. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Outline Background 1 Bayesian Inference Markov chain Monte Carlo 2 Metropolis-Hastings Algorithm Hamiltonian Monte Carlo Scalable MCMC Fixed-Form Variational Bayes 3 Lower Bounds and Free Energy Variational Bayes as Linear Regression Variational Hamiltonian Monte Carlo 4 Approximation with Random Bases Variational HMC Experiments Conclusion 5 Cheng Zhang UCI Variational HMC Jan 6, 2017 2 / 27

  3. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Bayesian Inference Bayesian Inference Bayesian inference model D = { y 1 , . . . , y N } : observed data θ ∈ R d : model parameter p ( D| θ ) : model density p ( θ ) : prior Cheng Zhang UCI Variational HMC Jan 6, 2017 3 / 27

  4. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Bayesian Inference Bayesian Inference Bayesian inference model D = { y 1 , . . . , y N } : observed data θ ∈ R d : model parameter p ( D| θ ) : model density p ( θ ) : prior Goal : learning parameter θ from data p ( θ |D ) = p ( D| θ ) · p ( θ ) ∝ p ( D| θ ) · p ( θ ) p ( D ) Cheng Zhang UCI Variational HMC Jan 6, 2017 3 / 27

  5. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Bayesian Inference Bayesian Inference Bayesian inference model D = { y 1 , . . . , y N } : observed data θ ∈ R d : model parameter p ( D| θ ) : model density p ( θ ) : prior Goal : learning parameter θ from data p ( θ |D ) = p ( D| θ ) · p ( θ ) ∝ p ( D| θ ) · p ( θ ) p ( D ) Difficulty : p ( D ) unknown ⇒ intractable posterior distribution p ( θ |D ) e.g., probabilistic graphical models, Bayesian hierarchical models Cheng Zhang UCI Variational HMC Jan 6, 2017 3 / 27

  6. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Bayesian Inference Bayesian Inference Bayesian inference model D = { y 1 , . . . , y N } : observed data θ ∈ R d : model parameter p ( D| θ ) : model density p ( θ ) : prior Goal : learning parameter θ from data p ( θ |D ) = p ( D| θ ) · p ( θ ) ∝ p ( D| θ ) · p ( θ ) p ( D ) Difficulty : p ( D ) unknown ⇒ intractable posterior distribution p ( θ |D ) e.g., probabilistic graphical models, Bayesian hierarchical models Two popular approximations Markov chain Monte Carlo . Sample by running a Markov chain : asymptotically unbiased but computationally slow Variational Bayes . Approximate via tractable distributions : computationally fast but may result in poor approximation. Cheng Zhang UCI Variational HMC Jan 6, 2017 3 / 27

  7. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Metropolis-Hastings Algorithm Markov chain Monte Carlo Intuitive idea : evolve a Markov chain to sample from a target distribution π ( θ ) (M ETROPOLIS et al. 1953). Cheng Zhang UCI Variational HMC Jan 6, 2017 4 / 27

  8. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Metropolis-Hastings Algorithm Markov chain Monte Carlo Intuitive idea : evolve a Markov chain to sample from a target distribution π ( θ ) (M ETROPOLIS et al. 1953). Conditions for transition kernel T ( ·|· ) Irreducibility : any state has positive probability of visiting any other state. Aperiodicity : The chain should not get trapped in cycles. Detailed Balance condition (sufficient) : π ( θ ) T ( θ ′ | θ ) = π ( θ ′ ) T ( θ | θ ′ ) Cheng Zhang UCI Variational HMC Jan 6, 2017 4 / 27

  9. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Metropolis-Hastings Algorithm Markov chain Monte Carlo Intuitive idea : evolve a Markov chain to sample from a target distribution π ( θ ) (M ETROPOLIS et al. 1953). Conditions for transition kernel T ( ·|· ) Irreducibility : any state has positive probability of visiting any other state. Aperiodicity : The chain should not get trapped in cycles. Detailed Balance condition (sufficient) : π ( θ ) T ( θ ′ | θ ) = π ( θ ′ ) T ( θ | θ ′ ) Metropolis-Hastings algorithms (one iteration) sample θ ′ ∼ q ( θ ′ | θ ) 1 update the current state to θ ′ with probability α ( θ , θ ′ ) = min [ 1 , π ( θ ′ ) q ( θ | θ ′ ) π ( θ ) q ( θ ′| θ ) ] 2 Cheng Zhang UCI Variational HMC Jan 6, 2017 4 / 27

  10. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Metropolis-Hastings Algorithm Markov chain Monte Carlo Intuitive idea : evolve a Markov chain to sample from a target distribution π ( θ ) (M ETROPOLIS et al. 1953). Conditions for transition kernel T ( ·|· ) Irreducibility : any state has positive probability of visiting any other state. Aperiodicity : The chain should not get trapped in cycles. Detailed Balance condition (sufficient) : π ( θ ) T ( θ ′ | θ ) = π ( θ ′ ) T ( θ | θ ′ ) Metropolis-Hastings algorithms (one iteration) sample θ ′ ∼ q ( θ ′ | θ ) 1 update the current state to θ ′ with probability α ( θ , θ ′ ) = min [ 1 , π ( θ ′ ) q ( θ | θ ′ ) π ( θ ) q ( θ ′| θ ) ] 2 Pros & Cons for simple MCMCs (e.g., RWM and Gibbs sampling) Pro : easy to implement and computationally cheap Con : slow mixing due to random walk behaviors, especially in complicate, high-dimensional models. Cheng Zhang UCI Variational HMC Jan 6, 2017 4 / 27

  11. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Hamiltonian Monte Carlo Hamiltonian Monte Carlo Intuition : Leveraging a Hamiltonian dynamical system to generate trial moves in MCMC samplers. (D UANE et al. 1987, N EAL 2011) Cheng Zhang UCI Variational HMC Jan 6, 2017 5 / 27

  12. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Hamiltonian Monte Carlo Hamiltonian Monte Carlo Intuition : Leveraging a Hamiltonian dynamical system to generate trial moves in MCMC samplers. (D UANE et al. 1987, N EAL 2011) Model based energy function : the Hamiltonian H ( θ , r ) = U ( θ ) + K ( r ) Potential : U ( θ ) = − log p ( θ , D ) = − [ log p ( θ ) + log p ( D| θ )] Kinetic : K ( r ) = 1 2 r ⊺ M − 1 r ⇒ π ( r ) ∼ N ( 0 , M ) The joint density of z = ( θ , r ) is π ( z ) ∝ exp ( − U ( θ ) − K ( r )) ∝ p ( θ |D ) · π ( r ) Cheng Zhang UCI Variational HMC Jan 6, 2017 5 / 27

  13. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Hamiltonian Monte Carlo Hamiltonian Monte Carlo Intuition : Leveraging a Hamiltonian dynamical system to generate trial moves in MCMC samplers. (D UANE et al. 1987, N EAL 2011) Model based energy function : the Hamiltonian H ( θ , r ) = U ( θ ) + K ( r ) Potential : U ( θ ) = − log p ( θ , D ) = − [ log p ( θ ) + log p ( D| θ )] Kinetic : K ( r ) = 1 2 r ⊺ M − 1 r ⇒ π ( r ) ∼ N ( 0 , M ) The joint density of z = ( θ , r ) is π ( z ) ∝ exp ( − U ( θ ) − K ( r )) ∝ p ( θ |D ) · π ( r ) Hamilton’s equations : d θ d r dt = ∇ r H = ∇ r K ( r ) , dt = −∇ θ H = −∇ θ U ( θ ) s : R 2 d → R 2 d , z ( 0 ) = z �→ z ∗ = z ( s ) Hamiltonian flow φ H Cheng Zhang UCI Variational HMC Jan 6, 2017 5 / 27

  14. Background Markov chain Monte Carlo Fixed-Form Variational Bayes Variational Hamiltonian Monte Carlo Conclusion Références Hamiltonian Monte Carlo Hamiltonian Monte Carlo Intuition : Leveraging a Hamiltonian dynamical system to generate trial moves in MCMC samplers. (D UANE et al. 1987, N EAL 2011) Model based energy function : the Hamiltonian H ( θ , r ) = U ( θ ) + K ( r ) Potential : U ( θ ) = − log p ( θ , D ) = − [ log p ( θ ) + log p ( D| θ )] Kinetic : K ( r ) = 1 2 r ⊺ M − 1 r ⇒ π ( r ) ∼ N ( 0 , M ) The joint density of z = ( θ , r ) is π ( z ) ∝ exp ( − U ( θ ) − K ( r )) ∝ p ( θ |D ) · π ( r ) Hamilton’s equations : d θ d r dt = ∇ r H = ∇ r K ( r ) , dt = −∇ θ H = −∇ θ U ( θ ) s : R 2 d → R 2 d , z ( 0 ) = z �→ z ∗ = z ( s ) Hamiltonian flow φ H Properties : reversibility , volume preservation and constant Hamiltonian over time t Cheng Zhang UCI Variational HMC Jan 6, 2017 5 / 27

Recommend


More recommend