implicit generation and generalization with energy based
play

Implicit Generation and Generalization with Energy Based Models - PowerPoint PPT Presentation

Implicit Generation and Generalization with Energy Based Models Yilun Du and Igor Mordatch Energy-Based Model E( x ) Distribution defined by energy function Train to maximize data likelihood gradient: Generate model samples


  1. Implicit Generation and Generalization with Energy Based Models Yilun Du and Igor Mordatch

  2. Energy-Based Model E( x ) • Distribution defined by energy function • Train to maximize data likelihood • gradient: • Generate model samples implicitly via Langevin Dynamics see [LeCun et al, 2006] for review

  3. Energy-Based Model E( x ) data • Distribution defined by energy function x + • Train to maximize data likelihood • gradient: • Generate model samples implicitly via Langevin Dynamics

  4. Energy-Based Model E( x ) data • Distribution defined by energy function x + x - hallucination • Train to maximize data likelihood • gradient: • Generate model samples implicitly via Langevin Dynamics See [Turner, 2006] for derivation

  5. Energy-Based Model E( x ) x 0 • Distribution defined by energy function x - • Train to maximize data likelihood • gradient: • Generate model samples implicitly via stochastic optimization Langevin Dynamics [Welling and Teh, 2011]

  6. Why Energy-Based Generative Models? 1 Implicit Generation • Flexibility • One Object to Learn • Compositionalitly • Generic Initialization and Computation Time 2 Intriguing Properties • Robustness • Online Learning

  7. Why Do EBMs Work Now? More compute and modern deep learning practices Faster Sampling • Continuous gradient based sampling using Langevin Dynamics • Replay buffer of past samples (similar to persistent CD) Stability improvements • Constrain Lipschitz constant of energy function (spectral norm) • Smoother activations (swish) • And others ...

  8. Comparison to Other Generative Models • gradient:

  9. ImageNet 128x128

  10. Cross Class Mapping

  11. Cross Class Mapping

  12. Surprising Benefits of Energy-Based Models • Robustness • Continual Learning • Compositionality • Trajectory Modeling

  13. Surprising Benefits of Energy-Based Models • Robustness • Continual Learning • Compositionality • Trajectory Modeling

  14. Out-of-Distribution Relative Likelihoods Also observed by [Hendrycks et al 2018] and [Nalisnick et al 2019]

  15. Out-of-Distribution Relative Likelihoods Also observed by [Hendrycks et al 2018] and [Nalisnick et al 2019]

  16. Out-of-Distribution Relative Likelihoods Also observed by [Hendrycks et al 2018] and [Nalisnick et al 2019]

  17. Out-of-Distribution Generalization • Following [Hendrycks and Gimpel, 2016]

  18. Robust Classification

  19. Robust Classification (recent follow-up submission at ICLR 2020 improves baseline EBM performance)

  20. Surprising Benefits of Energy-Based Models • Robustness • Continual Learning • Compositionality • Trajectory Modeling

  21. Continual Learning: Split MNIST Evaluation by [Hsu at al, 2019]

  22. Continual Learning: Split MNIST Evaluation by [Hsu at al, 2019]

  23. Continual Learning: Split MNIST EBM: 64.99 ± 4.27 (10 seeds) Evaluation by [Hsu at al, 2019]

  24. Continual Learning: Split MNIST EBM: 64.99 ± 4.27 Would any generative model work instead? Doesn’t look like it: Evaluation by [Hsu at al, 2019] VAE: 40.04 ± 1.31

  25. Surprising Benefits of Energy-Based Models • Robustness • Continual Learning • Compositionality • Trajectory Modeling

  26. Compositionality via Sum of EBMs [Hinton, 1999] [Mnih and Hinton, 2005] Specify a concept by successively adding constraints

  27. Compositionality via Sum of Energies Specify a concept by successively adding constraints Compositional Visual Generation with EBMs [Du, Li, Mordatch, 2019]

  28. Compositionality via Sum of Energies Specify a concept by successively adding constraints Compositional Visual Generation with EBMs [Du, Li, Mordatch, 2019]

  29. Compositionality via Sum of Energies Specify a concept by successively adding constraints Compositional Visual Generation with EBMs [Du, Li, Mordatch, 2019]

  30. Compositionality via Sum of Energies Specify a concept by successively adding constraints Compositional Visual Generation with EBMs [Du, Li, Mordatch, 2019]

  31. Compositionality via Sum of Energies Specify a concept by successively adding constraints Compositional Visual Generation with EBMs [Du, Li, Mordatch, 2019]

  32. Surprising Benefits of Energy-Based Models • Robustness • Continual Learning • Compositionality • Trajectory Modeling

  33. EBMs for Trajectory Modeling and Control [Du, Lin, Mordatch, 2019] • Train energy to model pairwise state transitions s t , s t+1 • Trajectory probability: s t s t+1 s T s 1 -E(s t , s t+1 )

  34. EBMs for Trajectory Modeling and Control [Du, Lin, Mordatch, 2019] • Train energy to model pairwise state transitions s t , s t+1 • Generate trajectories that achieve specific tasks: EBM Task R(s t ) s t s T s 1 (similar to direct trajectory optimization)

  35. EBMs for Control

  36. Source Code • Images • https://github.com/openai/ebm_code_release • Trajectories • https://github.com/yilundu/model_based_planning_ebm • Compositionality • https://drive.google.com/file/d/ 138w7Oj8rQl_e40_RfZJq2WKWb41NgKn3 • Interactive Notebook • https://drive.google.com/file/d/ 1fCFRw_YtqQPSNoqznIh2b1L2baFgLz4W/view

Recommend


More recommend