the big problem with meta learning and how bayesians can
play

The Big Problem with Meta-Learning and How Bayesians Can Fix It - PowerPoint PPT Presentation

The Big Problem with Meta-Learning and How Bayesians Can Fix It Chelsea Finn Stanford training data test datapoint Braque Cezanne By Braque or Cezanne? How did you accomplish this? Through previous experience. How might you get a


  1. The Big Problem with Meta-Learning 
 and How Bayesians Can Fix It Chelsea Finn Stanford

  2. training data test datapoint Braque Cezanne By Braque or Cezanne?

  3. How did you accomplish this? Through previous experience.

  4. How might you get a machine to accomplish this task? Modeling image formaKon Geometry Fewer human priors, more data -driven priors SIFT features, HOG features + SVM Greater success. Fine-tuning from ImageNet features Domain adaptaKon from other painters ??? Can we explicitly learn priors from previous experience that lead to efficient downstream learning? Can we learn to learn?

  5. Outline 1. Brief overview of meta-learning 2. The problem: peculiar, lesser-known, yet ubiquitous 3. Steps towards a solution

  6. How does meta-learning work? An example. Given 1 example of 5 classes: Classify new examples test set training data

  7. How does meta-learning work? An example. training meta-training classes … … Given 1 example of 5 classes: Classify new examples meta-testing T test test set training data

  8. How does meta-learning work? One approach : parameterize learner by neural network 4 0 1 2 3 4 y ts = f ( 𝒠 tr , x ts ; θ ) (Hochreiter et al. ’91, Santoro et al. ’16, many others)

  9. How does meta-learning work? Another approach : embed optimization inside the learning process 4 r θ L y ts = f ( 𝒠 tr , x ts ; θ ) 0 1 2 3 4 (Maclaurin et al. ’15, Finn et al. ’17, many others)

  10. The Bayesian perspective p ( ϕ | θ ) meta-learning <~> learning priors from data (Grant et al. ’18, Gordon et al. ’18, many others)

  11. Outline 1. Brief overview of meta-learning 2. The problem: peculiar, lesser-known, yet ubiquitous 3. First steps towards a solution

  12. How we construct tasks for meta-learning. 𝒠 tr x ts 0 1 2 3 4 2 4 0 1 2 3 4 3 1 T 3 0 1 2 3 4 4 3 Randomly assign class labels to image classes for each task —> Tasks are mutually exclusive . Algorithms must use training data to infer label ordering.

  13. What if label order is consistent? 𝒠 tr x ts 0 1 2 3 4 2 4 0 1 2 3 4 3 1 T 3 0 2 3 4 1 1 2 Tasks are non-mutually exclusive : a single function can solve all tasks. The network can simply learn to classify inputs, irrespective of 𝒠 tr

  14. The network can simply learn to classify inputs, irrespective of 𝒠 tr 4 1 2 3 4 0 4 r θ L 0 1 2 3 4

  15. What if label order is consistent? 𝒠 tr x ts 0 1 2 3 4 2 4 0 1 2 3 4 3 1 T 3 0 2 3 4 1 1 2 For new image classes: can’t make predictions w/o 𝒠 tr T test training data test set

  16. Is this a problem? - No : for image classi fi cation, we can just shu ffl e labels* - No , if we see the same image classes as training (& don’t need to adapt at meta-test time) - But, yes , if we want to be able to adapt with data for new tasks.

  17. Another example “hammer” “close drawer” “stack” meta-training … T 50 “close box” T test If you tell the robot the task goal, the robot can ignore the trials. T Yu, D Quillen, Z He, R Julian, K Hausman, C Finn, S Levine. Meta-World . CoRL ‘19

  18. Another example Model can memorize the canonical orientations of the training objects. Yin, Tucker, Yuan, Levine, Finn. Meta-Learning without Memorization . ‘19

  19. Can we do something about it?

  20. If tasks mutually exclusive : single function cannot solve all tasks (i.e. due to label shu ffl ing, hiding information) If tasks are non - mutually exclusive : single function can solve all tasks y ts = f θ ( D tr multiple solutions to the i , x ts ) meta-learning problem 𝒠 tr One solution: θ memorize canonical pose info in & ignore i 𝒠 tr Another solution: θ carry no info about canonical pose in , acquire from i An entire spectrum of solutions based on how information fl ows. Suggests a potential approach: control information fl ow. Yin, Tucker, Yuan, Levine, Finn. Meta-Learning without Memorization . ‘19

  21. If tasks are non - mutually exclusive : single function can solve all tasks y ts = f θ ( D tr multiple solutions to the i , x ts ) meta-learning problem 𝒠 tr One solution: θ memorize canonical pose info in & ignore i 𝒠 tr Another solution: θ carry no info about canonical pose in , acquire from i An entire spectrum of solutions based on how information fl ows. one option: max I ( ̂ y ts , 𝒠 tr | x ts ) Meta-regularization minimize meta-training loss + information in θ ℒ ( θ , 𝒠 meta − train ) + β D KL ( q ( θ ; θ μ , θ σ ) ∥ p ( θ )) θ Places precedence on using information from over storing info in . 𝒠 tr Can combine with your favorite meta-learning algorithm. Yin, Tucker, Yuan, Levine, Finn. Meta-Learning without Memorization . ‘19

  22. Omniglot without label shu ffl ing: “non-mutually-exclusive” Omniglot On pose prediction task: (and it’s not just as simple as standard regularization) TAML: Jamal & Qi. Task-Agnostic Meta-Learning for Few-Shot Learning . CVPR ‘19 Yin, Tucker, Yuan, Levine, Finn. Meta-Learning without Memorization . ‘19

  23. Does meta-regularization lead to better generalization? P ( θ ) θ Let be an arbitrary distribution over that doesn’t depend on the meta-training data. P ( θ ) = 𝒪 ( θ ; 0 , I ) (e.g. ) 1 − δ For MAML, with probability at least , ∀ θ μ , θ σ error on the meta-regularization generalization meta-training set error β With a Taylor expansion of the RHS + a particular value of —> recover the MR MAML objective . Proof: draws heavily on Amit & Meier ‘18 Yin, Tucker, Yuan, Levine, Finn. Meta-Learning without Memorization . ‘19

  24. CS330: Deep Multi-Task & Meta-Learning Want to Learn More? Lecture videos coming out soon! Working on Meta-RL? Try out the Meta-World benchmark Collaborators T Yu, D Quillen, Z He, R Julian, K Hausman, C Finn, S Levine. Meta-World . CoRL ‘19 Yin, Tucker, Yuan, Levine, Finn. Meta-Learning without Memorization . ‘19

Recommend


More recommend