hyperparameter optimization
play

Hyperparameter Optimization Albert-Ludwigs-Universitt Freiburg - PowerPoint PPT Presentation

Surrogate Benchmarks for Hyperparameter Optimization Albert-Ludwigs-Universitt Freiburg Holger Hoos Katharina Eggensperger Kevin Leyton-Brown Frank Hutter University of British Columbia University of Freiburg {hoos,kevinlb}@cs.ubc.ca


  1. Surrogate Benchmarks for Hyperparameter Optimization Albert-Ludwigs-Universität Freiburg Holger Hoos Katharina Eggensperger Kevin Leyton-Brown Frank Hutter University of British Columbia University of Freiburg {hoos,kevinlb}@cs.ubc.ca {eggenspk,fh}@cs.uni-freiburg.de

  2. Problem: Evaluation of Methods for Hyperparameter Optimization is expensive ! Albert-Ludwigs-Universität Freiburg

  3. Outline  Benchmarking Hyperparameter Optimization Methods  Constructing Surrogates  Using Surrogate Benchmarks MetaSEL’14 3 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  4. Outline  Benchmarking Hyperparameter Optimization Methods  Constructing Surrogates  Using Surrogate Benchmarks MetaSEL’14 4 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  5. Bayesian Optimization Methods Configuration Optimizer space Λ Uses internal model M λ i Performance Run algorithm f( λ i ) with configuration λ i MetaSEL’14 5 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  6. What do we need for an empirical comparison  Standard benchmark problems  Easy-to-use software Then:  Run each optimizer on each benchmark X multiple times MetaSEL’14 6 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  7. What do we need for an empirical comparison  Standard benchmark problems  Easy-to-use software Then:  Run each optimizer on each benchmark X multiple times Evaluation of X is expensive MetaSEL’14 6 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  8. Benchmarking hyperparameter optimization methods Neural Network, configuration space Λ : MetaSEL’14 7 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  9. Benchmarking hyperparameter optimization methods Neural Network, configuration space Λ : categorical hyperparameter MetaSEL’14 7 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  10. Benchmarking hyperparameter optimization methods Neural Network, configuration space Λ : conditional hyperparameter MetaSEL’14 7 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  11. Benchmarking hyperparameter optimization methods Neural network 8 MetaSEL’14 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  12. Benchmarking hyperparameter optimization methods Neural network 8 MetaSEL’14 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  13. Outline  Benchmarking Hyperparameter Optimization Methods  Constructing Surrogates  Using Surrogate Benchmarks MetaSEL’14 9 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  14. Surrogate Benchmark 𝑌 ′ • cheap-to-evaluate • Can be used like the real benchmark X • Behaves like X Configuration 𝜇 𝑌 Performance 𝑔(𝜇) MetaSEL’14 10 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  15. Surrogate Benchmark 𝑌 ′ • cheap-to-evaluate • Can be used like the real benchmark X • Behaves like X Configuration 𝜇 Regression model 𝑌 ′ Performance 𝑔(𝜇) MetaSEL’14 11 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  16. Constructing a Surrogate for Benchmark X 1. Collect data 2. Choose a regression model 3. Train and store model MetaSEL’14 12 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  17. 1. Collect data for benchmark X 𝜇 1 , 𝑔 𝜇 1 , … , 𝜇 𝑜 , 𝑔 𝜇 𝑜 Training data:  Dense sampling in high performance regions  Good overall coverage MetaSEL’14 13 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  18. 1. Collect data for benchmark X 𝜇 1 , 𝑔 𝜇 1 , … , 𝜇 𝑜 , 𝑔 𝜇 𝑜 Training data:  Dense sampling in high performance regions Run optimizers on benchmark X  Good overall coverage MetaSEL’14 13 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  19. 1. Collect data for benchmark X 𝜇 1 , 𝑔 𝜇 1 , … , 𝜇 𝑜 , 𝑔 𝜇 𝑜 Training data:  Dense sampling in high performance regions Run optimizers on benchmark X  Good overall coverage Run random search on benchmark X MetaSEL’14 13 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  20. 2. Choice of Regression Models Ridge Regression K-nearest neighbour Gradient Boosting Linear Regression Random Forests Gaussian Processes Bayesian Neural Network SVM MetaSEL’14 14 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  21. 2. Choice of Regression Models Ridge Regression K-nearest neighbour Gradient Boosting Linear Regression Random Forests Gaussian Processes Bayesian Neural Network SVM MetaSEL’14 15 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  22. 2. Choice of Regression Models Can we quantify the performance of a new optimizer?  Leave-one-optimizer-out setting - Train model on data gathered by all but one optimizer - Test on remaining data MetaSEL’14 16 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  23. 2. Choice of Regression Models Leave-one-optimizer-out setting Random forest prediction Neural Network True performance MetaSEL’14 17 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  24. 2. Choice of Regression Models Leave-one-optimizer-out setting Random Forest Neural Network MetaSEL’14 17 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  25. 2. Choice of Regression Models Leave-one-optimizer-out setting Random Forest Gaussian Process k-nearest-neighbour Gradient Boosting nuSVR Neural Network MetaSEL’14 17 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  26. 2. Choice of Regression Models Ridge Regression K-nearest neighbour Gradient Boosting Linear Regression Random Forests Gaussian Processes Bayesian Neural Network SVM MetaSEL’14 18 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  27. Outline  Benchmarking Hyperparameter Optimization Methods  Constructing Surrogates  Using Surrogate Benchmarks MetaSEL’14 19 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  28. Using Surrogate Benchmarks Neural Network MetaSEL’14 20 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  29. Using Surrogate Benchmarks Neural Network Real Benchmark MetaSEL’14 20 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  30. Using Surrogate Benchmarks Neural Network GP-based benchmark Real Benchmark RF-based benchmark MetaSEL’14 20 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  31. Using Surrogate Benchmarks Neural Network GP-based benchmark Real Benchmark RF-based benchmark One optimization run: 40h <200s <200s <1.5h <1.5h Whole comparison: 50d MetaSEL’14 21 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  32. Applications  Extensive testing at early development stages  Fast comparison of different hyperparameter optimization methods  Metaoptimization of existing hyperparameter optimization methods MetaSEL’14 23 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  33. Conclusion Can we construct cheap-to evaluate and realistic hyperparameter optimization benchmarks ? Yes, based on random forests and Gaussian process regression models  MetaSEL’14 24 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  34. Conclusion Can we construct cheap-to evaluate and realistic hyperparameter optimization benchmarks ? Yes, based on random forests and Gaussian process regression models But, some work needs to be done for high dimensional benchmarks. MetaSEL’14 24 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  35. This presentation was supported by an ECCAI travel award and the ECCAI sponsors Thank you for your attention Albert-Ludwigs-Universität Freiburg More information on hyperparameter optimization benchmarks can be found on automl.org/hpolib

  36. Regression models MetaSEL’14 26 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

  37. Benchmarks MetaSEL’14 27 Surrogates for Hyperparameter Optimization Benchmarks – Eggensperger, Hutter, Hoos, and Leyton-Brown

Recommend


More recommend