sampling effect on performance prediction of configurable
play

Sampling Effect on Performance Prediction of Configurable Systems : - PowerPoint PPT Presentation

Sampling Effect on Performance Prediction of Configurable Systems : A Case Study Juliana Alves Pereira, Mathieu Acher, Hugo Martin, Jean-Marc Jezequel 1 Configurable systems 2 Configurable systems 2 Configurable systems Pros


  1. Sampling Effect on Performance Prediction of Configurable Systems : A Case Study Juliana Alves Pereira, Mathieu Acher, Hugo Martin, Jean-Marc Jezequel 1

  2. Configurable systems 2

  3. Configurable systems 2

  4. Configurable systems Pros ● Adaptive ● Lots of options 2

  5. Configurable systems Pros ● Adaptive ● Lots of options Cons ● Lots of options (and interactions) ● Increasingly complex 2

  6. Configurable systems Pros ● Adaptive ● Lots of options Cons ● Lots of options (and interactions) ● Increasingly complex Machine learning to the rescue 2

  7. Machine Learning and Configurable systems 3

  8. Machine Learning and Configurable systems Sampling 3

  9. Machine Learning and Configurable systems Sampling Measuring 3

  10. Machine Learning and Configurable systems Sampling Measuring Learning 3

  11. Machine Learning and Configurable systems Sampling Measuring Validation Learning 3

  12. Machine Learning and Configurable systems Sampling Measuring Validation Learning 3

  13. Distance-Based Sampling of Software Configuration Spaces 4

  14. Distance-Based Sampling of Software Configuration Spaces ● C. Kaltenecker, A. Grebhahn, N. Siegmund, J. Guo and S. Apel, "Distance-Based Sampling of Software Configuration Spaces," 2019 IEEE/ACM 41st International Conference on Software Engineering (ICSE) , Montreal, QC, Canada, 2019, pp. 1084-1094. 4

  15. Distance-Based Sampling of Software Configuration Spaces ● C. Kaltenecker, A. Grebhahn, N. Siegmund, J. Guo and S. Apel, "Distance-Based Sampling of Software Configuration Spaces," 2019 IEEE/ACM 41st International Conference on Software Engineering (ICSE) , Montreal, QC, Canada, 2019, pp. 1084-1094. ● Proposing a new sampling solution : Distance-Based Sampling 4

  16. Distance-Based Sampling of Software Configuration Spaces ● C. Kaltenecker, A. Grebhahn, N. Siegmund, J. Guo and S. Apel, "Distance-Based Sampling of Software Configuration Spaces," 2019 IEEE/ACM 41st International Conference on Software Engineering (ICSE) , Montreal, QC, Canada, 2019, pp. 1084-1094. ● Proposing a new sampling solution : Distance-Based Sampling ● Empirical study on 10 subject systems and 6 sampling strategies 4

  17. Sampling strategies ● Coverage-based 5

  18. Sampling strategies ● Coverage-based ● Solver-based ● Randomized solver-based 5

  19. Sampling strategies ● Coverage-based ● Solver-based ● Randomized solver-based ● Random 5

  20. Sampling strategies ● Coverage-based ● Solver-based ● Randomized solver-based ● Random ● Distance-based ● Diversified distance-based 5

  21. Subject systems ● 7z ● BerkeleyDB-C ● Dune MGS ● HIPAcc ● Java GC ● LLVM ● LRZIP ● Polly ● VPXENC ● x264 6

  22. Subject systems Experiment setup ● 7z ● Machine learning based on multiple ● BerkeleyDB-C linear regression and feature-forward ● Dune MGS selection ● HIPAcc ● Mean Relative Error (MRE) ● Java GC ● LLVM ● LRZIP ● Polly ● VPXENC ● x264 6

  23. Results ● Coverage-based is dominant at low sample size ● Diversified distance-based is dominant on higher sample size ● Diversified distance-based is close to random sampling accuracy, even better in some cases 7

  24. Is it true?

  25. Replicating the experiment 9

  26. Replicating the experiment ● Subject system : x264, video encoder 9

  27. Replicating the experiment ● Subject system : x264, video encoder 9

  28. Replicating the experiment ● Subject system : x264, video encoder ● Changing the input video : 17 videos 9

  29. Replicating the experiment ● Subject system : x264, video encoder ● Changing the input video : 17 videos ● Changing the measured non-functional property 9

  30. Experimental setup What does vary? ● Sampling strategy (6 strategies) ● Sample size (3 sample size) ● Encoded video (17 videos) ● System configuration (1152 configurations) ● Measured property (Encoding time, encoding size) 10

  31. Experimental setup What does vary? ● Sampling strategy (6 strategies) ● Sample size (3 sample size) ● Encoded video (17 videos) ● System configuration (1152 configurations) ● Measured property (Encoding time, encoding size) What doesn’t vary? ● Learning algorithm (Performance-Influence Model) ● Learning algorithm hyperparameters ● Configurable Software (x264) ● Version ● Hardware 10

  32. Experimental setup What does vary? ● Sampling strategy (6 strategies) ● Sample size (3 sample size) ● Encoded video (17 videos) 🔵 ● System configuration (1152 configurations) ● Measured property (Encoding time, encoding size) 🔵 What doesn’t vary? ● Learning algorithm (Performance-Influence Model) ● Learning algorithm hyperparameters Configurable Software (x264) 🔵 ● ● Version ● Hardware 10

  33. Experimental setup What does vary? ● Sampling strategy (6 strategies) ● Sample size (3 sample size) ● Encoded video (17 videos) 🔵 ● System configuration (1152 configurations) ● Measured property (Encoding time, encoding size) 🔵 What doesn’t vary? ● Learning algorithm (Performance-Influence Model) ● Learning algorithm hyperparameters Configurable Software (x264) 🔵 ● ● Version 🔶 Hardware 🔶 ● 10

  34. Results 11

  35. 11 Results table for encoding time

  36. 11 Results table for encoding time

  37. 11 Results table for encoding time

  38. 11 Results table for encoding time

  39. 11 Results table for encoding time

  40. 11 Results table for encoding time

  41. 11 Results table for encoding time

  42. 11 Results table for encoding size

  43. 11 Results table for encoding size

  44. 11 Results table for encoding size

  45. 11 Results table for encoding size

  46. 11 Results table for encoding size

  47. 11 Results table for encoding size

  48. Results 11

  49. Results ● High variation between videos, between non-functional properties 11

  50. Results ● High variation between videos, between non-functional properties ● Encoding time : ○ Similar results ○ Random sampling dominant over Diversified Distance-based sampling 11

  51. Results ● High variation between videos, between non-functional properties ● Encoding time : ○ Similar results ○ Random sampling dominant over Diversified Distance-based sampling ● Encoding size : ○ Random sampling and randomized solver-based sampling overall dominant ○ Most strategies present good and similar accuracy for higher sample size 11

  52. Replicability ● Fully replicable experiment 12

  53. Replicability ● Fully replicable experiment 12

  54. Replicability ● Fully replicable experiment ● Dataset for video encoding time and size available 12

  55. Replicability ● Fully replicable experiment ● Dataset for video encoding time and size available ● Docker image with all data and scripts for performance prediction and results aggregation : https://github.com/jualvespereira/ICPE2020 12

  56. What’s next? 13

  57. What’s next? ● How do version and hardware affect the sampling effectiveness? 13

  58. What’s next? ● How do version and hardware affect the sampling effectiveness? ● How does machine learning technique affect the sampling effectiveness? 13

  59. What’s next? ● How do version and hardware affect the sampling effectiveness? ● How does machine learning technique affect the sampling effectiveness? ● How to leverage the fact that some sampling strategies overperform by focusing on important options? 13

  60. Conclusion 14

  61. Conclusion ● Random sampling is a strong baseline, hard to challenge 14

  62. Conclusion ● Random sampling is a strong baseline, hard to challenge ● Diversified distance-based sampling is a strong alternative 14

  63. Conclusion ● Random sampling is a strong baseline, hard to challenge ● Diversified distance-based sampling is a strong alternative ● Researchers should be aware that effectiveness of sampling strategies can be biased by inputs and performance property used 14

Recommend


More recommend