how can we best combine forecasts for added value
play

How can we best combine forecasts for added value? J. Broecker 1 , - PowerPoint PPT Presentation

Thorpex Montreal 2004 How can we best combine forecasts for added value? J. Broecker 1 , L. Clarke 1 , D. Kilminster 2 and L.A. Smith 1 , 2 1. Department of Statistics, London School of Economics, UK 2. Pembroke College, Oxford, UK Thorpex


  1. Thorpex Montreal 2004 How can we best combine forecasts for added value? J. Broecker 1 , L. Clarke 1 , D. Kilminster 2 and L.A. Smith 1 , 2 1. Department of Statistics, London School of Economics, UK 2. Pembroke College, Oxford, UK

  2. Thorpex Montreal 2004 How can we best combine forecasts for added value? How might we use TIGGE? J. Broecker 1 , L. Clarke 1 , D. Kilminster 2 and L.A. Smith 1 , 2 1. Department of Statistics, London School of Economics, UK 2. Pembroke College, Oxford, UK

  3. Motivation 1 Forecast improvement can be achieved in two ways: Motivation Overview Combining Evaluation Example

  4. Motivation 1 Forecast improvement can be achieved in two ways: Motivation Overview • improving the models (strategic) Combining Evaluation Example

  5. Motivation 1 Forecast improvement can be achieved in two ways: Motivation Overview • improving the models (strategic) Combining • using the available information more effectively (tactical) Evaluation Example

  6. Motivation 1 Forecast improvement can be achieved in two ways: Motivation Overview • improving the models (strategic) Combining • using the available information more effectively (tactical) Evaluation Example THORPEX : “THORPEX will develop, demonstrate and evaluate a multi-model, multi-analysis and multi-national ensemble prediction system, referred to as TIGGE.”

  7. Overview 2 • Combining Simulations Motivation • Evaluation Overview – skill scores → Broecker – bootstrapping and meaningful skill comparison Combining • Example - combining ECMWF and NCEP Evaluation Example

  8. Inputs 3 Motivation Climatological Distribution Overview Combining Evaluation Example

  9. Inputs 3 Motivation Climatological Dressed Point Distribution Forecast Overview x Combining Evaluation Example Roulston & Smith, Tellus 55 2003 Raftery et al. Univ. Washington Dept. of Stat. Tech. Report 440 2003

  10. Inputs 3 Motivation Climatological Dressed Point Ensemble Distribution Forecast Product Overview x x x x x x x x x x x Combining Evaluation Example Roulston & Smith, Tellus 55 2003 Raftery et al. Univ. Washington Dept. of Stat. Tech. Report 440 2003

  11. Inputs 3 Motivation Climatological Dressed Point Ensemble Distribution Forecast Product Overview x x x x x x x x x x x Combining Evaluation Example Combined Forecast Roulston & Smith, Tellus 55 2003 Raftery et al. Univ. Washington Dept. of Stat. Tech. Report 440 2003

  12. Combining: Skill Scores 4 The combination is based on the skill of the Motivation final forecast Overview Combining s = S ( f, o ) Evaluation forecast distribution f Example verifying observation o skill score S skill of forecast f s

  13. Combining 5 One combination method is to take a weighted sum of the component distributions Motivation � f = α i f i Overview Combining Choose α i that maximises the skill score � s � = max �S ( f, o ) � Evaluation Example over a set of historical forecast-verification pairs Ignorance : s = − log p ( o ) Roulston & Smith, Monthly Weather Review 130

  14. Evaluation:Bootstrapping 6 Motivation Overview Combining Skill Evaluation Example Model A Model B Model C Out of sample, particular location, particular lead time, particular target

  15. Evaluation:Bootstrapping 6 Motivation Overview Combining Skill Evaluation Example Model A Model B Model C Out of sample, particular location, particular lead time, particular target

  16. The Right Comparison 7 We do not want to compare the uncertainty in the average performance of two models. Motivation Overview We want the uncertainty in the comparative Combining performance of the models to each other. Evaluation Bootstrap the difference: � s A − s B � BS Not the difference of the bootstraps: � s A � BS − � s B � BS Example

  17. Comparative Skill 8 Motivation Overview Comparitive Skill Combining Evaluation Example A vs B A vs C B vs C

  18. Example 9 Predicting temperature at Heathrow. Motivation Using Overview • NCEP high resolution • NCEP ensemble • ECMWF high resolution Combining • ECMWF ensemble Evaluation Evaluating using Ignorance - out of sample Example

  19. Example - one lead time 10 Motivation Overview Comparitive Skill Combining Evaluation Example Lead Time (Days) Combination vs NCEP Hi Res. IGN: smaller = good

  20. Example - one lead time 10 Motivation Overview Comparitive Skill Combining Evaluation Example Lead Time (Days) Combination vs NCEP Hi Res. Combination vs NCEP Ens. IGN: smaller = good

  21. Example - one lead time 10 Motivation Overview Comparitive Skill Combining Evaluation Example Lead Time (Days) Combination vs NCEP Hi Res. Combination vs NCEP Ens. Combination vs ECMWF Hi Res. IGN: smaller = good

  22. Example - one lead time 10 Motivation Overview Comparitive Skill Combining Evaluation Example Lead Time (Days) Combination vs NCEP Hi Res. Combination vs NCEP Ens. Combination vs ECMWF Hi Res. IGN: smaller = good Combination vs ECMWF Ens.

  23. Example - all lead times 11 Motivation Overview Comparitive Skill Combining Evaluation Example Lead Time (Days) Combination vs NCEP Hi Res. IGN: smaller = good

  24. Example - all lead times 11 Motivation Overview Comparitive Skill Combining Evaluation Example Lead Time (Days) Combination vs NCEP Hi Res. Combination vs NCEP Ens. IGN: smaller = good

  25. Example - all lead times 11 Motivation Overview Comparitive Skill Combining Evaluation Example Lead Time (Days) Combination vs NCEP Hi Res. Combination vs NCEP Ens. Combination vs ECMWF Hi Res. IGN: smaller = good

  26. Example - all lead times 11 Motivation Overview Comparitive Skill Combining Evaluation Example Lead Time (Days) Combination vs NCEP Hi Res. Combination vs NCEP Ens. Combination vs ECMWF Hi Res. IGN: smaller = good Combination vs ECMWF Ens.

  27. Summary 12 We have: Motivation • presented a user-orientated methodology for com- bining simulations • whatever combination method, the evaluation must Overview be robust • dressing method, combination method and size of Combining forecast-verifi cation archive affects performance • potential relevance to TIGGE, provides a frame- Evaluation work for allowing users to extract the forecast in- formation most relevant to them Example

Recommend


More recommend