Verification of Renewable Energy Forecasts

Similar documents
Evaluating Forecast Quality

Non-parametric Probabilistic Forecasts of Wind Power: Required Properties and Evaluation

Benchmark study of Icing Forecasts Do they add value? Ben Martinez, Pierre Julien Trombe Vattenfall R&D

Ensemble Verification Metrics

Verification of Continuous Forecasts

Reliability diagrams for nonparametric density forecasts of continuous variables: accounting for serial correlation

Wind power prediction risk indices based on numerical weather prediction ensembles

Verification of ensemble and probability forecasts

1 (19) Hannele Holttinen, Jussi Ikäheimo

FORECASTING: A REVIEW OF STATUS AND CHALLENGES. Eric Grimit and Kristin Larson 3TIER, Inc. Pacific Northwest Weather Workshop March 5-6, 2010

VERFICATION OF OCEAN WAVE ENSEMBLE FORECAST AT NCEP 1. Degui Cao, H.S. Chen and Hendrik Tolman

Model verification / validation A distributions-oriented approach

Verification of Probability Forecasts

Enhancing Weather Information with Probability Forecasts. An Information Statement of the American Meteorological Society

Methods of forecast verification

GDP growth and inflation forecasting performance of Asian Development Outlook

Basic Verification Concepts

Verification of Weather Warnings

Anemos.Rulez: Extreme Event Prediction and Alarming to Support Stability of Energy Grids

Basic Verification Concepts

Wind-Wave Probabilistic Forecasting based on Ensemble Predictions

Probabilistic maximum-value wind prediction for offshore environments

Proper Scores for Probability Forecasts Can Never Be Equitable

Operational use of ensemble hydrometeorological forecasts at EDF (french producer of energy)

Model verification and tools. C. Zingerle ZAMG

Application and verification of ECMWF products 2015

J2.4 SKILLFUL SEASONAL DEGREE-DAY FORECASTS AND THEIR UTILITY IN THE WEATHER DERIVATIVES MARKET

Denver International Airport MDSS Demonstration Verification Report for the Season

CAISO Participating Intermittent Resource Program for Wind Generation

GL Garrad Hassan Short term power forecasts for large offshore wind turbine arrays

FORECASTING OF WIND GENERATION The wind power of tomorrow on your screen today!

Modelling Wind Farm Data and the Short Term Prediction of Wind Speeds

Task 36 Forecasting for Wind Power

This wind energy forecasting capability relies on an automated, desktop PC-based system which uses the Eta forecast model as the primary input.

Verification challenges in developing TC response procedures based on ensemble guidance

Verification of the operational NWP models at DWD - with special focus at COSMO-EU

Forecasting scenarios of wind power generation for the next 48 hours to assist decision-making in the Australian National Electricity Market

Verification of nowcasts and short-range forecasts, including aviation weather

Perfect and Imperfect Competition in Electricity Markets

Forecast Verification Analysis of Rainfall for Southern Districts of Tamil Nadu, India

Wind resource assessment and wind power forecasting

Probabilistic verification

Gefördert auf Grund eines Beschlusses des Deutschen Bundestages

Current best practice of uncertainty forecast for wind energy

Benchmark of forecasting models

Wind Power Forecasting

Warwick Business School Forecasting System. Summary. Ana Galvao, Anthony Garratt and James Mitchell November, 2014

FORECASTING poor air quality events associated with

Revisiting predictability of the strongest storms that have hit France over the past 32 years.

Probabilistic Forecasts of Wind Power Generation by Stochastic Differential Equation Models

Modelling residual wind farm variability using HMMs

Allison Monarski, University of Maryland Masters Scholarly Paper, December 6, Department of Atmospheric and Oceanic Science

About Nnergix +2, More than 2,5 GW forecasted. Forecasting in 5 countries. 4 predictive technologies. More than power facilities

Categorical Verification

High-resolution forecasting of wind power generation with regime switching models and off-site observations

Upscaled and fuzzy probabilistic forecasts: verification results

Current verification practices with a particular focus on dust

operational status and developments

Spatial forecast verification

Cómo contar la predicción probabilistica? Part I: The problem with probabilities

Prashant Pant 1, Achal Garg 2 1,2 Engineer, Keppel Offshore and Marine Engineering India Pvt. Ltd, Mumbai. IJRASET 2013: All Rights are Reserved 356

High Wind and Energy Specific Models for Global. Production Forecast

Research and application of locational wind forecasting in the UK

Generation of Statistical Scenarios of Short-term Wind Power Production

J11.5 HYDROLOGIC APPLICATIONS OF SHORT AND MEDIUM RANGE ENSEMBLE FORECASTS IN THE NWS ADVANCED HYDROLOGIC PREDICTION SERVICES (AHPS)

Wind Power Production Estimation through Short-Term Forecasting

Predicting the Electricity Demand Response via Data-driven Inverse Optimization

Optimization of the forecasting of wind energy production Focus on the day ahead forecast

Wind power and management of the electric system. EWEA Wind Power Forecasting 2015 Leuven, BELGIUM - 02/10/2015

Chapter 6 Impact of Stochastic Renewable Energy Generation on Market Quantities 6.1 Introduction

Inflow Forecasting for Hydropower Operations: Bayesian Model Averaging for Postprocessing Hydrological Ensembles

Seamless water forecasting for Australia

Skilful seasonal predictions for the European Energy Industry

Evaluation of simple wind power forecasting methods applied to a long-term wind record from Scotland

The Role of Meteorological Forecast Verification in Aviation. Günter Mahringer, November 2012

Standardized Anomaly Model Output Statistics Over Complex Terrain.

Validation of Forecasts (Forecast Verification) Overview. Ian Jolliffe

Part 1: You are given the following system of two equations: x + 2y = 16 3x 4y = 2

The Weather Information Value Chain

CHAPTER 6 CONCLUSION AND FUTURE SCOPE

Probabilistic seasonal forecast verification

Calibration of ECMWF forecasts

Verification of wind forecasts of ramping events

Focus on Spatial Verification Filtering techniques. Flora Gofa

Space-time trajectories of wind power generation: Parameterized precision matrices under a Gaussian copula approach

Regional Production, Quarterly report on the daily analyses and forecasts activities, and verification of the CHIMERE performances

Exemplar for Internal Achievement Standard. Mathematics and Statistics Level 3

Aggregate Forecasting of Wind Generation on the Irish Grid Using a Multi-Scheme Ensemble Prediction System

Forecast verification at Deltares

NHC Ensemble/Probabilistic Guidance Products

Energy Forecasting Customers: Analysing end users requirements Dec 3rd, 2013 Carlos Alberto Castaño, PhD Head of R&D

Industrial Engineering Prof. Inderdeep Singh Department of Mechanical & Industrial Engineering Indian Institute of Technology, Roorkee

Proper Security Criteria Determination in a Power System with High Penetration of Renewable Resources

Quantile Forecasting of Wind Power Using Variability Indices

Analog Ensemble for Probabilis1c Weather Predic1ons

Ensemble forecasting: Error bars and beyond. Jim Hansen, NRL Walter Sessions, NRL Jeff Reid,NRL May, 2011

The Making Of A Great Contraction. With A Liquidity Trap And A Jobless Recovery. Columbia University

Bringing Renewables to the Grid. John Dumas Director Wholesale Market Operations ERCOT

Adaptation for global application of calibration and downscaling methods of medium range ensemble weather forecasts

Probabilistic Weather Forecasting and the EPS at ECMWF

Transcription:

Verification of Renewable Energy Forecasts Pierre Pinson Technical University of Denmark. DTU Electrical Engineering - Centre for Electric Power and Energy mail: ppin@dtu.dk - webpage: www.pierrepinson.com 16 April 2018 31761 - Renewables in Electricity Markets 1

Learning objectives Through this lecture and additional study material, it is aimed for the students to be able to: 1 Explain what makes renewable energy forecasts of different quality and value 2 Describe how one may evaluate the quality of different forms of forecasts 3 Appraise how different scores and diagnostic tools should be used and interpreted 31761 - Renewables in Electricity Markets 2

A few interesting quotes on forecasting Some of my favorites: Prediction is very difficult, especially if it s about the future Nils Bohr, Nobel laureate in Physics Forecasting is the art of saying what will happen, and then explaining why it didn t! Anonymous It is far better to foresee even without certainty than not to foresee at all Henri Poincaré A good sample is gathered at: Exeter University - famous forecasting quotes 31761 - Renewables in Electricity Markets 3

Let s accept it... Forecasts are always wrong! Bad forecasts translate to consequences - these may be: security issues in, e.g., offshore wind farm maintenance financial losses for those participating in the markets overall decrease in social welfare blackouts! (well, hopefully not)... but definitely, harsh criticism on using renewables for supplying us with electricity 31761 - Renewables in Electricity Markets 4

Outline 1 What makes a good forecast? 2 Test case and general considerations 3 Verification of point (deterministic) forecasts scores diagnostic tools 4 Verification of probabilistic forecasts attributes of forecast quality scores diagnostic tools 31761 - Renewables in Electricity Markets 5

1 What makes a good forecast? 31761 - Renewables in Electricity Markets 6

The nature of goodness in forecasting Following Murphy (ref. and link below), the nature of goodness in weather forecasting (same goes for other types of forecasts) consists in: 31761 - Renewables in Electricity Markets 7

The nature of goodness in forecasting Following Murphy (ref. and link below), the nature of goodness in weather forecasting (same goes for other types of forecasts) consists in: Forecast consistency: Forecasts should correspond to the forecaster s best judgement on future events, based on the knoweldge available at the time of issuing the forecasts 31761 - Renewables in Electricity Markets 8

The nature of goodness in forecasting Following Murphy (ref. and link below), the nature of goodness in weather forecasting (same goes for other types of forecasts) consists in: Forecast consistency: Forecasts should correspond to the forecaster s best judgement on future events, based on the knoweldge available at the time of issuing the forecasts Forecast quality: Forecasts should describe future events as good as possible, regardless of what these forecasts may be used for 31761 - Renewables in Electricity Markets 9

The nature of goodness in forecasting Following Murphy (ref. and link below), the nature of goodness in weather forecasting (same goes for other types of forecasts) consists in: Forecast consistency: Forecasts should correspond to the forecaster s best judgement on future events, based on the knoweldge available at the time of issuing the forecasts Forecast quality: Forecasts should describe future events as good as possible, regardless of what these forecasts may be used for Forecast value: Forecasts should bring additional benefits (monetary or others) when used as input to decision-making [Extra reading: AH Murphy (1993). What is a good forecast? An essay on the nature of goodness in weather forecasting. Weather and Forecasting 8: 281 293 (pdf)] 31761 - Renewables in Electricity Markets 10

Illustrative example (1) You are in charge of optimal maintenance planning at Horns Rev, and have booked both a vessel and an helicopter for onsite service (for a cost of 100.000e) The conditions for this to happen at time t + k are wind speed: u t+k 15 m.s -1 wave height: h t+k 1.8 m 24 hours before service (time t), this is your last chance to cancel before huge financial penalties (another 100.000e) Your two forecasters (Foresight and Blindspot) tell you that: Foresight Blindspot û t+k t 12.6 m.s -1 3.4 m.s -1 ĥ t+k t 1.6 m 0.2 m In both cases, you go ahead with the planned service... 31761 - Renewables in Electricity Markets 11

Illustrative example (1, continued) At time t + k, this is what actually happened: Foresight Blindspot û t+k t 12.6 m.s -1 3.4 m.s -1 ĥ t+k t 1.6 m 0.2 m u t+k 12.3 m.s -1 h t+k 1.45 m In both cases, your overall cost is 100.000e, Both Foresight and Blindspot served their purpose, since you made the right decision... Forecast value is good You might want to have a chat with Blindspot, since its forecast quality appears to be far from good! 31761 - Renewables in Electricity Markets 12

Illustrative example (2) The boy who cried wolf (Tale from Ancient Greece) - revisited. Rogue Trading R made huge losses last year, due to expensive upregulation events... It is therefore decided to get a new forecaster that would be good at predicting them Foresight and Blindspot are in competition for the job The score is simple: Sc = 100 the higher the better! (0 is worst, 100 is best) #{events leading to upregulation predicted} #{events leading to upregulation} 31761 - Renewables in Electricity Markets 13

Illustrative example (2, continued) If you were Foresight and Blindspot, what would you do? 31761 - Renewables in Electricity Markets 14

Illustrative example (2, continued) If you were Foresight and Blindspot, what would you do? The two competitors have sharpened their strategy: Strategy Foresight Always predict need for upregulation! Blindspot Do your best to find when upregulation will occur... The results on the benchmarking exercise are such that: #{market time units} = 8760 #{events leading to upregulation} = 3237 #{events leading to upregulation predicted by Foresight} = 3237 #{events leading to upregulation predicted by Blindspot} = 2500 Their scores: Foresight Blindspot Sc 100% 77.2% 31761 - Renewables in Electricity Markets 15

Illustrative example (2, continued) If you were Foresight and Blindspot, what would you do? The two competitors have sharpened their strategy: Strategy Foresight Always predict need for upregulation! Blindspot Do your best to find when upregulation will occur... The results on the benchmarking exercise are such that: #{market time units} = 8760 #{events leading to upregulation} = 3237 #{events leading to upregulation predicted by Foresight} = 3237 #{events leading to upregulation predicted by Blindspot} = 2500 Their scores: Foresight Blindspot Sc 100% 77.2% Foresight gets the job! 31761 - Renewables in Electricity Markets 16

Illustrative example (2, continued) The consequences are: even though never missing on upregulation events, Rogue Trading R will always miss the down regulation ones eventually, the financial loss may still be there... and possibly much higher than expected 31761 - Renewables in Electricity Markets 17

Illustrative example (2, continued) The consequences are: even though never missing on upregulation events, Rogue Trading R will always miss the down regulation ones eventually, the financial loss may still be there... and possibly much higher than expected A more consistent way to evaluate these forecasters would be to consider: event happens no event event predicted HIT FALSE ALARM event not predicted MISS CORRECT REJECTION And a proper score, ensuring forecast consistency, is: Sc = 100 #{hits} #{hits} + #{misses} + #{false alarms} The higher the better! (0 is worst, 100 is best) (This score is called the Threat Score (TS)) 31761 - Renewables in Electricity Markets 18

Illustrative example (2, continued) In the present case: Foresight Blindspot #{hits} 3237 2320 #{misses} 0 917 #{false alarms} 5523 180 #{correct rejections} 0 5343 The resulting Threat Score (TS) values are: Foresight Blindspot TS 36.9% 67.9% Conclusions: if using a proper score... Blindspot should have gotten the job! I can promise that Rogue Trading R would have lower financial losses 31761 - Renewables in Electricity Markets 19

2 Test case and general considerations 31761 - Renewables in Electricity Markets 20

Test case: the Klim wind farm The wind farm: full name: Klim Fjordholme onshore/offshore: onshore year of commissioning: 1996 nominal capacity (P n): 21 MW number of turbines in farm: 35 average annual electricity generation: 49 GWh data available: 1999-2003 (for some researchers) temporal resolution: 5 mins, and hourly averages forecasts: deterministic and probabilistic A link to the online description: Vattenfall s Klim wind farm The wind farm has been recommissioned recently: NordJyske online article 31761 - Renewables in Electricity Markets 21

Splitting of available data Forecasting is about being able to predict future events, in new situations not only explain what happen in the past... One need to verify forecasts on data that has not been used for the modelling! power [MW] 0 5 10 15 20 25 Modelling Evaluation 1 1 2002 1 7 2002 31 12 2002 Here we will focus on the last 6 months of 2002, while giving examples for some other periods 31761 - Renewables in Electricity Markets 22

3 Verification of point (deterministic) forecasts 31761 - Renewables in Electricity Markets 23

Visual inspection of forecasts Visual inspection allows you to develop susbtantial insight on forecast quality... This comprises a qualitative analysis only What do you think of these two? Are they good or bad? Forecast issued on 16 November 2001 (18:00) Forecast issued on 23 December 2003 (12:00) 31761 - Renewables in Electricity Markets 24

Various types of forecast error patterns Errors in renewable energy generation (but also load, price, etc.) are most often driven by weather forecasts errors Typical error patterns are: amplitude errors (left, below) phase errors (right, below) Forecast issued on 29 March 2003 (12:00) Forecast issued on 6 November 2002 (00:00) 31761 - Renewables in Electricity Markets 25

Quantitative analysis and the forecast error For continuous variables such as renewable energy generation (but also electricity prices or electric load for instance) qualitative analysis ought to be complemented by a quantitative analysis these are based on scores and diagnostic tools The base concept is that of the forecast error: ε t+k t = y t+k ŷ t+k t, P n ε t+k t P n where ŷ t+k t is the forecast issued at time t for time t + k y t+k is the observation at time t + k P n is the nominal capacity of the wind farm It can be calculated directly for the quantity of interest as a normalized version, for instance by dividing by the nominal capacity of the wind farm if evaluating wind power forecasts: ε t+k t = y t+k ŷ t+k t P n, 1 ε t+k t 1 31761 - Renewables in Electricity Markets 26

Forecast error: examples Example 1: If the 24-ahead prediction for Klim is of 18 MW, while the observation is 15.5MW ε t+k t = 2.5MW (if not normalized) ε t+k t = 0.119 (or, -11.9%, if normalized) Example 2: forecast issued on the 6 November 2002 (00:00) Forecast and observations Corresponding forecast errors (Note that we prefer to work with normalized errors from now on...) 31761 - Renewables in Electricity Markets 27

Scores for point forecast verification One cannot look at all forecasts, observations, and forecasts errors over a long period of time Scores are to be used to summarize aspects of forecast accuracy... The most common scores include, as function of the lead time k: bias (or Nbias, for the normalized version) bias(k) = 1 T T t=1 ε t+k t Mean Absolute Error (MAE) (or NMAE, for the normalized version) MAE(k) = 1 T T t=1 ε t+k t Root Mean Square Error (RMSE) (or NRMSE, for the normalized version) [ 1 1 2 RMSE(k) = T t=1 T t+k t] ε2 MAE and RMSE are negatively-oriented (the lower, the better) Let us illustrate their advantages and drawbacks... (black board illustration) 31761 - Renewables in Electricity Markets 28

Example: calculating a few scores at Klim Period: 1.7.2012-31.12.2012 Forecats quality necessarily degrades with further lead times For instance, for 24-ahead forecasts: bias is close to 0, while NMAE and NRMSE are of 8% and 12%, respectively on average, there is ± 1.68 MW between forecasts and measurements 31761 - Renewables in Electricity Markets 29

Comparing against benchmark approaches Forecasts from advanced methods are expected to outperform simple benchmarks! Two typical benchmarks are (to be further discussed in Lecture 11): Persistence ( what you see is what you get ): ŷ t+k t = y t, k = 1, 2,... Climatology (the once and for all strategy): ŷ t+k t = ȳ t, k = 1, 2,... where ȳ t is the average of all measurements available up to time t A skill score informs of the relative quality of a method vs. a relevant benchmark, for a given lead time k: where SSc(k) = 1 Sc adv(k), SSc 1 (possibly expressed in %) Sc ref (k) Sc can be MAE, RMSE, etc., Sc adv is score value for the advanced method, and Sc ref is for the benchmark 31761 - Renewables in Electricity Markets 30

Example: benchmarking at Klim Great! My forecasts are way better than the benchmarks considered (in terms of RMSE) Additional comments: persistence is difficult to outperform for short lead times the opposite holds for climatology 31761 - Renewables in Electricity Markets 31

4 Verification of probabilistic forecasts? 31761 - Renewables in Electricity Markets 32

Well... it is a bit more difficult Evaluating probabilistic forecasts is more involved than evaluating point predictions! Can you tell if this single forecast is good or not? 31761 - Renewables in Electricity Markets 33

Attributes of probabilistic forecast quality How do you want your forecasts? Reliable? (also referred to as probabilistic calibration ) Sharp? (i.e., informative) Skilled? (all-round performance, and of higher quality than some benchmark) Of high resolution? (i.e., resolving among situations with various uncertainty levels) etc. 31761 - Renewables in Electricity Markets 34

Probabilistic calibration Calibration is about respecting the probabilistic contract: for a quantile forecast ˆq (α) with nominal level α = 0.5, one expect that the t+k t observations y t+k are to be less than ˆq (α) 50% of the times t+k t for an interval forecast Î (β) with nominal coverage rate β = 0.9, one expect that the t+k t observations y t+k are to be covered by Î (β) 90% of the times t+k t further than that, since an interval forecast Î (β) is composed by two quantile t+k t forecasts with nominal levels α and α, one evaluates these two quantile forecasts finally for predictive densities ˆF t+k t, composed by a number m of quantile forecasts with nominal levels {α 0, α 1, α 2,..., α m}, all these quantile forecasts are evaluated, individually To do it in practice, we take a frequentist approach... we simply count! 31761 - Renewables in Electricity Markets 35

Assessing calibration For a given quantile forecast ˆq (α) t+k t and the corresponding observation y t+k, the indicator variable ξ (α) t,k is given by ξ (α) t,k = 1{y t+k < ˆq (α) t+k t } = { 1, if y t+k < ˆq (α) t+k t (HIT) 0, otherwise (MISS) By counting the number of hits over your set of forecasts, one obtains the empirical level of these quantile forecasts The empirical level a (α) k a (α) k where n (α) k n (α) k = n(α) k T is the sum of hits: = #{ξ (α) t,k is given by the mean of ξ (α) t,k = 1} = T t=1 ξ(α) t,k over the set of T quantile forecasts, 31761 - Renewables in Electricity Markets 36

Example calibration assessment at Klim with reliability diagrams The calibration assessment can be summarized in reliability diagrams Here example for our probabilistic forecasts at Klim: period: 1.7.2002-31.12.2002 predictive densities composed by quantile forecasts with nominal levels {0.05, 0.1,..., 0.45, 0.55,..., 0.9, 0.95} quantile forecasts are evaluated one by one, and their empirical levels are reported vs. their nominal levels The closest to the diagonal, the better! 31761 - Renewables in Electricity Markets 37

Sharpness Sharpness is about the concentration of probability A perfect probabilistic forecast gives a probability of 100% on a single value! Consequently, a sharpness assessment boils down to evaluating how tight the predictive densities are... The width of a given interval forecast Î (β) t+k t is given by the distance between its two bounds δ (beta) t,k = ˆq (α) t+k t ˆq(α) t+k t The sharpness of these interval forecasts is obtained by calculating their average width over the evaluation period: δ (beta) (k) = 1 T T t=1 δ(beta) t,k This is done for all the intervals composing the predictive densities 31761 - Renewables in Electricity Markets 38

Example: sharpness evaluation at Klim Period: 1.7.2012-31.12.2012 Predictive densities are composed by interval forecasts with nominal coverage rates β = 0.1, 0.2,..., 0.9 The intervals width increase with the lead time, reflecting higher forecast uncertainty 31761 - Renewables in Electricity Markets 39

Overall skill assessment The skill of probabilistic forecasts can be assessed by scores, like MAE and RMSE for the deterministic forecasts. The most common skill score for predictive densities is the Continuous Ranked Probability Score (CRPS) For a given predictive density ˆF t+k t and corresponding observation y t+k, CRPS t,k = y ( ) 2 ˆF t+k t (y) 1{y t+k y} dy cumulative probability [p.u.] 0.00 0.25 0.50 0.75 1.00 Observation y=0.45 0.00 0.25 0.50 0.75 1.00 power generation [p.u.] The CRPS score value is then given by taking its average for each of the predictive densities and corresponding observation over the evaluation period: CRPS(k) = 1 T T t=1 CRPS t,k 31761 - Renewables in Electricity Markets 40

Example: CRPS evaluation at Klim Period: 1.7.2012-31.12.2012 Probabilistic forecast quality also degrades with further lead times For instance, for 24-ahead forecasts, CRPS is equal to 7% of nominal capacity CRPS and MAE (for deterministic forecasts) can be directly compared... This CRPS value of 7% is better than the MAE value of 8% in the previous example for deterministic forecasts 31761 - Renewables in Electricity Markets 41

Now you should be ready to evaluate/handle forecasts in the real world! [Extra reading: Jolliffe IT, Stephenson DB (2011). Forecast Verification: A Practitioner s Guide in Atmospheric Science (2nd Ed.). Wiley (link to pdf cannot be provided)] 31761 - Renewables in Electricity Markets 42