Evaluating Forecast Quality

Similar documents
Ensemble Verification Metrics

Verification of Probability Forecasts

Methods of forecast verification

Towards Operational Probabilistic Precipitation Forecast

Predicting uncertainty in forecasts of weather and climate (Also published as ECMWF Technical Memorandum No. 294)

Probabilistic verification

Categorical Verification

Verification of African RCOF Forecasts

VERFICATION OF OCEAN WAVE ENSEMBLE FORECAST AT NCEP 1. Degui Cao, H.S. Chen and Hendrik Tolman

QUANTIFYING THE ECONOMIC VALUE OF WEATHER FORECASTS: REVIEW OF METHODS AND RESULTS

Probabilistic seasonal forecast verification

Verification of Renewable Energy Forecasts

Model verification / validation A distributions-oriented approach

Drought forecasting methods Blaz Kurnik DESERT Action JRC

Global Flash Flood Forecasting from the ECMWF Ensemble

REPORT ON APPLICATIONS OF EPS FOR SEVERE WEATHER FORECASTING

Feature-specific verification of ensemble forecasts

Florida State University Libraries

Validation of Forecasts (Forecast Verification) Overview. Ian Jolliffe

Complimentary assessment of forecast performance with climatological approaches

The Impact of Horizontal Resolution and Ensemble Size on Probabilistic Forecasts of Precipitation by the ECMWF EPS

Monthly probabilistic drought forecasting using the ECMWF Ensemble system

Enhancing Weather Information with Probability Forecasts. An Information Statement of the American Meteorological Society

Predictability from a Forecast Provider s Perspective

Application and verification of the ECMWF products Report 2007

Basic Verification Concepts

Forecast Verification Analysis of Rainfall for Southern Districts of Tamil Nadu, India

Extracting probabilistic severe weather guidance from convection-allowing model forecasts. Ryan Sobash 4 December 2009 Convection/NWP Seminar Series

Monthly forecast and the Summer 2003 heat wave over Europe: a case study

Deterministic and Probabilistic prediction approaches in Seasonal to Inter-annual climate forecasting

Verification of nowcasts and short-range forecasts, including aviation weather

4.3.2 Configuration. 4.3 Ensemble Prediction System Introduction

Probabilistic Weather Forecasting and the EPS at ECMWF

North Carolina State University Land Grant University founded in undergraduate fields of study 80 master's degree areas 51 doctoral degree

Seasonal Climate Watch July to November 2018

Standardized Verification System for Long-Range Forecasts. Simon Mason

Verification of ensemble and probability forecasts

Detective Miss Marple in Murder at the Gallop. 3rd verification workshop, ECMWF 2007, Göber

Verification of Weather Warnings

Verification of WAQUA/DCSMv5 s operational water level probability forecasts

Basic Verification Concepts

Seasonal Climate Watch January to May 2016

Severe storm forecast guidance based on explicit identification of convective phenomena in WRF-model forecasts

Upscaled and fuzzy probabilistic forecasts: verification results

statistical methods for tailoring seasonal climate forecasts Andrew W. Robertson, IRI

Application and verification of ECMWF products: 2010

Verification of the operational NWP models at DWD - with special focus at COSMO-EU

The ECMWF Extended range forecasts

Verification of ECMWF products at the Deutscher Wetterdienst (DWD)

The Latest Science of Seasonal Climate Forecasting

Climate Forecasts and Forecast Uncertainty

Evaluation of Ensemble Icing Probability Forecasts in NCEP s SREF, VSREF and NARRE-TL Systems

Verification and performance measures of Meteorological Services to Air Traffic Management (MSTA)

ECMWF 10 th workshop on Meteorological Operational Systems

DROUGHT FORECASTING IN CROATIA USING STANDARDIZED PRECIPITATION INDEX (SPI)

Performance evaluation of binary classifiers

A Local Ensemble Prediction System for Fog and Low Clouds: Construction, Bayesian Model Averaging Calibration, and Validation

Seasonal Climate Watch June to October 2018

Accounting for the effect of observation errors on verification of MOGREPS

Radius of reliability: A distance metric for interpreting and verifying spatial probability forecasts

PRICING AND PROBABILITY DISTRIBUTIONS OF ATMOSPHERIC VARIABLES

Focus on parameter variation results

Seasonal Climate Watch September 2018 to January 2019

Introduction to Artificial Intelligence. Prof. Inkyu Moon Dept. of Robotics Engineering, DGIST

Model error and seasonal forecasting

Guidance on Verification of Operational Seasonal Climate Forecasts

Proper Scores for Probability Forecasts Can Never Be Equitable

Verification Methods for High Resolution Model Forecasts

Jaclyn Ashley Shafer. Bachelor of Science Meteorology Florida Institute of Technology 2006

ENSO-DRIVEN PREDICTABILITY OF TROPICAL DRY AUTUMNS USING THE SEASONAL ENSEMBLES MULTIMODEL

Verification of ECMWF products at the Finnish Meteorological Institute

Ensemble forecast and verification of low level wind shear by the NCEP SREF system

Precipitation verification. Thanks to CMC, CPTEC, DWD, ECMWF, JMA, MF, NCEP, NRL, RHMC, UKMO

Communicating uncertainty from short-term to seasonal forecasting

Probabilistic Quantitative Precipitation Forecasts for Tropical Cyclone Rainfall

Benchmark study of Icing Forecasts Do they add value? Ben Martinez, Pierre Julien Trombe Vattenfall R&D

Work on on Seasonal Forecasting at at INM. Dynamical Downscaling of of System 3 And of of ENSEMBLE Global Integrations.

Coefficients for Debiasing Forecasts

The COLA Anomaly Coupled Model: Ensemble ENSO Prediction

Assessment of Ensemble Forecasts

Ensemble based first guess support towards a risk-based severe weather warning service

The Analysis of the Thunderstorm Forecasts on the Friuli-Venezia Giulia Plain

Using time-lag ensemble techniques to assess behaviour of high-resolution precipitation forecasts

Spatial forecast verification

3.6 NCEP s Global Icing Ensemble Prediction and Evaluation

Revisiting predictability of the strongest storms that have hit France over the past 32 years.

Challenges of Communicating Weather Information to the Public. Sam Lashley Senior Meteorologist National Weather Service Northern Indiana Office

Spatial verification of NWP model fields. Beth Ebert BMRC, Australia

Model verification and tools. C. Zingerle ZAMG

Adaptation for global application of calibration and downscaling methods of medium range ensemble weather forecasts

S e a s o n a l F o r e c a s t i n g f o r t h e E u r o p e a n e n e r g y s e c t o r

Analysis and accuracy of the Weather Research and Forecasting Model (WRF) for Climate Change Prediction in Thailand

I.Classical probabilities

Update on CoSPA Storm Forecasts

Guidance on Verification of Operational Seasonal Climate Forecasts

Denver International Airport MDSS Demonstration Verification Report for the Season

Simulating the Value of El Niño Forecasts for the Panama Canal. Nicholas Graham and Konstantine Georgakakos. Carlos Vargas and Modesto Echevers

Application and verification of ECMWF products 2015

Developing Operational MME Forecasts for Subseasonal Timescales

Focus on Spatial Verification Filtering techniques. Flora Gofa

Transcription:

Evaluating Forecast Quality Simon J. Mason International Research Institute for Climate Prediction

Questions How do we decide whether a forecast was correct? How do we decide whether a set of forecasts is correct consistently enough to be considered good? How can we answer any of these questions when forecasts are expressed probabilistically?

How do we decide whether a forecast was correct? Discrete Continuous Deterministic Probabilistic It will rain tomorrow There is a 50% chance of rain tomorrow There will be 10 mm of rain tomorrow There is a 50% chance of more than 10 mm of rain tomorrow

The method of forecast verification depends upon the type of information provided in the forecast: Deterministic - discrete Forecast: Verification: It will rain tomorrow. Does it rain? Yes or no? Deterministic - continuous Forecast: Verification: There will be 10 mm of rain tomorrow. Does it rain (approximately) 10 mm?

Probabilistic discrete and continuous Forecast: Verification: There is a 50% chance of (more than 10 mm of) rain tomorrow. Yes! As long as what happened was not given a 0% chance of occurring, a probabilistic forecast cannot be wrong.

Conclusion Prediction is very difficult, especially if it s about the future. Nils Bohr No probabilistic forecasting is very easy because the forecast is never wrong!?

How do we decide whether a set of forecasts is correct consistently enough to be considered good? Deterministic - discrete FORECASTS OBSERVATIONS Yes No Yes Hit Miss No False alarm Correct Rejection

Hit Rate How many of the events were forecast? number of hits Hit rate = 100% number of events Wall Street indices predicted nine out of the last five recessions! Newsweek, 19 Sept., 1966. Correct rejections are correct forecasts. False-alarm rate = number of false alarms number of nonevents

Hit Score How many times was the forecast correct? number of correct forecasts Hit score = 100% number of forecasts number of hits and correct rejections Hit score = 100% number of forecasts A measure of forecast accuracy.

Finley s Tornado Forecasts A set of tornado forecasts for the U.S. Midwest published in 1884. FORECASTS OBSERVATIONS Tornado No tornado Total Tornado 28 23 51 No tornado 72 2680 2752 Total 100 2703 2803 28+ 2680 Hit score = 100% = 96.6% 2803

No Tornado Forecasts A better score can be achieved by issuing no forecasts of tornadoes! FORECASTS OBSERVATIONS Tornado No tornado Total Tornado 0 51 51 No tornado 0 2752 2752 Total 0 2803 2803 0+ 2752 Hit score = 100% = 98.2% 2803

Problems Accuracy: The forecasts correspond well with the observations, but: Uncertainty: Tornadoes are rare events so it is easy to score a correct rejection. There is very little uncertainty. Skill: The accuracy of one set of forecasts compared to that of another set.

How do we decide whether a set of forecasts is good? Is this set of forecasts better than another (simple) set? Often the comparison is with a simple forecast strategy such as: climatology; persistence; perpetual forecasts; random forecasts.

Hit Skill Score How many more times was the forecast correct compared to a reference forecast strategy? # correct - # expected correct Hit skill score = 100% # forecasts - # expected correct ( ) 28+ 2680 2752 Hit skill score = 100% = 86. 3% 2803-2752 Exercise: calculate the hit skill score for Finley s forecasts compared to a strategy of random guessing.

How do we decide whether a set of forecasts is correct consistently enough to be considered good? Deterministic continuous Correlation A measure of association e.g., does rainfall increase if forecast rainfall increases?

A set of perfect forecasts!

Mean squared error mean squared error = total of squared errors number of forecasts A measure of forecast accuracy. Composed of: correlation; mean bias; amplitude bias. Can be converted to a skill score.

How do we decide whether one set of probabilistic forecasts is better than another? Brier score Measures the mean-squared error of probability forecasts. Brier score = total of squared probabilit y errors number of forecasts If an event was forecast with a probability of 60%, and the event occurred, the probability error is: 60% - 100% = -40%

Verification of probabilistic forecasts How do we know if a probabilistic forecast was correct? A probabilistic forecast can never be wrong! As soon as a forecast is expressed probabilistically, all possible outcomes are forecast. However, the forecaster s level of confidence can be correct or incorrect = reliable. Is the forecaster over- / under-confident?

Verification of probabilistic forecasts How do we know if a forecaster is over- / underconfident Whenever a forecaster says there is a high probability of rain tomorrow, it should rain more frequently than when the forecaster says there is a low probability of rain.

Forecast reliability A forecast is consistent if the forecast probability is a true estimate of the forecaster s level of confidence. If forecasts are reliable, the forecaster s confidence is appropriate. If forecasts are consistent and reliable, the probability that the event will occur is the same as the forecast probability.

Desired characteristics of forecasts: Probabilities should be reliable. Probabilities should be sharp. Reliability is a function of forecast accuracy. Assuming the forecasts are reliable, sharpness is a function of predictability.

Reliability Diagrams For all forecasts of a given confidence, identify how often the event occurs. If the proportion of times that the event occurs is the same as the forecast probability, the probabilities are reliable (or well calibrated). A plot of relative frequency of occurrence against forecast probability will be a diagonal line if the forecasts are reliable. Problem: large number of forecasts required.

Reliability Diagrams Reliability diagrams for forecasts of Nino3.4 sea-surface temperatures.

Reliability Diagrams Exercise: diagnose the following reliability curves.

Relative (or Receiver) Operating Characteristics (ROC) Convert probabilistic forecasts to deterministic forecasts by issuing a warning if the probability exceeds a threshold minimum. By raising the threshold less warnings are likely to be issued - reducing the potential of issuing a false alarm, but increasing the potential of a miss. By lowering the threshold more warnings are likely to be issued - reducing the potential of a miss, but increasing the potential of a false alarm.

Relative (or Receiver) Operating Characteristic (ROC) Are (proportionately) more warnings issued for events than for non-events? Plot the hit rate against the false alarm rate for varying thresholds. The ROC is used to estimate whether forecasts are potentially useful.

Hit rate Defines: the proportion of events for which a warning was provided correctly Estimates: the probability that an event will be forewarned hit rate = number of hits number of events

False alarm rate Defines: the proportion of non-events for which a warning was provided incorrectly Estimates: the probability that a warning will be provided incorrectly for a non-event false alarm rate = number of false alarms number of nonevents

Relative (or Receiver) Operating Characteristics (ROC) ROC curves for ECHAM 3.6 simulations of March-May belowand above-normal precipitation over eastern Africa (10-10 S, 30-50 E)

Conclusions Even for deterministic forecasts, there is no single measure that gives a comprehensive summary of forecast quality: - accuracy - skill - uncertainty Probabilistic forecasts address the two fundamental questions: - What is going to happen? - How confident can we be that it is going to happen? Both these aspects require verification.

Conclusions The most important aspects of forecast quality: - the most likely outcome must be the one that occurs most frequently; - confidence in the forecast must be appropriate (reliability); - forecast probabilities should be sharp (without compromising reliability). Good forecasts must address: - consistency - quality - value

Recommended readings Forecast verification and quality: Murphy, A. H., 1991: Forecast verification: its complexity and dimensionality. Monthly Weather Review, 105, 803 816. Murphy, A. H., 1993: What is a good forecast? An essay on the nature of goodness in weather forecasting. Weather and Forecasting, 8, 281 293. Wilks, D. S., 1995: Statistical Methods in the Atmospheric Sciences, Academic Press, San Diego. Chapter 7, Forecast verification, pp 233 283.