Methods of forecast verification

Similar documents
Verification at JMA on Ensemble Prediction

4.3.2 Configuration. 4.3 Ensemble Prediction System Introduction

Standardized Verification System for Long-Range Forecasts. Simon Mason

Ensemble Verification Metrics

Verification of Probability Forecasts

Evaluating Forecast Quality

VERFICATION OF OCEAN WAVE ENSEMBLE FORECAST AT NCEP 1. Degui Cao, H.S. Chen and Hendrik Tolman

Exercise Part 1 - Producing Guidance and Verification -

Products of the JMA Ensemble Prediction System for One-month Forecast

Akira Ito & Staffs of seasonal forecast sector

JMA s Seasonal Prediction of South Asian Climate for Summer 2018

Seasonal Climate Watch September 2018 to January 2019

Tokyo Climate Center Website (TCC website) and its products -For monitoring the world climate and ocean-

Seasonal Climate Watch July to November 2018

JMA s Ensemble Prediction System for One-month and Seasonal Predictions

Drought forecasting methods Blaz Kurnik DESERT Action JRC

Seasonal Climate Watch June to October 2018

Shuhei Maeda Climate Prediction Division Global Environment and Marine Department Japan Meteorological Agency

Winter Forecast for GPC Tokyo. Shotaro TANAKA Tokyo Climate Center (TCC) Japan Meteorological Agency (JMA)

Probabilistic verification

Atmospheric circulation analysis for seasonal forecasting

How to Use the Guidance Tool (Producing Guidance and Verification)

Development of Multi-model Ensemble technique and its application Daisuke Nohara

Predicting uncertainty in forecasts of weather and climate (Also published as ECMWF Technical Memorandum No. 294)

Performance of Multi-Model Ensemble(MME) Seasonal Prediction at APEC Climate Center (APCC) During 2008

Program of the morning lectures

Model error and seasonal forecasting

Extracting probabilistic severe weather guidance from convection-allowing model forecasts. Ryan Sobash 4 December 2009 Convection/NWP Seminar Series

Wassila Mamadou Thiaw Climate Prediction Center

Towards Operational Probabilistic Precipitation Forecast

Verification of Continuous Forecasts

WMO Lead Centre activities for global sub-seasonal MME prediction

Model verification / validation A distributions-oriented approach

Monthly probabilistic drought forecasting using the ECMWF Ensemble system

How to use the guidance tool (Producing Guidance and Verification)

Interpretation of Outputs from Numerical Prediction System

Developing Operational MME Forecasts for Subseasonal Timescales

Seasonal Prediction, based on Canadian Seasonal to Interannual Prediction system (CanSIPS) for the Fifth South West Indian Ocean Climate Outlook Forum

The Impact of Horizontal Resolution and Ensemble Size on Probabilistic Forecasts of Precipitation by the ECMWF EPS

ECMWF 10 th workshop on Meteorological Operational Systems

DROUGHT FORECASTING IN CROATIA USING STANDARDIZED PRECIPITATION INDEX (SPI)

Seasonal Climate Watch April to August 2018

Five years of limited-area ensemble activities at ARPA-SIM: the COSMO-LEPS system

A simple method for seamless verification applied to precipitation hindcasts from two global models

Probabilistic seasonal forecast verification

Extended-range/Monthly Predictions. WGSIP, Trieste

S e a s o n a l F o r e c a s t i n g f o r t h e E u r o p e a n e n e r g y s e c t o r

EMC Probabilistic Forecast Verification for Sub-season Scales

Application and verification of the ECMWF products Report 2007

Categorical Verification

Severe storm forecast guidance based on explicit identification of convective phenomena in WRF-model forecasts

Seasonal Hydrometeorological Ensemble Prediction System: Forecast of Irrigation Potentials in Denmark

Downscaling of ECMWF seasonal integrations by RegCM

statistical methods for tailoring seasonal climate forecasts Andrew W. Robertson, IRI

Background of Symposium/Workshop Yuhei Takaya Climate Prediction Division Japan Meteorological Agency

TCC Recent Development and Activity

Environment and Climate Change Canada / GPC Montreal

Recent Developments in Climate Information Services at JMA. Koichi Kurihara Climate Prediction Division, Japan Meteorological Agency

Feature-specific verification of ensemble forecasts

Ensemble forecast and verification of low level wind shear by the NCEP SREF system

What is one-month forecast guidance?

WMO LC-LRFMME Website User Manual

Standardized Anomaly Model Output Statistics Over Complex Terrain.

Forecasting precipitation for hydroelectric power management: how to exploit GCM s seasonal ensemble forecasts

Verification of the operational NWP models at DWD - with special focus at COSMO-EU

Monthly forecast and the Summer 2003 heat wave over Europe: a case study

Spatial forecast verification

Basic Verification Concepts

Assessment of Ensemble Forecasts

Verification of the Seasonal Forecast for the 2005/06 Winter

NOTES AND CORRESPONDENCE. Improving Week-2 Forecasts with Multimodel Reforecast Ensembles

TCC Training Seminar on 17 th Nov 2015 JMA s Ensemble Prediction Systems (EPSs) and their Products for Climate Forecast.

3.6 NCEP s Global Icing Ensemble Prediction and Evaluation

Assessing high resolution forecasts using fuzzy verification methods

Seasonal Predictions for South Caucasus and Armenia

An extended re-forecast set for ECMWF system 4. in the context of EUROSIP

Climate Outlook for Pacific Islands for July December 2017

Verification of nowcasts and short-range forecasts, including aviation weather

The benefits and developments in ensemble wind forecasting

Verification statistics and evaluations of ECMWF forecasts in

QUANTIFYING THE ECONOMIC VALUE OF WEATHER FORECASTS: REVIEW OF METHODS AND RESULTS

The 2010/11 drought in the Horn of Africa: Monitoring and forecasts using ECMWF products

Detective Miss Marple in Murder at the Gallop. 3rd verification workshop, ECMWF 2007, Göber

Climate Outlook for Pacific Islands for December 2017 May 2018

Introduction of Seasonal Forecast Guidance. TCC Training Seminar on Seasonal Prediction Products November 2013

Model verification and tools. C. Zingerle ZAMG

Basic Verification Concepts

Focus on parameter variation results

Application and verification of ECMWF products: 2010

Global climate predictions: forecast drift and bias adjustment issues

Tokyo Climate Center s activities as RCC Tokyo

REPORT ON APPLICATIONS OF EPS FOR SEVERE WEATHER FORECASTING

Verification of WAQUA/DCSMv5 s operational water level probability forecasts

JP3.7 SHORT-RANGE ENSEMBLE PRECIPITATION FORECASTS FOR NWS ADVANCED HYDROLOGIC PREDICTION SERVICES (AHPS): PARAMETER ESTIMATION ISSUES

Exploring ensemble forecast calibration issues using reforecast data sets

Experimental MOS Precipitation Type Guidance from the ECMWF Model

Multi-Model Ensembles in NWS Climate Prediction Center Subseasonal to Seasonal Forecasts: Metrics for Impact Events

1.10 A NEW SPATIAL SCALE DECOMPOSITION OF THE BRIER SCORE FOR THE VERIFICATION OF PROBABILISTIC LIGHTNING FORECASTS. Abstract

Upgrade of JMA s Typhoon Ensemble Prediction System

Probabilistic Forecast Verification. Yuejian Zhu EMC/NCEP/NOAA

Transcription:

Methods of forecast verification Kiyotoshi Takahashi Climate Prediction Division Japan Meteorological Agency 1

Outline 1. Purposes of verification 2. Verification methods For deterministic forecasts For probabilistic forecasts Standardised Verification System for Long-Range Forecasts (SVSLRF) 3. Results on the TCC web page 2

1. Purposes of verification Forecast verification is a process of assessing quality of forecasts. to monitor forecast quality how accurate are forecasts and are they improving? to guide forecasters and users help forecasters understand model characteristics help us provide higher value forecasts to users to guide future developments of system identify model faults and improve systems help us compare and evaluate different forecasts (model or guidance) 3

2. Verification methods For deterministic forecast ACOR, RMSE, Bias (ME),MSSS (ROC) For probabilistic forecast ROC, Reliability diagram, BSS (Brel, Bres) 4

Root Mean Square Error (RMSE) 1 RMSE N N F i O i i1 Range: 0 to infinity, Perfect score: 0. 2 F : forecast O : observation N :sample size RMSE measures absolute magnitude of the forecast error. It does not indicate the direction the error. RMSE small RMSE large 5

Mean Error (ME) ME 1 N N F i O i i1 F : forecast O : observation N :sample size Range: variable, Perfect score: 0. ME measures average magnitude of the forecast error. It indicates the direction the error. RMSE 2 2 1 N N i 1 2 2 ME e e xi ai ME RMSE can be divided into ME(systematic error) and random error ( ). e 2 6

Anomaly Correlation (AC) Range: -1 to 1. Perfect score: 1. AC measures correspondence or phase difference between forecast and observation, subtracting out the climatological mean at each point. N i i i N i i i N i i i i i C O C F C O C F AC 1 2 1 2 1 ) ( ) ( ) )( ( climatology observation forecast : : : C O F 7 AC <<1 AC 1

AC positive ( 1) RMSE large AC negative RMSE large AC positive( 1) RMSE small AC negative RMSE small 8

Mean Squared Skill Score (MSSS) MSSS 1 MSE MSE c Perfect score: 1 (when MSE=0) Climatology forecast score: 0 where MSE is the mean squared error MSE 1 N N F i O i i1 2 F : forecast O : observation and MSE c is the MSE of climatology forecast. MSSS can be expanded (Murphy,1988) as 2 2 s f s f f o MSSS 2 r fo 2 so so so 2n 1 2n 1 1 2 n 1 n 1 s : squared root of variance r : correlation f : forecast o : observation 1 2 3 n 1 Fi F O i O n i1 rfo The first 3 terms are related to S f SO 1 phase error (through the correlation) 2amplitude errors (through the ratio of the forecast to observed variances) 3bias error 9

Reliability diagram The reliability diagram plots the observed frequency(y-axis) against the forecast probability(x-axis). The diagonal line indicates perfect reliability (observed frequency equal to forecast probability for each category). climatology perfect reliability Brier Scores forecast frequency Points below(above) the diagonal line indicate overforecasting (underforecasting). 10

BS 1 N Brier (skill) score Brier score measures mean squared error of the probability forecasts. N i1 ( p i o i 2 ) Range: 0 to 1. Perfect score: 0 Random:1/3 p o i i N : forecast : observed : sample size probability Brier skill score measures skill relative to a reference forecast (usually climatology). BSS 1 BS BS reference Climatology: occurrence(0or 1) o( 1 o) Range: minus infinity to 1. BSS=0 indicates no skill when compared to the reference forecast. Perfect score: 1. 11

Decomposition of the Brier score Murphy(1973) showed that the Brier score could be partitioned into three terms(for K probability classes and N samples). These terms are shown separately to attribute sources of error. o :climatological occurrence BS 1 K K 1 2 n ( p o ) n ( o o) 2 k k k k k N N k1 k1 o(1 o) reliability (brel) resolution (bres) uncertainty (bunc) the mean squared difference between the forecast probability and the observed frequency. Perfect score: 0 the mean squared difference between the observed frequency and climatological frequency. - indicates the degree to which the forecast can separete different situations. climatologial forecast score:0 Perfect score: o( 1 o) measures the variability of the observations. 12

Brier skill score Brier skill score = the relative skill of the probabilistic forecast to the climatology BSS BS BS BS BS clim 1 0 BSc lim clim Reliability skill score brel Brel 1 BS c lim Skill score BS c bunc lim score score Range: minus infinity to 1. Perfect score: 1 BSS=0 indicates no skill when compared to the climatology. BSS>0 : better than clim. Resolution skill score Bres bres BS clim Perfect score: 1 Perfect score: 1 forecast perfect forecast score score reference reference Score 100 The larger these skill scores are, the better. 13

Interpretation of Reliability diagram and BSS Event : Z500 Anomaly > 0 Northern Hemisphere Spring of 2008 (2008/2/28 ~2008/5/29) First week forecast (day 2-8) Third and fourth week (day 16-29) overforecasting underforecasting BSS>0 better than climatology BSS<0 inferior to climatology 14

Relative Operating Characteristic (ROC) ROC is created by plotting the hit rate(y-axis) against the false alarm rate(x-axis) using increasing probability thresholds to make the yes/no decision. The area under the ROC curve (=ROC area) is frequently used as a score. 15

Forecast Steps for making ROC diagram 1. For each forecast probability category, count the number of hits,misses, false alarms, and correct non-events 2. Compute the hit rate and false alarm rate in each category k hit rate k = hits k / (hits k + misses k ) false alarm rate k = false alarms k / (false alarms k + correct non-events k ) 3. Plot hit rate vs false alarm rate 4. ROC area is the integrated area under the ROC curve yes no yes hits false alarms no misses correct nonevents total Observed Observed yes Observed no Contingency Table 16

Interpretation of ROC curves Event: Z500 anomaly > 0 Spring 2008 1st week forecast 3rd and 4th week forecast Perfect performance high resolution (high potential skill) low resolution (low potential skill) ROC is not sensitive to bias in forecasts. Forecasts with bias may have a good ROC curve if resolution is still good. In this sense, the ROC can be considered as a measure of potential usefulness. On the other hand, reliability diagram is sensitive to the bias. It is needed to see both the ROC and the reliability diagram. 17

SVSLRF (Standard Verification System for Long-Range Forecast) WMO standard tool to verify skill in seasonal models It was introduced by the Commission for Basic Systems (CBS) of the World Meteorological Organization (WMO) in December, 2002. Users can appropriately evaluate forecast skill with common measures. 18

Outline of SVSLRF Mandatory part Parameters Verification regions Deterministic forecasts Probabilistic forecasts Level 1 T2m anomaly Tropics(20S-20N) MSSS ROC curves Precipitation anomaly (Nino3.4 Index) Northern extratropics(20n-90n) Southern extratropics(20s-90s) (N/A) ROC areas Reliability diagrams Frequency histograms Level 2 T2m anomaly Precipitation anomaly (SST anomaly) Grid-point verification on a 2.5 by 2.5 grid MSSS and its three-term decomposition at each gridpoint ROC areas at each grid-point Level 3 T2m anomaly Precipitation anomaly (SST anomaly) Grid-point verification on a 2.5 by 2.5 grid 3 by 3 contingency tables at each grid-point ROC reliability tables at each grid-point 19

LC-SVSLRF Lead Centre for the Long-Range Forecast Verification System Australian Bureau of Meteorology (BOM) Meteorological Service of Canada (MSC) URL: http://www.bom.gov.au/wmo/lrfvs/ The Manual is here Lead Center web site You can get scores of several centers on Lead Centre website. 20

3. Verification results on TCC web page 1-month forecast 3-month forecast Warm/Cold season forecast 21

Verification of operational 1-month forecasts Deterministic forecast Ensemble mean forecast error maps, RMSE and Anomaly Correlation Probabilistic forecast Time sequence of RMSE and AC 22

Verification of 1-month Ensemble mean forecast maps (Deterministic) Z500 over the Northern Hemisphere Observation Forecast Error Stream function (850, 200hPa) RMSE and Anomaly Corr. 23 23

Verification of 1-month Ensemble mean forecast maps (Deterministic) Z500 over the Northern Hemisphere Observation Forecast Error ANL-Clim (JRA-25/JCDAS) FCT-Clim (model) 24

Verification of operational 1-month forecasts Probabilistic forecast Reliability diagrams and Brier skill scores ROC curves and area for each season 25

Verification of hindcast based on SVSLRF 26

Hindcast verification methods based on Standardised Verification System for Long-Range Forecasts (SVSLRF) Verification of deterministic forecasts Mean Square Skill Score (MSSS) Contingency tables Verification of probabilistic forecasts Reliability diagrams ROC curves and ROC areas 27

Examples of MSSS for precipitation Dec-Feb started on 10 Nov. Jun-Aug started on 10 May. Positive MSSS indicates that the forecast is better than climatological forecast. 28

Deterministic Summary Index Random Climatology Perfect RMSE >0 0 ME 0 AC -1 +1 MSSS 0 +1 MSE >0 >0 0 Probabilistic Index Random Climatology Perfect Reliability diagram o( 1 o) Fit to the diagonal line BS 1/3 bunc= 0 BSS(x100) 100> 0 +100 Brel(x100) 100> 100> +100 Bres(x100) 100> 0 +100 Roc area(x100) 100> 50 +100 29

References Murphy, A.H., 1973: A new vector partition of the probability score. J. Appl. Meteor., 12, 595-600. Murphy, A.H., 1988: Skill scores based on the mean square error and their relationships to the correlation coefficient. Mon. Wea. Rev., 16, 2417-2424. http://www.bom.gov.au/wmo/lrfvs/index.html http://www.ecmwf.int/newsevents/meetings/workshops/2007/jwgv/index.html 30