Current verification practices with a particular focus on dust

Similar documents
Spatial Forecast Verification Methods

Spatial forecast verification

Spatial verification of NWP model fields. Beth Ebert BMRC, Australia

Verification of ensemble and probability forecasts

Basic Verification Concepts

Verification Methods for High Resolution Model Forecasts

Basic Verification Concepts

Implementation of global surface index at the Met Office. Submitted by Marion Mittermaier. Summary and purpose of document

Generating probabilistic forecasts from convectionpermitting. Nigel Roberts

Feature-specific verification of ensemble forecasts

Operational convective scale NWP in the Met Office

Using time-lag ensemble techniques to assess behaviour of high-resolution precipitation forecasts

Verification and performance measures of Meteorological Services to Air Traffic Management (MSTA)

The Nowcasting Demonstration Project for London 2012

COMPOSITE-BASED VERIFICATION OF PRECIPITATION FORECASTS FROM A MESOSCALE MODEL

Verification Marion Mittermaier

Identifying skilful spatial scales using the Fractions Skill Score

Xinhua Liu National Meteorological Center (NMC) of China Meteorological Administration (CMA)

Update on CoSPA Storm Forecasts

Nesting and LBCs, Predictability and EPS

A statistical approach for rainfall confidence estimation using MSG-SEVIRI observations

Intercomparison of spatial forecast verification methods: A review and new project launch

Verification of nowcasts and short-range forecasts, including aviation weather

Model enhancement & delivery plans, RAI

Focus on Spatial Verification Filtering techniques. Flora Gofa

THE MODEL EVALUATION TOOLS (MET): COMMUNITY TOOLS FOR FORECAST EVALUATION

WMO Aeronautical Meteorology Scientific Conference 2017

Representing model error in the Met Office convection permitting ensemble prediction system

Precipitation verification. Thanks to CMC, CPTEC, DWD, ECMWF, JMA, MF, NCEP, NRL, RHMC, UKMO

Verification of Probability Forecasts

Medium-range Ensemble Forecasts at the Met Office

Comparison of the NCEP and DTC Verification Software Packages

Description of the case study

Merging Rain-Gauge and Radar Data

10A.1 The Model Evaluation Tool

Concepts of Forecast Verification FROST-14

Scatterometer Wind Assimilation at the Met Office

Global Flash Flood Forecasting from the ECMWF Ensemble

Analysis and accuracy of the Weather Research and Forecasting Model (WRF) for Climate Change Prediction in Thailand

Extracting probabilistic severe weather guidance from convection-allowing model forecasts. Ryan Sobash 4 December 2009 Convection/NWP Seminar Series

SAFNWC/MSG Dust flag.

The Impact of Horizontal Resolution and Ensemble Size on Probabilistic Forecasts of Precipitation by the ECMWF EPS

Convective-scale data assimilation at the UK Met Office

Probabilistic Weather Forecasting and the EPS at ECMWF

Verification of the operational NWP models at DWD - with special focus at COSMO-EU

ICAM conference 6 June 2013 Kranjska Gora (SLO) Objective forecast verification of WRF compared to ALARO and the derived INCA-FVG outputs

Severe storm forecast guidance based on explicit identification of convective phenomena in WRF-model forecasts

Application and verification of ECMWF products 2013

Application and verification of ECMWF products 2010

Recent advances in Tropical Cyclone prediction using ensembles

Application and verification of ECMWF products in Croatia - July 2007

Standardized Verification System for Long-Range Forecasts. Simon Mason

Object-oriented verification of WRF forecasts from 2005 SPC/NSSL Spring Program

Developments towards multi-model based forecast product generation

Validation Report for Precipitation products from Cloud Physical Properties (PPh-PGE14: PCPh v1.0 & CRPh v1.0)

EUMETSAT Hydrological SAF H05 product development at CNMCA

ECMWF products to represent, quantify and communicate forecast uncertainty

Application and verification of ECMWF products in Norway 2008

Application and verification of the ECMWF products Report 2007

Categorical Verification

Predicting rainfall using ensemble forecasts

Verification of Low Clouds and Fog using a Spatial Verification Method. Gema Morales and Javier Calvo NWP Area AEMET

Application and verification of ECMWF products: 2010

The ECMWF Extended range forecasts

Ensemble Verification Metrics

Satellite Radiance Data Assimilation at the Met Office

Current and future configurations of MOGREPS-UK. Susanna Hagelin EWGLAM/SRNWP, Rome, 4 Oct 2016

Application and verification of ECMWF products 2016

Observations needed for verification of additional forecast products

MESO-NH cloud forecast verification with satellite observation

The development of a Kriging based Gauge and Radar merged product for real-time rainfall accumulation estimation

Verifying Ensemble Forecasts Using A Neighborhood Approach

Denver International Airport MDSS Demonstration Verification Report for the Season

A new mesoscale NWP system for Australia

Methods of forecast verification

1 st FORALPS CONFERENCE

TAF and TREND Verification

NEW SCHEME TO IMPROVE THE DETECTION OF RAINY CLOUDS IN PUERTO RICO

Direct assimilation of all-sky microwave radiances at ECMWF

YOPP archive: needs of the verification community

Convective-scale NWP for Singapore

Verification of the operational NWP models at DWD - with special focus at COSMO-EU. Ulrich Damrath

Assessing high resolution forecasts using fuzzy verification methods

Detective Miss Marple in Murder at the Gallop. 3rd verification workshop, ECMWF 2007, Göber

Global NWP Index documentation

Cb-LIKE: thunderstorm forecasts up to 6 hrs with fuzzy logic

Satellite-Based Detection of Fog and Very Low Stratus

LATE REQUEST FOR A SPECIAL PROJECT

2012 and changes to the Rapid Refresh and HRRR weather forecast models

Verification of wind forecasts of ramping events

Improving real time observation and nowcasting RDT. E de Coning, M Gijben, B Maseko and L van Hemert Nowcasting and Very Short Range Forecasting

Remote Ground based observations Merging Method For Visibility and Cloud Ceiling Assessment During the Night Using Data Mining Algorithms

Complimentary assessment of forecast performance with climatological approaches

Spatial Verification. for Ensemble. at DWD

Assimilation of MSG visible and near-infrared reflectivity in KENDA/COSMO

9.4. Jennifer L. Mahoney NOAA Research-Forecast Systems Laboratory, Boulder, Colorado

Performance of a Probabilistic Cloud-to-Ground Lightning Prediction Algorithm

Evaluating Forecast Quality

Towards Operational Probabilistic Precipitation Forecast

Upgrade of JMA s Typhoon Ensemble Prediction System

Transcription:

Current verification practices with a particular focus on dust Marion Mittermaier and Ric Crocker

Outline 1. Guide to developing verification studies 2. Observations at the root of it all 3. Grid-to-point, point-to-grid, grid-to-grid? Traditional verification and the double penalty effect 4. When and why do we need new spatial methods? 5. The appeal of evaluating synoptic evolution object-based methods come into their own 6. Dust!

Simple guide for developing verification studies 1. Consider the users of the forecasts of the verification information 2. What aspects of forecast quality are of interest for the user? 3. Develop verification questions to evaluate those aspects/attributes training notes

Simple guide for developing verification studies 2 4. Identify observations that represent the event being forecast, including the Element (e.g., temperature, precipitation) Temporal resolution Spatial resolution and representation Thresholds, categories, etc. 5. Identify multiple verification attributes that can provide answers to the questions of interest 6. Select measures and graphics that appropriately measure and represent the attributes of interest 7. Identify a standard of comparison that provides a reference level of skill (e.g., persistence, climatology, old model) training notes

Matching forecasts and observations May be the most difficult (and time consuming) part of the verification process! Many factors need to be taken into account, e.g. Identifying observations that represent the forecast event Example: Precipitation accumulation over an hour at a point For a gridded forecast there are many options for the matching process Point-to-grid Match obs to closest gridpoint Grid-to-point Interpolate? Take largest value? Grid-to-grid different resolutions, use area-weighted averaging? training notes

Role of observations Essential for verification, but need to be treated with respect. QC is important! Forecasts need to be well posed to facilitate matching with observations. Observations need to be appropriate to capture the events of interest. Observational uncertainty should be taken into account in whatever way possible. Error/uncertainty sources Biases in frequency or value Instrument error Random error or noise Reporting errors Reporting of errors Subjective obs (e.g., STORM data) Representativeness error Precision error Conversion error Analysis error Forecast error

F The double penalty Consider forecasts and observations O F O of some dichotomous field on a grid: O F O F O F O Closeness not rewarded Detail is penalised unless exactly correct O F - higher resolution is more detailed! CSI = 0 for first 4; CSI > 0 for the 5th CSI hits hits false alarms misses training notes

Traditional v neighbourhood verification Make use of spatial verification methods which compare single observations to a forecast neighbourhood around the observation location. Represents a fundamental departure from our current verification system strategy where the emphasis is on extracting the nearest GP or bilinear interpolation to get matched forecast-ob pair.

Matching forecasts and observations Matching approach can impact verification results and interpretation Impact of land-sea points? On biases? Point-to-Grid Grid-to-Point training notes

Matching forecasts and Example: Two approaches: observations 2 Match rain gauge to nearest gridpoint or Interpolate grid values to rain gauge location Crude assumption: equal weight to each gridpoint Differences in results associated with matching: representativeness error 20 20 20 20 10 10 0 20 0 20 Obs=10 Fcst=0 Obs=10 Fcst=15 training notes

Forecast accuracy: the relative index 36-h comparison (length of 1.5 km UKV forecast) with global model (17/25 km) of nearest model grid point over the UK. Value added by km-scale NWP. Relative nature takes account of weather dependence. Both series are subject to upgrades so moving goal posts, i.e. relative

17 37 x 37 17 Spatial sampling Forecast neighbourhood x Make use of spatial verification methods which compare single observations to a forecast neighbourhood around the observation location. SO-NF Observation Represents a fundamental departure from our current verification system strategy where the emphasis is on extracting the nearest GP or bilinear interpolation to get matched forecast-ob pair. NOT upscaling/ smoothing! Only ~130 1.5 km grid points in >500 000 domain used to assess entire forecast! Note the variability in the neighbourhoods.

Time series skill against persistence MOGREPS-UK @ 2.2 km UKV @ 1.5 km MOGREPS-UK is already outperforming UKV

Synoptic evolution: Feature-based assessment

MODE Method for Object-based Diagnostic Evaluation Davis et al., MWR, 2006 Two parameters: 1. Convolution radius 2. Threshold Highly configurable Attributes: Centroid difference, Angle difference, Area ratio etc Focus is on spatial properties, especially the spatial biases

MODE methodology Identification Measure Attributes Merging Matching Comparison Summarize Convolution threshold process Fuzzy Logic Approach Compare forecast and observed attributes Merge single objects into clusters Compute interest values Identify matched pairs Accumulate and examine comparisons across many cases

MODE example: cloud Tracking cloud free areas Using UKPP cloud analysis UK4 UK4 Analysis UKV Analysis UKV Mittermaier and Bullock, 2013)

MODE example: 250 hpa jets t+24h t+96h

Evaluation of dust

Entity-based approach (CRA) Ebert and McBride (2000) Define entities using threshold (Contiguous Rain Areas) Horizontally translate the forecast until a pattern matching criterion is met: minimum total squared error between forecast and observations maximum correlation maximum overlap The displacement is the vector difference between the original and final locations of the forecast. Observed Forecast

CRA details MSEtotal = MSEdisp + MSEvol + MSEpattern The difference between the MSE before and after the feature has been displaced gives MSEdisp= MSEtotal MSEshifted The volume error represents the bias in mean intensity MSEvol = (F X)^2 where F and X are the mean forecast and observed values after displacement The pattern error accounts for differences in the fine structure with MSEpattern = MSEshifted MSEvol

Comparison details 12 km Unified model configuration called the South Asia Model or SAM. Compared to AERONET aerosol optical depth (AOD) and AOD product from Spinning Enhanced Visible and Infrared Imager (SEVIRI) instrument on MSG. SEVIRI viewing area only covers ~half of model domain. AOD can not be calculated for areas contaminated with cloud, or over the sea. Both SEVIRI and forecast put on 0.15deg grid.

Case 1: 26 January 2010

Case 2: 18 February 2010

Case 3: 2 March 2010

CRA time series

Contingency table for matching CRA objects By defining criteria for the centre of mass of isolated objects, a contingency table can be created to see how the model performs as a function of threshold

AERONET site-specific verification Little difference in forecasts with lead time Noisy

AERONET ETS @ Mussafa Not much difference in lead time

Conclusions

Conclusions Point and grid verification, will not necessarily give the same answer. Neither will traditional methods and new spatial methods. New methods can be really useful but not all new spatial methods are equally useful for specific applications. A tangible impact of the double penalty effect will depend on the parameter and the resolution of your forecast/observation. Understanding the limitations of your observations is critical. The most challenging aspect of a verification process is making the decision on what to do and how to do it. Calculating the actual scores is often trivial. Data preparation and choice of methodology are the most important aspects to get right.

Questions?