Verifying Ensemble Forecasts Using A Neighborhood Approach

Similar documents
Extracting probabilistic severe weather guidance from convection-allowing model forecasts. Ryan Sobash 4 December 2009 Convection/NWP Seminar Series

Severe storm forecast guidance based on explicit identification of convective phenomena in WRF-model forecasts

Using Convection-Allowing Models to Produce Forecast Guidance For Severe Thunderstorm Hazards via a Surrogate-Severe Approach!

NEXT-DAY CONVECTION-ALLOWING WRF MODEL GUIDANCE: A SECOND LOOK AT 2 VS. 4 KM GRID SPACING

Comparison of Convection-permitting and Convection-parameterizing Ensembles

Aaron Johnson* and Xuguang Wang. University of Oklahoma School of Meteorology and Center for Analysis and Prediction of Storms (CAPS), Norman, OK

Spatial forecast verification

Spatial verification of NWP model fields. Beth Ebert BMRC, Australia

Evaluation of Ensemble Icing Probability Forecasts in NCEP s SREF, VSREF and NARRE-TL Systems

SPC Ensemble Applications: Current Status and Future Plans

CAPS Storm-Scale Ensemble Forecasting (SSEF) System

VERIFICATION OF PROXY STORM REPORTS DERIVED FROM ENSEMBLE UPDRAFT HELICITY

Convection-Permitting Ensemble Forecasts at CAPS for Hazardous Weather Testbed (HWT)

THE DECORRELATION SCALE: METHODOLOGY AND APPLICATION FOR PRECIPITATION FORECASTS

THE MODEL EVALUATION TOOLS (MET): COMMUNITY TOOLS FOR FORECAST EVALUATION

OPTIMIZING PROBABILISTIC HIGH-RESOLUTION ENSEMBLE GUIDANCE FOR HYDROLOGIC PREDICTION

Precipitation verification. Thanks to CMC, CPTEC, DWD, ECMWF, JMA, MF, NCEP, NRL, RHMC, UKMO

Upscaled and fuzzy probabilistic forecasts: verification results

1. INTRODUCTION 3. PROJECT DATAFLOW

Precipitation Calibration for the NCEP Global Ensemble Forecast System

Convection-Resolving NWP with WRF. Section coordinator Ming Xue University of Oklahoma

A Cycled GSI+EnKF and Storm-Scale Ensemble Forecasting (SSEF) Experiment

Verification of ensemble and probability forecasts

Identifying skilful spatial scales using the Fractions Skill Score

Concepts of Forecast Verification FROST-14

Center for Analysis and Prediction of Storms 1 and School of Meteorology 2 University of Oklahoma, Norman, OK 73072

Using time-lag ensemble techniques to assess behaviour of high-resolution precipitation forecasts

Thunderstorm-Scale EnKF Analyses Verified with Dual-Polarization, Dual-Doppler Radar Data

The Impact of Horizontal Resolution and Ensemble Size on Probabilistic Forecasts of Precipitation by the ECMWF EPS

COMPOSITE-BASED VERIFICATION OF PRECIPITATION FORECASTS FROM A MESOSCALE MODEL

Assimilation of cloud/precipitation data at regional scales

Weather Hazard Modeling Research and development: Recent Advances and Plans

32 AN OVERVIEW OF CAPS STORMSCALE ENSEMBLE FORECAST FOR THE 2015 NOAA HWT SPRING FORECASTING EXPERIMENT

CAPS Storm-Scale Ensemble Forecast for the NOAA Hazardous Weather Testbed (HWT) Spring Experiment

Basic Verification Concepts

Basic Verification Concepts

A comparison of QPF from WRF simulations with operational NAM and GFS output using multiple verification methods

Categorical Verification

University of Oklahoma, Norman, OK INTRODUCTION

Introduction to NCEP's time lagged North American Rapid Refresh Ensemble Forecast System (NARRE-TL)

New Development in CAPS Storm-Scale Ensemble Forecasting System for NOAA 2012 HWT Spring Experiment

Assessing high resolution forecasts using fuzzy verification methods

A WRF-based rapid updating cycling forecast system of. BMB and its performance during the summer and Olympic. Games 2008

Toward improved initial conditions for NCAR s real-time convection-allowing ensemble. Ryan Sobash, Glen Romine, Craig Schwartz, and Kate Fossell

Adaptation for global application of calibration and downscaling methods of medium range ensemble weather forecasts

Convection-Allowing Models (CAMs) A discussion on Creating a Roadmap to a Unified CAM-based Data-Assimilation and Forecast Ensemble

3.6 NCEP s Global Icing Ensemble Prediction and Evaluation

Spatial methods for assessing radar data assimilation performance at RHM: a framework

Probabilistic Forecast Verification. Yuejian Zhu EMC/NCEP/NOAA

Deterministic and Ensemble Storm scale Lightning Data Assimilation

Object-based evaluation of the impact of horizontal grid spacing on convection-allowing forecasts

A Gaussian Mixture Model Approach to

Feature-specific verification of ensemble forecasts

457 EVALUATION OF CAPS MULTI-MODEL STORM-SCALE ENSEMBLE FORECAST FOR THE NOAA HWT 2010 SPRING EXPERIMENT

Convective-scale NWP for Singapore

Aaron Johnson School of Meteorology, University of Oklahoma and Center for Analysis and Prediction of Storms, Norman, Oklahoma

Verification Methods for High Resolution Model Forecasts

Application and verification of ECMWF products 2010

Application and verification of ECMWF products 2015

Global Flash Flood Forecasting from the ECMWF Ensemble

Towards Operational Probabilistic Precipitation Forecast

ICAM conference 6 June 2013 Kranjska Gora (SLO) Objective forecast verification of WRF compared to ALARO and the derived INCA-FVG outputs

Spatial Forecast Verification Methods

DEFINITION OF DRY THUNDERSTORMS FOR USE IN VERIFYING SPC FIRE WEATHER PRODUCTS. 2.1 Data

Application and verification of ECMWF products 2008

MULTI-RESOLUTION ENSEMBLE FORECASTS OF AN OBSERVED TORNADIC THUNDERSTORM SYSTEM, PART I: COMPARISON OF COARSE AND FINE GRID EXPERIMENTS

J11.1 AN OVERVIEW OF THE 2013 NOAA HAZARDOUS WEATHER TESTBED SPRING FORECASTING EXPERIMENT

Chengsi Liu 1, Ming Xue 1, 2, Youngsun Jung 1, Lianglv Chen 3, Rong Kong 1 and Jingyao Luo 3 ISDA 2019

Variable-Resoluiton Global Atmospheric Modeling Spanning Convective to Planetary Scales

A new mesoscale NWP system for Australia

Surface Hydrology Research Group Università degli Studi di Cagliari

Model verification and tools. C. Zingerle ZAMG

Towards a Definitive High- Resolution Climate Dataset for Ireland Promoting Climate Research in Ireland

Comparison of the NCEP and DTC Verification Software Packages

Implementation of global surface index at the Met Office. Submitted by Marion Mittermaier. Summary and purpose of document

Performance of TANC (Taiwan Auto- Nowcaster) for 2014 Warm-Season Afternoon Thunderstorm

An Investigation of Reforecasting Applications for NGGPS Aviation Weather Prediction: An Initial Study of Ceiling and Visibility Prediction

Drought forecasting methods Blaz Kurnik DESERT Action JRC

HPC Ensemble Uses and Needs

Christopher J. Melick*,1,2, Israel L. Jirak 1, James Correia Jr. 1,2, Andrew R. Dean 1, and Steven J. Weiss 1

MULTIRESOLUTION ENSEMBLE FORECASTS OF AN OBSERVED TORNADIC THUNDERSTORM SYSTEM. PART II: STORM-SCALE EXPERIMENTS

J th Conference on Aviation, Range, and Aerospace Meteorology, 90th AMS Annual Meeting, Atlanta, GA, Jan, 2010

ENSEMBLE FORECASTS AND VERIFICATION OF THE MAY 2015 MULTI-HAZARD SEVERE WEATHER EVENT IN OKLAHOMA. Austin Coleman 1, Nusrat Yussouf 2

Amy Harless. Jason Levit, David Bright, Clinton Wallace, Bob Maxson. Aviation Weather Center Kansas City, MO

Exploring ensemble forecast calibration issues using reforecast data sets

9.4. Jennifer L. Mahoney NOAA Research-Forecast Systems Laboratory, Boulder, Colorado

2012 and changes to the Rapid Refresh and HRRR weather forecast models

Developments towards multi-model based forecast product generation

A 21st century HRRR-based approach to estimating probable maximum precipitation to enhance dam safety and community resilience

Numerical Weather Prediction: Data assimilation. Steven Cavallo

Standardized Verification System for Long-Range Forecasts. Simon Mason

Combining Deterministic and Probabilistic Methods to Produce Gridded Climatologies

46 AN OVERVIEW OF THE 2014 NOAA HAZARDOUS WEATHER TESTBED SPRING FORECASTING EXPERIMENT

EWGLAM/SRNWP National presentation from DMI

Motivation & Goal. We investigate a way to generate PDFs from a single deterministic run

Focus on Spatial Verification Filtering techniques. Flora Gofa

10.3 PROBABILISTIC FORECASTS OF LOCATION AND TIMING OF CONVECTION DURING THE 2012 NOAA HAZARDOUS WEATHER TESTBED SPRING FORECASTING EXPERIMENT

10A.1 The Model Evaluation Tool

J11.5 HYDROLOGIC APPLICATIONS OF SHORT AND MEDIUM RANGE ENSEMBLE FORECASTS IN THE NWS ADVANCED HYDROLOGIC PREDICTION SERVICES (AHPS)

Peter P. Neilley. And. Kurt A. Hanson. Weather Services International, Inc. 400 Minuteman Road Andover, MA 01810

Transcription:

Verifying Ensemble Forecasts Using A Neighborhood Approach Craig Schwartz NCAR/MMM schwartz@ucar.edu Thanks to: Jack Kain, Ming Xue, Steve Weiss

Theory, Motivation, and Review

Traditional Deterministic Point-by-point Verification Put the forecast and observations on the same grid Define an event (e.g. precip. exceedance) Perform a point-by-point comparison of the observations and the forecasts for each of the N grid points Sort the points by the degree of event overlap Typically organized into a 2 x 2 contingency table

Traditional Ensemble Guidance Define an event (e.g. precip. exceedance) At each grid point, count how many members forecast the event then divide by the total number of ensemble members Creates a field of probabilities from the direct model output Often use reliability diagrams, ROC curves to assess aspects of ensemble performance

Forecast Quality and Value It would be ideal if objective statistics agreed with subjective impressions But often, objective verification metrics do not corroborate the perceived value of a forecast This issue of quality versus value is magnified when examining high-resolution model output Finer grids are more heavily penalized for displacement errors

It s a beautiful day in the neighborhood High-resolution models are not accurate at the grid scale To account for spatial displacement errors, specify a Radius of Influence (r) about each grid point Define an event (e.g. precip. exceedance) Count number of events within r, then divide by total number of points within the neighborhood to determine the probability the event will occur at each grid point Creates a probabilistic forecast from a deterministic forecast a Neighborhood Probability (NP)

Schematic Example r = 2.5 times the grid spacing The event has occurred in the shaded boxes Circular geometry: 21 total boxes in neighborhood Event occurs in 8 boxes P = 8/21 = 38% Hypothetical model output + + + + + + + + r + + + + + + r + + + + + + + + + + +

Example Applied to Model and Obs Model and obs on same grid P = 8/21 = 38% P = 8/21 = 38% A perfect forecast using this neighborhood approach

Example Event: Apply Neighborhood Precip. to Obs 5.0 as mm/hr well WRF4 1-hr precip. WRF4 Obs 5 mm/hr Direct model output < 5 mm/hr Binary Field Fractions, r = 25 km Fractions, r = 75 km

Compare fractions with fractions Fractions skill score (FSS) where No overlap of non-zero fractions Model fractional field Observational fractional field

Neighborhood approach applied to an ensemble Apply the neighborhood approach to each of the 10 ensemble members separately Average all probabilistic fields cn n1 n2 p1 p2 ph1 ph2 ph3 ph4 ph5 AVERAGE Neighborhood Ensemble Probability (NEP) 10 probabilistic fields generated via neighborhood approach Alternatively, average the traditional (point-based) ensemble probability over the neighborhood

What do these fields look like? NEP; r = 25 km NEP; r = 75 km 1.0 mm/hr threshold NEP; r = 125 km Traditional prob.

Examples

The NOAA HWT Spring Experiment (formerly the NSSL/SPC Spring Experiment) Annual experiments since 2000 Participants and contributions from external organizations Fosters communication between researchers and forecasters while examining new technology Usually a focus on severe weather Since 2004, focus on evaluating high-resolution NWP models

Ran by CAPS WRF-ARW core Ensemble Configurations 10 member 4 km ensemble (2007,2008) 18 member 4 km ensemble (2009) 30-33 hour forecasts Convection resolved explicitly SE2007: 2100 UTC initialization SE2008, 2009: 0000 UTC initialization Radar assimilation in 2008, 2009 Computational Domain Computational (SE2008) Domain (SE2007)

Model Climatology Examined each model forecast hour, averaged over all days in the 2007 (2008) Spring Experiment Emphasis on hourly precipitation Chose accumulation thresholds to define events Used NCEP Stage II data as observational truth All objective statistics computed over this domain

FSS as a function of r (SE2007) Data aggregated from f21-f33 (1800-0600Z)

FSS as a function of r (SE2008) Data aggregated from f18-f30 (1800-0600Z)

ROC area > 0.70 indicates useful forecasting system ROC Area

http://hwt.nssl.noaa.gov/spring_2009/probgrids_precip.php?date=20090515&hour=24 NEP, r = 50 km for threshold of 2.0 5.0 mm/hr Stage II Precip.

NEP Conclusions The neighborhood ensemble probabilities yielded better forecasts in terms of quality than the traditional ensemble probability, on average over the Experiment NEP focuses on scales where the model can be expected to be accurate Can apply this approach to other discontinuous fields (updraft helicity, simulated reflectivity)

Questions Do you think applying a neighborhood approach is appropriate for coarse grids? What is the coarsest grid that necessitates the use of a neighborhood method? Is the neighborhood ensemble probability approach useful for a large ensemble? How should a neighborhood approach be applied to continuous fields?

Dual uses of Fractional Fields 1) Can be used as an end-product Probabilistic guidance tool 2) Can be used as an intermediate step in objective model verification Fractions, r = 25 km

http://hwt.nssl.noaa.gov/spring_2009/probgrids_precip.php?date=20090513&hour=24 NEP, r = 50 km for threshold of 5.0 mm/hr Stage II Precip.

http://hwt.nssl.noaa.gov/spring_2009/probgrids_precip.php?date=20090603&hour=24 NEP, r = 50 km for threshold of 2.0 mm/hr Stage II Precip.

1-hr Precipitation: 2100 UTC 29 May

SE2007 WRF Model Configurations Ran daily by CAPS WRF-ARW core 10 member 4 km ensemble Single 2 km deterministic forecast 33 hour forecasts Convection resolved explicitly 2100 UTC initialization Cold start no regional data assimilation Computational Domain (SE2007) WORF from Star Trek

ROC Curves Data aggregated from f21-f33 (1800-0600Z)

FSS as a function of time (SE2007)

Variations in IC/LBC and model physics (SE2007)

Apply Neighborhood to Obs as well WRF4 Obs r = 25 km r = 75 km

Acknowledgements

Contingency Table for Dichotomous Events Observed Yes No Yes a b a+b Forecast No c d c+d a+c b+d N Standard 2 x 2 contingency table Bias = (a+b)/(a+c) Threat Score = a/(a+b+c) Equitable Threat Score = (a-e)/(a+b+c-e) where e = (a+b)*(a+c) But this is a point-by-point verification technique

Overall Conclusions High-resolution models have the potential to significantly improve weather forecasting Immediately increasing resolution beyond ~ 4 km when additional computer power becomes available might not be the best use of the new resources Post-processing techniques can be explored Note of caution: As computer power continues to increase, the conclusion about grid spacing may change Radar assimilation seems to be useful for short-range forecasts

Verification Approach When possible, statistics were computed on native grids For some metrics, it was necessary that the models and observations be on the same grid In these instances, the model output was interpolated to the Stage II grid (~ 4.7 km grid length) Statistical significance tested with a bootstrapping method Focus on afternoon (1800-0600 UTC; f21-f33) period to evaluate the utility of the models as next-day guidance