ATCF: LESSONS LEARNED ON TC CONSENSUS FORECASTING. John Knaff, NOAA/NESDIS and Buck Sampson, NRL, Monterey

Similar documents
11A.1 PREDICTION OF TROPICAL CYCLONE TRACK FORECAST ERROR FOR HURRICANES KATRINA, RITA, AND WILMA

Experiments with a Simple Tropical Cyclone Intensity Consensus

JTWC use of ensemble products. Matthew Kucas Joint Typhoon Warning Center Pearl Harbor, HI

A Deterministic Rapid Intensification Aid

2017 Year in review: JTWC TC Activity, Forecast Challenges, and Developmental Priorities

The Evolution and Use of Objective Forecast Guidance at NHC

A Multi-Model Ensemble for Western North Pacific Tropical Cyclone Intensity Prediction

Evaluation of long-term trends in tropical cyclone intensity forecasts

Tropical Cyclone Track Forecasts Using an Ensemble of Dynamical Models

NAVAL POSTGRADUATE SCHOOL THESIS

OPERATIONAL CONSIDERATIONS FOR HURRICANE MODEL DIAGNOSTICS / VERIFICATION

Satellite Applications to Hurricane Intensity Forecasting

Extratropical transition of tropical cyclones: Operational forecasting challenges. Matthew Kucas Joint Typhoon Warning Center Pearl Harbor, HI, USA

Ensemble Prediction Systems

Model assessment using a multi-metric ranking technique

Brian Strahl, JTWC Buck Sampson, NRL-MRY AG2 Jack Tracey, JTWC. Forward, Ready, Responsive Decision Superiority. Joint Typhoon Warning Center

Methodology for 2013 Stream 1.5 Candidate Evaluation

TROPICAL CYCLONE PROBABILITY PRODUCTS LECTURE 1C: WIND PROBABILITY

2006 Annual Tropical Cyclone Report U.S. Naval Maritime Forecast Center/ Joint Typhoon Warning Center Pearl Harbor, Hawaii

GFDL Hurricane Model Ensemble Performance During the 2012 Hurricane Season

IMPROVEMENT IN HURRICANE

Application of Kalman Filter for Adjusting Tropical Cyclone Track Forecasts

Methodology for 2014 Stream 1.5 Candidate Evaluation

EMC multi-model ensemble TC track forecast

Impact of Stochastic Convection on Ensemble Forecasts of Tropical Cyclone Development

Tropical Cyclone Track Prediction

TC STRUCTURE GUIDANCE UPDATES

The UW-CIMSS Advanced Dvorak Technique (ADT) : An Automated IR Method to Estimate Tropical Cyclone Intensity

Predicting Tropical Cyclone Formation and Structure Change

Upgrade of JMA s Typhoon Ensemble Prediction System

NHC Ensemble/Probabilistic Guidance Products

The impact of assimilation of microwave radiance in HWRF on the forecast over the western Pacific Ocean

Performance of the Navy s Tropical Cyclone Prediction Model in the Western North Pacific Basin during 1996

Using NOGAPS Singular Vectors to Diagnose Large-scale Influences on Tropical Cyclogenesis

Use of the GFDL Vortex Tracker

NAVAL POSTGRADUATE SCHOOL THESIS

A New Method for Estimating Tropical Cyclone Wind Speed Probabilities

NOTES AND CORRESPONDENCE. Statistical Postprocessing of NOGAPS Tropical Cyclone Track Forecasts

Tropical Cyclone Formation/Structure/Motion Studies

11A.3 The Impact on Tropical Cyclone Predictions of a Major Upgrade to the Met Office Global Model

Predicting Tropical Cyclone Formation and Structure Change

Wave Analysis Guidance for Tropical Cyclone Forecast Advisories

Dynamical Statistical Seasonal Prediction of Atlantic Hurricane Activity at NCEP

On the Ability of Global Ensemble Prediction Systems to Predict Tropical Cyclone Track Probabilities

Operational and Statistical Prediction of Rapid Intensity Change. Mark DeMaria and Eric Blake, NCEP/NHC John Kaplan, AOML/HRD

Verification of Tropical Storm Track Prediction in Southeast Asia Using GFS Model

The Dynamics of Tropical Cyclones

Tropical Cyclone Track Prediction

P4.1 CONSENSUS ESTIMATES OF TROPICAL CYCLONE INTENSITY USING MULTISPECTRAL (IR AND MW) SATELLITE OBSERVATIONS

SIXTH INTERNATIONAL WORKSHOP on TROPICAL CYCLONES. Topic 1.5 : Operational guidance and skill in forecasting structure change

Comparison of Typhoon Track Forecast using Dynamical Initialization Schemeinstalled

Motivation & Goal. We investigate a way to generate PDFs from a single deterministic run

Consistent Tropical Cyclone Wind and Wave Forecasts for the U.S. Navy

NWS and Navy Plans for the ATCF and AWIPS2

28th Conference on Hurricanes and Tropical Meteorology, 28 April 2 May 2008, Orlando, Florida.

Derrick Herndon and Chris Velden University of Wisconsin - Madison Cooperative Institute for Meteorological Satellite Studies

Hurricane Track Prediction Using a Statistical Ensemble of Numerical Models

NOAA HFIP R&D Activities Summary: Recent Results and Operational Implementation

T-PARC and TCS08 (Submitted by Pat Harr, Russell Elsberry and Tetsuo Nakazawa)

The Impact of Horizontal Resolution and Ensemble Size on Probabilistic Forecasts of Precipitation by the ECMWF EPS

Recent advances in Tropical Cyclone prediction using ensembles

Verification of TC Tracks and Intensity Forecasts

Enhancing Weather Information with Probability Forecasts. An Information Statement of the American Meteorological Society

Early Guidance for Tropical Weather Systems

Forecast Challenges of the 2017 Hurricane Season and NHC Priorities for 2018

Uncertainty in Operational Atmospheric Analyses. Rolf Langland Naval Research Laboratory Monterey, CA

Comparison of Three Tropical Cyclone Intensity Datasets

Improvements in Hurricane Initialization and Forecasting at NCEP with Global and Regional (GFDL) models

Numerical Weather Prediction: Data assimilation. Steven Cavallo

Probabilistic Evaluation of Prediction and Dynamics of Super Typhoon MEGI (2010)

NWS Operational Marine and Ocean Forecasting. Overview. Ming Ji. Ocean Prediction Center National Weather Service/NCEP. CIOSS/CoRP

NAVAL POSTGRADUATE SCHOOL THESIS

SIXTH INTERNATIONAL WORKSHOP on TROPICAL CYCLONES. Special Focus Topic 3a : Sharing experiences in operational consensus forecasting.

11A.3 THE IMPACT OF DROPSONDE DATA FROM DOTSTAR ON TROPICAL CYCLONE TRACK FORECASTING

A High-Quality Tropical Cyclone Reanalysis Dataset Using 4DVAR Data Assimilation Technique

Evaluation of a Multigrid Barotropic Tropical Cyclone Track Model

The Impacts on Extended-Range Predictability of Midlatitude Weather Patterns due to Recurving Tropical Cyclones

Tropical Cyclone Intensity Forecast Error Predictions and Their Applications

National Oceanic and Atmospheric Administration Hurricane Forecast Improvement Program Five-Year Strategic Plan

EARLY WARNINGS FOR HURRICANES

Overview of HFIP FY10 activities and results

4.3.2 Configuration. 4.3 Ensemble Prediction System Introduction

August Forecast Update for Atlantic Hurricane Activity in 2012

Science Objectives contained in three categories

Coupled Global-Regional Data Assimilation Using Joint States

TSR TROPICAL STORM TRACKER LAUNCH

The Properties of Convective Clouds Over the Western Pacific and Their Relationship to the Environment of Tropical Cyclones

Focus on parameter variation results

Leveraging ISI Multi-Model Prediction for Navy Operations: Proposal to the Office of Naval Research

1.2 DEVELOPMENT OF THE NWS PROBABILISTIC EXTRA-TROPICAL STORM SURGE MODEL AND POST PROCESSING METHODOLOGY

AN UPDATE ON UW-CIMSS SATELLITE-DERIVED WIND DEVELOPMENTS

Ensemble Prediction Systems

Application of NWP products and meteorological information processing system in Hong Kong

Will it rain? Predictability, risk assessment and the need for ensemble forecasts

2014 real-time COAMPS-TC ensemble prediction

The 2009 Hurricane Season Overview

Revisiting the maximum intensity of recurving tropical cyclones

John Gyakum, Ron McTaggart-Cowan, Eyad Atallah, and Lance Bosart

Expansion of NCEP Operational Hurricane Weather Research and Forecast (HWRF) Model Forecast Guidance to all Global Tropical Cyclones

Peter P. Neilley. And. Kurt A. Hanson. Weather Services International, Inc. 400 Minuteman Road Andover, MA 01810

Transcription:

ATCF: LESSONS LEARNED ON TC CONSENSUS FORECASTING John Knaff, NOAA/NESDIS and Buck Sampson, NRL, Monterey

Definitions TC tropical cyclone Aid individual model forecast Guidance any method that provides a forecast Early guidance guidance available at the forecast time Late guidance guidance that is available after the forecast has been issued Interpolation a method that blends initial conditions and late guidance to form early guidance (add the deltas, applied at all times, fazed out) Track- positions (lat, lon) associated with a TC Intensity the maximum (1-minute) wind speed associated with a TC

History Consensus forecasting in economics in the late 1960 s Applied to Meteorology in the early 1970 s Applied to track forecasting in the 1990 s Applied to intensity forecasting in 2000 s

Common Methods Equally weight aids Weighted aids (typically based on RMS errors of recent performance) Regress on the aids (can provide negative weights) User (human) selected

Track Improvements The Beginning 1994 Buy-in complete 1998/9 JTWC (2014)

Intensity Improvements The Beginning 2006 Buy-in by forecasters 2009/10 DeMaria et al. (2014)

Recent Track Consensus Results Track forecasts are considered mature, statistical methods are no longer used, and consensus forecasting has been successful in reducing errors since the 1990s (i.e., when NWP track models started to be the most skillful) Analysis Caveats No bias correction performed Errors for individual aids are non-homogeneous Consensus forecasts is made up from available aids (variable consensus)

Example: wp182014 (Phanfone)

Aid Definitions NVGI/NGPI= NAVGEM EGRI= UK Met Office JGSI= Japanese Global GFNI= GFDL model AVNI= GFS HWFI= H-WRF ECM2= ECMWF CTCI=COAMPS-TC JENI=Japanese Ensemble AEMI=GFS ensemble mean ACEI=TC-ACCESS (Aussie Global) CONW= 2013 Consensus (average) COXX= Consensus Reruns WBAR= Weber barotropic model (removed 2013) All aids are early meaning they are there to use during the forecast process I in 4 th character indicates 6-h interpolation 2 in 4 th character indicates 12-h interpolation (ECM2)

[9] [1] [4] [2] [7] [8] [6] [3] [5] Performance Ranks Real-time aid performance (inhomogeneous) for WP 2012 season. WBAI is considerably less skillful at this point. We have added two new members (COTI and JENI) in the last year. Is it time to retire WBAI from the track consensus?

3 10 9 7 6 5 4 1 2 8 Performance Ranks Inhomogeneous aid track performance for CONW members and CONW reruns (COXX). COXX leader, HWFI, AVNI and AEMI very close.

[6] (3) 6 [2] (1) [1] 4 [4] (2) [8] (6) (4) 3 (7) (5) 5 [3] [7] 2 10 [5] 1 (9) (8) (10) 7 9 8 Ranks 2012 error 2013 error Impact on the Consensus Performance Ranks Here we remove one aid (listed at bottom of each bar) from COXX to get contribution of each aid. Higher error indicates more contribution. ECM2 largest contributor, CTCI contributes even though it had the highest errors. WP 2013 season (283 cases).

Concept disproven Current Performance Past Performance Aid 2012 2013 Contribution (2013) ECM2 2 4 1 NVGI/NGPI 6 6 3 GFNI 7 9 8 EGRI 4 5 6 JGSI 3 7 9 JENI 5 8 10 AVNI 1 3 2 HWFI N/A 2 4 AEMI N/A 1 5 CTCI 8 10 7 WBAR 9 REMOVED REMOVED Colors indicate models that share components (physics, initial conditions, etc.) Typically one of the group provides big contributions while the others contribute considerably less This is due to their relative independence

Comment on independence Model forecasts can be correlated. Those serial correlations result in a reduction of information or degrees freedom One can test The number of effective degrees of freedom See Sampson et al. (2008) Past co-variation If the number of aids is large one can combine strongly correlated models (e.g., a mean of an ensemble system)

Comment on user selected consensus Selected consensus was a hot idea for track forecasting in the early- 2000 s. Conceptually it seems a user could select and remove a bad forecast However, verification of the methods reveled that user selected consensus under performed the all member equally weighted methods Forecasters confirmed this result as use of selected consensus fell of over time Sampson et al. (2007) Relative Skill Frequency of use

Take away points Past and concurrent aid performance does not always indicate future performance. Past and concurrent performance does not always indicate how aid inclusion will impact the consensus forecast errors New aids become available and older aids become unavailable (argument for a variable approach) Independent modelling efforts seem to contribute more to the improvement of consensus forecasts (argument for many different efforts) Multi-model consensus guidance out performed ensemble system means. User selected consensus did not provide additional improvements, but required extra work Over the course of a season individual aids typically perform worst than the simple consensus, but not always

Simple example: Initial Intensity Consensus Results (Sampson et al. 2008) At the time (2006): Intensity forecasts had only recently become skillful (just a few models), & statistical methods were most skillful Few skillful models (2 or 3) Offers some insight to other practical issues Availability Variable vs. Fixed Analysis Caveats No bias correction performed Errors for individual aids are non-homogeneous

Potential members (in 2006) NGPI Navy NOGAPS model UKMI UK Met Office AVNI NCEP GFS model GFNI GFLD Regional Hurricane Model (NOGAPS) GFDI GFDL Regional Hurricane Model (GFS) AFWI Air Force Regional Hurricane Model DSHP Statistical Hurricane Intensity Prediction Scheme, accounting for decay over land. -All models are early except for DSHP

Step 1: Identify potential candidates 4 Skill at 72 h 3 2 1 Skillful SKILL vs Baseline Not Skillful

Step 2: Test % Improvement over CLIPER Skillful INT2=DSHP+GFDI INT3=DSHP+GFDI+GFNI Observations: 1. Adding the unskillful GFNI improved the consensus forecasts

Step 3: Practical considerations A three member variable consensus is available more often (in our case 91% vs 83%) Variable consensus forecasts can be more challenging for forecasters to interpret Fixed consensus forecasts are available for fewer forecast periods

Lessons learned from our experiences Consensus forecasts made by equally weighted aids work as well, if not better, than other approaches Increasing the number of more independent aids makes for better consensus guidance Aid independence is as important as aid skill when constructing consensus guidance That forming a consensus from skillful members was not sufficient to reduce the consensus mean error. The members must also demonstrate independence from each other. Selective consensus efforts thus far have not been worth the extra effort Trimming the poorest performing models is often a good strategy for building superior consensus guidance

From Economics & Finance.combination methods have gained even more ground in the forecasting literature, largely because of the strength of the empirical evidence suggesting that these methods systematically perform better than alternatives based on forecasts from a single model. Stable, equal weights have so far been the workhorse of the combination literature and have set a benchmark that has proved surprisingly difficult to beat. This is surprising since on theoretical grounds one would not expect any particular combination scheme to be dominant, since the various methods incorporate restrictions on the covariance matrix that are designed to trade off bias against reduced parameter estimation error. The optimal bias can be expected to vary across applications, and the scheme that provides the best trade-off is expected to depend on the sample size, the number of forecasting models involved, the ratio of the variance of individual models forecast errors as well as their correlations and the degree of instability in the underlying data generating process. - Timmermann (2005) "At least since the publication of The Combination of Forecasts (Bates and Granger [1969]), economists have known that combining forecasts from different sources can both improve accuracy and reduce forecaster error. In the intervening years, numerous studies have confirmed these conclusions, outlined conditions under which forecast combinations are most effective, and tried to explain why simple equal weights work so well relative to more sophisticated statistical techniques. Jones (2014)

References: Bates, J. M., and C. W. J. Granger (1969). The Combination of Forecasts," Operational Research Quarterly 20, 451-68. DeMaria, M., C.R. Sampson, J.A. Knaff and K.D. Musgrave, (2014). Is tropical cyclone intensity guidance improving? Bull. Amer. Meteor. Soc., 95, 387-398. Jones, Robert C. (2014). "Making Better Investment Decisions". The Journal of Portfolio Management 40 (2): 128 143. Joint Typhoon Warning Center (2014). 2013 Annual Tropical Cyclone Report: Western Pacific (Report). United States Navy, United States Air Force. Sampson, C.R., J.L. Franklin, J.A. Knaff, and M. DeMaria (2008). Experiments with a Simple Tropical Cyclone Intensity Consensus. Weather and Forecasting, 23, 304 312.