Some ideas for Ensemble Kalman Filter

Similar documents
Ensemble Kalman Filter potential

Recent Advances in EnKF

Relationship between Singular Vectors, Bred Vectors, 4D-Var and EnKF

Relationship between Singular Vectors, Bred Vectors, 4D-Var and EnKF

4D-Var or Ensemble Kalman Filter?

Accelerating the spin-up of Ensemble Kalman Filtering

Improved analyses and forecasts with AIRS retrievals using the Local Ensemble Transform Kalman Filter

Comparing Local Ensemble Transform Kalman Filter with 4D-Var in a Quasi-geostrophic model

4D-Var or Ensemble Kalman Filter? TELLUS A, in press

Data assimilation; comparison of 4D-Var and LETKF smoothers

Handling nonlinearity in Ensemble Kalman Filter: Experiments with the three-variable Lorenz model

Advances and Challenges in Ensemblebased Data Assimilation in Meteorology. Takemasa Miyoshi

Data assimilation for the coupled ocean-atmosphere

Estimating observation impact without adjoint model in an ensemble Kalman filter

Generating climatological forecast error covariance for Variational DAs with ensemble perturbations: comparison with the NMC method

The Local Ensemble Transform Kalman Filter (LETKF) Eric Kostelich. Main topics

Parameter Estimation in EnKF: Surface Fluxes of Carbon, Heat, Moisture and Momentum

WRF-LETKF The Present and Beyond

Ji-Sun Kang. Pr. Eugenia Kalnay (Chair/Advisor) Pr. Ning Zeng (Co-Chair) Pr. Brian Hunt (Dean s representative) Pr. Kayo Ide Pr.

Weight interpolation for efficient data assimilation with the Local Ensemble Transform Kalman Filter

Ensemble 4DVAR for the NCEP hybrid GSI EnKF data assimilation system and observation impact study with the hybrid system

STRONGLY COUPLED ENKF DATA ASSIMILATION

EnKF Localization Techniques and Balance

How 4DVAR can benefit from or contribute to EnKF (a 4DVAR perspective)

Weight interpolation for efficient data assimilation with the Local Ensemble Transform Kalman Filter

Simple Doppler Wind Lidar adaptive observation experiments with 3D-Var. and an ensemble Kalman filter in a global primitive equations model

Data Assimilation: Finding the Initial Conditions in Large Dynamical Systems. Eric Kostelich Data Mining Seminar, Feb. 6, 2006

Analysis sensitivity calculation in an Ensemble Kalman Filter

Asynchronous data assimilation

Some Applications of WRF/DART

Multivariate Correlations: Applying a Dynamic Constraint and Variable Localization in an Ensemble Context

Ensemble prediction and strategies for initialization: Tangent Linear and Adjoint Models, Singular Vectors, Lyapunov vectors

Can hybrid-4denvar match hybrid-4dvar?

P 1.86 A COMPARISON OF THE HYBRID ENSEMBLE TRANSFORM KALMAN FILTER (ETKF)- 3DVAR AND THE PURE ENSEMBLE SQUARE ROOT FILTER (EnSRF) ANALYSIS SCHEMES

Local Ensemble Transform Kalman Filter

Quarterly Journal of the Royal Meteorological Society !"#$%&'&(&)"&'*'+'*%(,#$,-$#*'."(/'*0'"(#"(1"&)23$)(4#$2#"( 5'$*)6!

Comparison of 3D-Var and LETKF in an Atmospheric GCM: SPEEDY

A Comparative Study of 4D-VAR and a 4D Ensemble Kalman Filter: Perfect Model Simulations with Lorenz-96

EFSO and DFS diagnostics for JMA s global Data Assimilation System: their caveats and potential pitfalls

Guo-Yuan Lien*, Eugenia Kalnay, and Takemasa Miyoshi University of Maryland, College Park, Maryland 2. METHODOLOGY

A simpler formulation of forecast sensitivity to observations: application to ensemble Kalman filters

Simultaneous estimation of covariance inflation and observation errors within an ensemble Kalman filter

Four-Dimensional Ensemble Kalman Filtering

Ensemble Assimilation of Global Large-Scale Precipitation

Kalman Filter and Ensemble Kalman Filter

Estimation of Surface Fluxes of Carbon, Heat, Moisture and Momentum from Atmospheric Data Assimilation

Ensemble-based Data Assimilation of TRMM/GPM Precipitation Measurements

Accounting for Model Errors in Ensemble Data Assimilation

Local Ensemble Transform Kalman Filter: An Efficient Scheme for Assimilating Atmospheric Data

EnKF Review. P.L. Houtekamer 7th EnKF workshop Introduction to the EnKF. Challenges. The ultimate global EnKF algorithm

Variable localization in an Ensemble Kalman Filter: application to the carbon cycle data assimilation

Monthly Weather Review The Hybrid Local Ensemble Transform Kalman Filter

Adaptive ensemble Kalman filtering of nonlinear systems

We honor Ed Lorenz ( ) who started the whole new science of predictability

Addressing the nonlinear problem of low order clustering in deterministic filters by using mean-preserving non-symmetric solutions of the ETKF

4DEnVar. Four-Dimensional Ensemble-Variational Data Assimilation. Colloque National sur l'assimilation de données

Comparison between Local Ensemble Transform Kalman Filter and PSAS in the NASA finite volume GCM: perfect model experiments

Abstract 2. ENSEMBLE KALMAN FILTERS 1. INTRODUCTION

Comparisons between 4DEnVar and 4DVar on the Met Office global model

Development of the Local Ensemble Transform Kalman Filter

UMD-UCB carbon cycle data assimilation system

Ensemble Kalman Filters for WRF-ARW. Chris Snyder MMM and IMAGe National Center for Atmospheric Research

Prof. Stephen G. Penny University of Maryland NOAA/NCEP, RIKEN AICS, ECMWF US CLIVAR Summit, 9 August 2017

Bred vectors: theory and applications in operational forecasting. Eugenia Kalnay Lecture 3 Alghero, May 2008

An implementation of the Local Ensemble Kalman Filter in a quasi geostrophic model and comparison with 3D-Var

Error (E) Covariance Diagnostics. in Ensemble Data Assimilation. Kayo Ide, Daryl Kleist, Adam Kubaryk University of Maryland

Introduction to data assimilation and least squares methods

Hybrid variational-ensemble data assimilation. Daryl T. Kleist. Kayo Ide, Dave Parrish, John Derber, Jeff Whitaker

Bred Vectors, Singular Vectors, and Lyapunov Vectors in Simple and Complex Models

Data Assimilation Development for the FV3GFSv2

(Extended) Kalman Filter

16. Data Assimilation Research Team

Inter-comparison of 4D-Var and EnKF systems for operational deterministic NWP

Coupled Ocean-Atmosphere Assimilation

HYBRID GODAS STEVE PENNY, DAVE BEHRINGER, JIM CARTON, EUGENIA KALNAY, YAN XUE

Observability, a Problem in Data Assimilation

Assimilating Nonlocal Observations using a Local Ensemble Kalman Filter

(Toward) Scale-dependent weighting and localization for the NCEP GFS hybrid 4DEnVar Scheme

Ting Lei, Xuguang Wang University of Oklahoma, Norman, OK, USA. Wang and Lei, MWR, Daryl Kleist (NCEP): dual resolution 4DEnsVar

In the derivation of Optimal Interpolation, we found the optimal weight matrix W that minimizes the total analysis error variance.

Numerical Weather Prediction: Data assimilation. Steven Cavallo

6.5 Operational ensemble forecasting methods

M.Sc. in Meteorology. Numerical Weather Prediction

Fundamentals of Data Assimilation

LETKF Data Assimilation System for KIAPS AGCM: Progress and Plan

Bred Vectors: A simple tool to understand complex dynamics

Hybrid Variational Ensemble Data Assimilation for Tropical Cyclone

The Local Ensemble Transform Kalman Filter and its implementation on the NCEP global model at the University of Maryland

The Local Ensemble Transform Kalman Filter and its implementation on the NCEP global model at the University of Maryland

How does 4D-Var handle Nonlinearity and non-gaussianity?

Convective-scale data assimilation in the Weather Research and Forecasting model using a nonlinear ensemble filter

Comparing Variational, Ensemble-based and Hybrid Data Assimilations at Regional Scales

J9.4 ERRORS OF THE DAY, BRED VECTORS AND SINGULAR VECTORS: IMPLICATIONS FOR ENSEMBLE FORECASTING AND DATA ASSIMILATION

Introduction to Data Assimilation. Reima Eresmaa Finnish Meteorological Institute

ABSTRACT. from a few minutes (e.g. cumulus convection) to years (e.g. El Niño). Fast time

4. DATA ASSIMILATION FUNDAMENTALS

Numerical Weather prediction at the European Centre for Medium-Range Weather Forecasts (2)

Introduction to ensemble forecasting. Eric J. Kostelich

P7.1 STUDY ON THE OPTIMAL SCANNING STRATEGIES OF PHASE-ARRAY RADAR THROUGH ENSEMBLE KALMAN FILTER ASSIMILATION OF SIMULATED DATA

Optimal Localization for Ensemble Kalman Filter Systems

Transcription:

Some ideas for Ensemble Kalman Filter Former students and Eugenia Kalnay UMCP Acknowledgements: UMD Chaos-Weather Group: Brian Hunt, Istvan Szunyogh, Ed Ott and Jim Yorke, Kayo Ide, and students Former students: Shu-Chih Yang, Takemasa Miyoshi, Junjie Liu, Hong Li, Chris Danforth Also: Malaquías Peña, Matteo Corazza, Pablo Grunman, DJ Patil, Ji-Sun Kang, Debra Baker, Steve Greybush, Tamara Singleton, Matt Hoffman, Steve Penny, Elana Fertig.

Some ideas for Ensemble Kalman Filter Basic idea: We are in competition with 4D-Var We take advantage of ideas and properties that were developed for 4D-Var and show that it can be done easier, cheaper and better with our EnKF, the LETKF (Hunt et al., 2007) And we don t need adjoint, priors, etc.

Some ideas to improve LETKF We can adapt ideas developed within 4D-Var:! No-cost smoother (Kalnay et al, Tellus)! Accelerating the spin-up: Running in place (Kalnay and Yang, QJRMS, submitted)! Outer loop and nonlinearities (Yang and Kalnay)! Forecast sensitivity to observations (Liu and Kalnay, QJRMS, 2008)! Coarse analysis resolution interpolating weights (Yang, Kalnay, Hunt, Bowler, QJ submitted)! Low-dimensional model bias correction (Li, Kalnay, Danforth, Miyoshi, MWR, submitted)

Local Ensemble Transform Kalman Filter (Ott et al, 2004, Hunt et al, 2004, 2007) (Start with initial ensemble) Observation operator ensemble observations ensemble forecasts Observations LETKF ensemble analyses Model Model independent (black box) Obs. assimilated simultaneously at each grid point 100% parallel: very fast No adjoint needed 4D LETKF extension

Localization based on observations Perform data assimilation in a local volume, choosing observations The state estimate is updated at the central grid red dot

Localization based on observations Perform data assimilation in a local volume, choosing observations The state estimate is updated at the central grid red dot All observations (purple diamonds) within the local region are assimilated The LETKF algorithm can be described in a single slide!

Local Ensemble Transform Kalman Filter (LETKF( LETKF) Globally: Forecast step: Analysis step: construct Locally: Choose for each grid point the observations to be used, and compute the local analysis error covariance and perturbations in ensemble space: Analysis mean in ensemble space: and add to b x n,k ( ) a = M n x n!1,k X b = "# x b 1! x b... x b K! x b $ % ; y b i = H (x b i ); Y b n = "# y b 1! y b... y b K! y b $ %!P a = "# ( K! 1)I + Y bt R!1 Y b!1 $ % ; W a = [(K! 1) P! a ] 1/2 W a w a =! P a Y bt R!1 (y o! y b ) to get the analysis ensemble in ensemble space The new ensemble analyses in model space are the columns of X a n = X b n W a + x b. Gathering the grid point analyses forms the new global analyses. Note that the the output of the LETKF are analysis weights w a and perturbation analysis matrices of weights W a. These weights multiply the ensemble forecasts.

Analysis time The 4D-LETKF produces an analysis in terms of weights of the ensemble forecast members at the analysis time t n, giving the trajectory that best fits all the observations in the assimilation window.

No-cost LETKF smoother ( ): apply at t n-1 the same weights found optimal at t n. It works for 3D- or 4D-LETKF The 4D-LETKF produces an analysis in terms of weights of the ensemble forecast members at the analysis time t n, giving the trajectory that best fits all the observations in the assimilation window.

No-cost LETKF smoother test on a QG model: It really works! LETKF analysis at time n x n a = x n f + X n f w n a LETKF Analysis Smoother analysis at time n-1!x a n!1 f = x n!1 f a + X n!1 w n Smoother reanalysis This very simple smoother allows us to go back and forth in time within an assimilation widow!!

Running in place to spin-up faster Kalnay and Yang (2008) 4D-Var spins-up faster than EnKF because it is a smoother: it keeps iterating until it fits the observations within the assimilation window as well as possible EnKF spins-up fast if starting from a good initial state, e.g., 3D-Var, but needs also an ensemble representing the errors of the day In a severe storm where radar observations start with the storm, there is little real time to spin-up Caya et al. (2005): EnKF is eventually better than 4D-Var (but it is too late to be useful, it misses the storm). Jidong Gao, (pers. comm. 2007): spin-up is the main obstacle for the use of EnKF for storm prediction.

Can we use the data more than once? Hunt et al., 2007: The background term represents the evolution of the maximum likelihood trajectory given all the observations in the past n!1 After the analysis a similar relationship is valid: "x! x n b # T $ % ( b Pn )!1 b " x! x n # T R j!1 " y o j! H j M tn x% ",t y #$ j &' o j! H j M tn x% (,t #$ j &' = T b " # x! x n % & j=1 $ % + " # y o! H n n x$ % ( )" y o n! H n x T R n!1 From here one can derive the linear KF equations Also the rule: Use the data once and then discard it # $ % = x! x a " # n ( b Pn )!1 b " x! x n T $ % # ( a Pn )!1 a " x! x n # % & + c $ % + c'

Can we use the data more than once? The rule: Use the data once and then discard it (Ide et al., 1997) makes sense when the analysis/forecasts are the most likely given all the past data, not when we start from scratch. We propose Running in place until we extract the maximum information form the observations. We need 4D-LETKF (Hunt et al, 2004) to use all the observations within an assimilation window at their right time A No-Cost Smoother (Kalnay et al., 2007b) An appropriate iterative scheme

Running in Place EnKF is a sequential data assimilation system where, after the new data is used at the analysis time, it should be discarded only if the previous analysis and the new background are the most likely states given the past observations. If the system has converged after the initial spin-up all the information from past observations is already included in the background. During the spin-up we should use the observations repeatedly if we can extract extra information. But we should avoid overfitting the observations

Running in Place algorithm (1) Cold-start the EnKF from any initial ensemble mean and random perturbations at t 0, and integrate the initial ensemble to t 1. The running in place loop with n=1, is:

Running in Place algorithm (2) a) Perform a standard EnKF analysis and obtain the analysis weights at t n, saving the mean square observations minus forecast (OMF) computed by the EnKF. b) Apply the no-cost smoother to obtain the smoothed analysis ensemble at t n-1 by using the same weights obtained at t n. c) Perturb the smoothed analysis ensemble with a small amount of random Gaussian perturbations, similar to additive inflation. d) Integrate the perturbed smoothed ensemble to t n. If the forecast fit to the observations is smaller than in the previous iteration according to some criterion, go to a) and perform another iteration. If not, let t n!1 " t n and proceed to the next assimilation window.

Running in Place algorithm (notes) Notes: c) Perturb the smoothed analysis ensemble with a small amount of random Gaussian perturbations, a method similar to additive inflation. This perturbation has two purposes: 1) Avoid reaching the same analysis as before, and 2) Encourage the ensemble to explore new unstable directions d) Convergence criterion: if OMF 2 (iter)! OMF 2 (iter + 1) OMF 2 (iter) > " with!! 5% do another iteration. Otherwise go to the next assimilation window.

Results with a QG model It works well (red) spin-up becomes as fast as 4D-Var (blue). With 5% criterion for reduction of OMF errors (red), it takes only 2-4 iterations. With 20 members (green) it still works.

Discussion of spin-up acceleration Number of iterations during spin-up: 2-4, computationally acceptable We could use the weights interpolation of Yang et al. (2008b) and run in place only where the action is. There are many applications where a fast spin-up is important. It could also help to handle the initial bias in biased systems, a very difficult problem (Miyoshi, pers. comm.)

Nonlinearities and outer loop The main disadvantage of EnKF is that it cannot handle nonlinear (non-gaussian) perturbations and therefore needs short assimilation windows. It doesn t t have the important outer loop so important in 3D- Var and 4D-Var (DaSilva, pers. comm.. 2006) Lorenz -3 variable model (Kalnay et al. 2007a Tellus), RMS analysis error 4D-Var LETKF Window=8 steps 0.31 0.30 (linear window) Window=25 steps 0.53 0.66 (nonlinear window) Long windows + Pires et al. => 4D-Var wins!

Outer loop in 4D-Var

Comparison of ensemble-based and variational-based data assimilation schemes in a Quasi-Geostrophic model. 3D-Var Hybrid (3DVar+20 BVs) 12-hour 4D-Var LETKF (40 ensemble) 24-hour 4D-Var EnKF does not handle well long windows because ensemble perturbations become non-gaussian. 4D-Var simply iterates and produces a more accurate control. We can imitate this with the outer loop idea for LETKF.

Nonlinearities and outer loop Outer loop: do the same as 4D-Var, and use the final weights to correct only the mean initial analysis, keeping the initial perturbations. Repeat the analysis once or twice. It centers the ensemble on a more accurate nonlinear solution. Miyoshi pointed out that Jaszwinski (1970) suggested this in a footnote!!!!! Lorenz -3 variable model RMS analysis error 4D-Var LETKF LETKF +outer loop Window=8 steps 0.31 0.30 0.27 Window=25 steps 0.53 0.66 0.48 Running in place further reduces RMS from 0.48 to 0.39!

Estimation of observation impact without adjoint in an ensemble Kalman filter Junjie Liu and Eugenia Kalnay

Background AIRS shortwave 4.180 µm AIRS shortwave 4.474 µm AIRS longwave 14-13 µm AMSU/A " The adjoint method proposed by Langland and Baker (2004) and Zhu and Gelaro (2007) quantifies the reduction in forecast error for each individual observation source " The adjoint method detects the observations which make the forecast worse. " The adjoint method requires adjoint model which is difficult to get.

Objective and outline Objective " Propose an ensemble sensitivity method to calculate observation impact without using adjoint model. Outline " Illustrate and derive the ensemble sensitivity method; " With Lorenz-40 variable model, compare the ensemble sensitivity method with adjoint method in # the ability to represent the actual error reduction; # the ability to detect the poor quality observations. " Summary

Schematic of the observation impact on the reduction of forecast error f e t!6 = x t!6! x t a e t!6 e t 0 f e t 0 = x t 0! x t a OBS. (Adapted from Langland and Baker, 2004) -6hr 00hr analysis t The only difference between e t 0 and e t!6 is the assimilation of observations at 00hr. " Observation impact on the reduction of forecast error: J = 1 2 (e T T t 0e t 0! e t!6 e t!6 )

The ensemble sensitivity method Euclidian cost function: J = 1 2 (e T T t 0e t 0! e t!6 e t!6 ) v 0 = y o b 0! h(x 0!6 ) Cost function as function of obs. increments: J = v 0,!J!v 0

The ensemble sensitivity method Euclidian cost function: J = 1 2 (e T T t 0e t 0! e t!6 e t!6 ) v 0 = y o b 0! h(x 0!6 ) Cost function as function of obs. Increments: J = v 0,!J!v 0 The sensitivity of cost function with respect to the assimilated observations:!j =!K T ft f 0 X t "6!v #$ % & #$ e t "6 + X t "6!K 0 v 0 % & 0 With this formula we can predict the impact of observations on the forecasts!

Ability to detect the poor quality observation Observation impact from LB (red) and from ensemble sensitivity method (green) Larger random error Biased observation case! Like adjoint method, ensemble sensitivity method can detect the observation poor quality (11 th observation location)! The ensemble sensitivity method has a stronger signal when the observation has negative impact on the forecast.

Summary for forecast sensitivity to obs. Derived a formula to calculate the observation impact based on the ensemble without using the adjoint model which usually is not available. The results based on Lorenz-40 variable model show that ensemble sensitivity method without using adjoint model gives results similar to adjoint method. Like adjoint method, ensemble sensitivity method can detect the observation which either has larger random error or has bias. Under such conditions, the ensemble sensitivity method has stronger and more accurate signal. It provides a powerful tool to check the quality of the observations.

Coarse analysis with interpolated weights Yang et al (2008) In EnKF the analysis is a weighted average of the forecast ensemble We performed experiments with a QG model interpolating weights compared to analysis increments. Coarse grids of 11%, 4% and 2% interpolated analysis points. 1/(3x3)=11% analysis grid

Coarse analysis with interpolated weights Weights vary on very large scales: they interpolate well. Interpolated weights are obtained even for data void areas.

Analysis increments With increment interpolation, the analysis is OK only with 50% analysis coverage With weight interpolation, there is almost no degradation! EnKF maintains balance and conservation properties

Impact of coarse analysis on accuracy With increment interpolation, the analysis degrades With weight interpolation, there is no degradation, the analysis is actually better!

Model error: comparison of methods to correct model bias and inflation Hong Li, Chris Danforth, Takemasa Miyoshi, and Eugenia Kalnay

Model error: If we assume a perfect model in EnKF, we underestimate the analysis errors (Li, 2007) perfect SPEEDY model imperfect model (obs from NCEP- NCAR Reanalysis NNR)

Why is EnKF vulnerable to model errors? Imperfect model Ens 1 Ens mean Ens 2 Truth Forecast error " In the theory of Extended Kalman filter, forecast error is represented by the growth of errors in IC and the model errors. P = M P M + Q f a T i x i! 1 a a i! 1 xi! 1 The ensemble spread is blind to model errors " However, in ensemble Kalman filter, error estimated by the ensemble spread can only represent the first type of errors. P i f! 1 K f #(x i " x f f )(x i k " 1 i=1 " x f ) T

We compared several methods to handle bias and random model errors perfect model imperfect model Low Dimensional Method to correct the bias (Danforth et al, 2007) combined with additive inflation

Discussion: 4D-Var vs.. EnKF war We should be able to adopt some simple strategies to capture the advantages of 4D-Var: Smoothing and running in place A simple outer loop to deal with nonlinearities Adjoint sensitivity without adjoint model Coarse resolution analysis without degradation Correcting model bias (Baek et al, 2006, Danforth et al, 2007, Li et al. submitted). The correction of model bias combined with additive inflation gives the best results It seems like there is nothing that 4D-Var can do that EnKF cannot do as well, usually simpler, cheaper and better. There is a WMO Workshop in Buenos Aires this November on 4D-Var and EnKF intercomparisons and a 2-week intensive course on data assimilation for Latin American scientists (140 applications!)

Bias removal schemes (Low Dimensional Method) 2.3 Low-dim method (Danforth et al, 2007: Estimating and correcting global weather model error. Mon. Wea. Rev, J. Atmos. Sci., 2007) Generate a long time series of model forecast minus reanalysis from the training period f x e x 6hr NNR t=0 model NNR NNR NNR x truth t=6hr We collect a large number of estimated errors and estimate from them bias, etc. NNR x f % L f f t a t n xn xn M xn M xn b! + 1 = + 1 " + 1 = ) " ( ) + + $ n, lel + l= 1 M! ( # m= 1 n, m f m Forecast error due to error in IC Time-mean model bias Diurnal model error State dependent model error

Low-dimensional method Include Bias, Diurnal and State-Dependent model errors: model error = b + "! n,l e l + "# n,m f m 2 l =1 10 m=1 Having a large number of estimated errors allows to estimate the global model error beyond the bias

SPEEDY 6 hr model errors against NNR (diurnal cycle) 1987 Jan 1~ Feb 15 Error anomalies e ' e e x6 hr = x i 6hr! x ( ) 6hr pc1 pc2 Leading EOFs for 925 mb TEMP For temperature at lower-levels, in addition to the time-independent bias, SPEEDY has diurnal cycle errors because it lacks diurnal radiation forcing