Local Ensemble Transform Kalman Filter

Similar documents
Local Ensemble Transform Kalman Filter: An Efficient Scheme for Assimilating Atmospheric Data

Addressing the nonlinear problem of low order clustering in deterministic filters by using mean-preserving non-symmetric solutions of the ETKF

Localization in the ensemble Kalman Filter

Efficient Data Assimilation for Spatiotemporal Chaos: a Local Ensemble Transform Kalman Filter

The Local Ensemble Transform Kalman Filter (LETKF) Eric Kostelich. Main topics

4D-Var or Ensemble Kalman Filter? TELLUS A, in press

Four-Dimensional Ensemble Kalman Filtering

A Comparative Study of 4D-VAR and a 4D Ensemble Kalman Filter: Perfect Model Simulations with Lorenz-96

Efficient Data Assimilation for Spatiotemporal Chaos: a Local Ensemble Transform Kalman Filter

(Extended) Kalman Filter

Development of the Local Ensemble Transform Kalman Filter

Kalman Filter and Ensemble Kalman Filter

Data Assimilation: Finding the Initial Conditions in Large Dynamical Systems. Eric Kostelich Data Mining Seminar, Feb. 6, 2006

Relative Merits of 4D-Var and Ensemble Kalman Filter

Quarterly Journal of the Royal Meteorological Society

Abstract 2. ENSEMBLE KALMAN FILTERS 1. INTRODUCTION

The Local Ensemble Transform Kalman Filter and its implementation on the NCEP global model at the University of Maryland

Assimilating Nonlocal Observations using a Local Ensemble Kalman Filter

Maximum Likelihood Ensemble Filter Applied to Multisensor Systems

Handling nonlinearity in Ensemble Kalman Filter: Experiments with the three-variable Lorenz model

The Local Ensemble Transform Kalman Filter and its implementation on the NCEP global model at the University of Maryland

Ensemble Kalman Filter potential

Optimal Localization for Ensemble Kalman Filter Systems

Adaptive ensemble Kalman filtering of nonlinear systems

A Note on the Particle Filter with Posterior Gaussian Resampling

Four-dimensional ensemble Kalman filtering

EnKF Review. P.L. Houtekamer 7th EnKF workshop Introduction to the EnKF. Challenges. The ultimate global EnKF algorithm

P 1.86 A COMPARISON OF THE HYBRID ENSEMBLE TRANSFORM KALMAN FILTER (ETKF)- 3DVAR AND THE PURE ENSEMBLE SQUARE ROOT FILTER (EnSRF) ANALYSIS SCHEMES

ABSTRACT. Sabrina Rainwater Doctor of Philosophy, Professor Brian R. Hunt Department of Mathematics and

Gaussian Filtering Strategies for Nonlinear Systems

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit

Comparison between Local Ensemble Transform Kalman Filter and PSAS in the NASA finite volume GCM perfect model experiments

Multivariate Correlations: Applying a Dynamic Constraint and Variable Localization in an Ensemble Context

Asynchronous data assimilation

Practical Aspects of Ensemble-based Kalman Filters

A Unification of Ensemble Square Root Kalman Filters. and Wolfgang Hiller

A variance limiting Kalman filter for data assimilation: I. Sparse observational grids II. Model error

A Local Ensemble Kalman Filter for Atmospheric Data Assimilation

EnKF Localization Techniques and Balance

Comparison between Local Ensemble Transform Kalman Filter and PSAS in the NASA finite volume GCM: perfect model experiments

Accelerating the spin-up of Ensemble Kalman Filtering

How 4DVAR can benefit from or contribute to EnKF (a 4DVAR perspective)

Data assimilation; comparison of 4D-Var and LETKF smoothers

Some ideas for Ensemble Kalman Filter

Review of Covariance Localization in Ensemble Filters

Improved analyses and forecasts with AIRS retrievals using the Local Ensemble Transform Kalman Filter

Relationship between Singular Vectors, Bred Vectors, 4D-Var and EnKF

Relationship between Singular Vectors, Bred Vectors, 4D-Var and EnKF

Fundamentals of Data Assimila1on

A Comparison of Hybrid Ensemble Transform Kalman Filter Optimum Interpolation and Ensemble Square Root Filter Analysis Schemes

A mechanism for catastrophic filter divergence in data assimilation for sparse observation networks

Forecasting and data assimilation

A Local Ensemble Kalman Filter for Atmospheric Data Assimilation

J1.3 GENERATING INITIAL CONDITIONS FOR ENSEMBLE FORECASTS: MONTE-CARLO VS. DYNAMIC METHODS

Data assimilation in high dimensions

DART_LAB Tutorial Section 2: How should observations impact an unobserved state variable? Multivariate assimilation.

Simple Doppler Wind Lidar adaptive observation experiments with 3D-Var. and an ensemble Kalman filter in a global primitive equations model

Comparing Variational, Ensemble-based and Hybrid Data Assimilations at Regional Scales

Improving GFS 4DEnVar Hybrid Data Assimilation System Using Time-lagged Ensembles

Quarterly Journal of the Royal Meteorological Society !"#$%&'&(&)"&'*'+'*%(,#$,-$#*'."(/'*0'"(#"(1"&)23$)(4#$2#"( 5'$*)6!

UNIVERSITY OF OKLAHOMA GRADUATE COLLEGE EFFECTS OF SEQUENTIAL OR SIMULTANEOUS ASSIMILATION OF OBSERVATIONS AND LOCALIZATION METHODS ON THE PERFORMANCE

Recent Advances in EnKF

Observation Bias Correction with an Ensemble Kalman Filter

Analysis sensitivity calculation in an Ensemble Kalman Filter

ABSTRACT. from a few minutes (e.g. cumulus convection) to years (e.g. El Niño). Fast time

Ensemble square-root filters

A local ensemble transform Kalman filter data assimilation system for the NCEP global model

Brian J. Etherton University of North Carolina

Aspects of the practical application of ensemble-based Kalman filters

An implementation of the Local Ensemble Kalman Filter in a quasi geostrophic model and comparison with 3D-Var

arxiv: v1 [physics.ao-ph] 23 Jan 2009

Introduction to ensemble forecasting. Eric J. Kostelich

EARLY ONLINE RELEASE

A Matrix-Free Posterior Ensemble Kalman Filter Implementation Based on a Modified Cholesky Decomposition

A new Hierarchical Bayes approach to ensemble-variational data assimilation

Accepted in Tellus A 2 October, *Correspondence

MAXIMUM LIKELIHOOD ENSEMBLE FILTER: THEORETICAL ASPECTS. Milija Zupanski

A data-driven method for improving the correlation estimation in serial ensemble Kalman filter

Ensemble 4DVAR for the NCEP hybrid GSI EnKF data assimilation system and observation impact study with the hybrid system

Data assimilation in high dimensions

ESTIMATING CORRELATIONS FROM A COASTAL OCEAN MODEL FOR LOCALIZING AN ENSEMBLE TRANSFORM KALMAN FILTER

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications

Monthly Weather Review (Proof Only)

A Moment Matching Particle Filter for Nonlinear Non-Gaussian. Data Assimilation. and Peter Bickel

Organization. I MCMC discussion. I project talks. I Lecture.

The Ensemble Kalman Filter: Theoretical Formulation and Practical Implementation

Implications of the Form of the Ensemble Transformation in the Ensemble Square Root Filters

A HYBRID ENSEMBLE KALMAN FILTER / 3D-VARIATIONAL ANALYSIS SCHEME

Variable localization in an Ensemble Kalman Filter: application to the carbon cycle data assimilation

Advances and Challenges in Ensemblebased Data Assimilation in Meteorology. Takemasa Miyoshi

State and Parameter Estimation in Stochastic Dynamical Models

Simultaneous estimation of covariance inflation and observation errors within an ensemble Kalman filter

Ensemble Data Assimilation and Uncertainty Quantification

Local Predictability of the Performance of an. Ensemble Forecast System

Adaptive ensemble Kalman filtering of nonlinear systems. Tyrus Berry and Timothy Sauer George Mason University, Fairfax, VA 22030

Limited-Area Ensemble-Based Data Assimilation

The Ensemble Kalman Filter: Theoretical Formulation and Practical Implementation

Multivariate localization methods for ensemble Kalman filtering

Data Assimilation for Dispersion Models

4D-Var or Ensemble Kalman Filter?

Transcription:

Local Ensemble Transform Kalman Filter Brian Hunt 11 June 2013

Review of Notation Forecast model: a known function M on a vector space of model states. Truth: an unknown sequence {x n } of model states to be estimated. Observations: a sequence {y o n } of vectors in observation space. Forward operator: a known function H n from model space to observation space. Observation error: ε n = y o n H n (x n ). Background (before data): superscript b. Analysis (after data): superscript a.

Kalman Filter Review The Kalman Filter tracks a mean xn b,a and covariance Pn b,a representing a Gaussian distribution of model states. The analysis step assumes Gaussian observation errors and computes (exactly if H n is linear) xn a and Pn a from xn b, Pn, b and the observation error covariance R n. The forecast step sets xn+1 b = M(x n a ). If M is nonlinear, there is no right way to propagate the covariance from Pn a to Pn+1 b.

Practical Difficulties If x is m-dimensional, then P is an m-by-m matrix If m is large, the computational cost of storing and manipulating P may be prohibitive. Linearizing the model around the forecast trajectory (as required by the Extended Kalman Filter) can be expensive. Ensemble Kalman Filters make a low-rank approximation to P and avoid explicit linearization of the forecast model.

Ensemble Kalman Filters Introduced by Geir Evensen [1994]. Use the sample mean and covariance of an ensemble of forecasts as the Kalman filter background. Form an ensemble of states that matches the Kalman Filter analysis mean and covariance and use it to initialize the next ensemble forecast. Assumption: the model is approximately linear within the range of the ensemble.

Ensemble Analysis Subscripts now index ensemble members and not time. Input: background ensemble x b 1,..., x b k with sample mean x b and covariance P b ; observation information y o, H, R. Output: analysis ensemble x a 1,..., x a k with sample mean x a and covariance P a determined as in the Kalman Filter.

Kalman Filter Equations Recall the Kalman filter analysis equations: K =P b H T [HP b H T + R] 1 x a = x b + K (y o H x b ) P a =(I KH)P b Note: x a x b is called the analysis increment. What should we do if H is nonlinear? How do we determine the analysis ensemble?

Nonlinear Forward Operator H Write P b = X b (X b ) T wherex b is the matrix whose ith column is (x b i x b )/ k 1. Let Y b = HX b if H is linear.; then: K =X b (Y b ) T [Y b (Y b ) T + R] 1 x a = x b + K (y o H x b ) P a =(X b KY b )(X b ) T For nonlinear H, let yi b = H(xi b ), let ȳ b be the mean of the yi b, and let Y b be a matrix whose ith column is (yi b ȳ b )/ k 1.

Perturbed Observations There are many possible analysis ensembles whose mean and covariance are x a and P a. One can apply the Kalman Filter update to each ensemble member, perturbing the observed values differently for each: x a i = x b i + K (y o + ε i y b i ) where ε i N(0, R) [Burgers et al. 1998, Houtekamer & Mitchell 1998]. Then if H is linear, one can show that the expected mean and covariance of the analysis ensemble are x a and P a.

Deterministic Approaches Square root filters track a matrix square root of the model state covariance matrix; in our case X b for which P b = X b (X b ) T. We seek a matrix X a of (scaled) analysis ensemble perturbations such that P a = X a (X a ) T. Add x a to the (scaled) columns of X a to get the analysis ensemble. Various approaches to determining X a from X b : EAKF [Anderson 2001], ETKF [Bishop et al. 2001, Wang et al. 2004], EnSRF [Whitaker & Hamill 2002]; see also [Tippett et al. 2003].

Reduced Rank: Pros and Cons The background covariance allows analysis increments only in the space spanned by the ensemble perturbations. Indeed, in all the approaches on the last two slides, each analysis ensemble member can be written x a i = x b + X b w i. If the ensemble is too small, the filter may fail. If an ensemble of reasonable size is successful, analysis computations can be done in the ensemble space spanned by the background ensemble perturbations.

Underlying Assumptions A linear combination of background ensemble states is a plausible state. The truth lies reasonably close to the ensemble space. More precisely, most of the uncertainty in the background state lies in the ensemble space. In particular, forecast uncertainty lies in a relatively low-dimensional space.

A 2.5 Day Ensemble Forecast

A 4.5 Day Ensemble Forecast

Spatial Localization For a spatially extended system, it may not be feasible to forecast an ensemble large enough to span the possible global states. If long-distance correlations are weak, the ensemble covariance will represent spurious long-distance correlations. Localization reduces the effect of spurious correlations and allows analysis increments from a higher dimensional space.

Localization Approaches A standard approach is to localize the background covariance P b by multiplying each element by a factor between 1 (for variables at the same grid point) and 0 (for variables farther apart than a chosen localization distance. Our approach [LEKF, Ott et al. 2004; LETKF, Hunt et al. 2007] is do a separate analysis for each grid point using only observations from within a local region that we choose. These analyses can be done in parallel.

Inflation does Time Localization Many filters use multiplicative covariance inflation, which at each step multiplies the forecast (background) covariance by ρ for an ad hoc parameter ρ > 1. This is equivalent to multiplying the term corresponding to observations from t analysis cycles ago in the least squares cost function by a weight ρ t. Covariance inflation compensates for effects of model error, model nonlinearity, etc.

LETKF Formalism LETKF stands for Local Ensemble Transform Kalman Filter; equivalent to LEKF but formulated more like ETKF. Recall P b = X b (X b ) T wherex b is a matrix whose ith column is (x b i x b )/ k 1. Consider model states x = x b + X b w where w is a k-dimensional vector. If w has mean 0 and covariance I, then x has mean x b and covariance P b.

Analysis in Ensemble Space Let yi b = H(xi b ), let ȳ b be their mean, and form the matrix Y b of perturbations like X b. Make the linear approximation: H( x b + X b w) ȳ b + Y b w Minimize the cost function: J(w) =ρ 1 w T w Compare to: + (ȳ b + Y b w y o ) T R 1 (ȳ b + Y b w y o ) J(x) =(x x b ) T (ρp b ) 1 (x x b ) + (H(x) y o ) T R 1 (H(x) y o )

Analysis in Ensemble Space The analysis mean w a and covariance A are: A =[I + (Y b ) T R 1 Y b ] 1 w a =A(Y b )R 1 (y o ȳ b ) Then x a = x b + X b w a and P a = X b A(X b ) T. Notice that the analysis equations in w coordinates depend only on the background ensemble {yi b } in observation space and the observation data y o and R. The matrix that is inverted to find A is k-by-k and has no small eigenvalues.

Asynchronous Observations With this formulation, the filter easily becomes 4D (properly takes into account temporal information) when observations are asynchronous. When mapping the background ensemble members xi b into observation space, use the background state at the appropriate time for each observation. Assumption: a linear combination of ensemble trajectories is an approximate model trajectory.

Choice of Analysis Ensemble To form the analysis ensemble vectors wi a, add to w a the columns of the symmetric square root W a = [(k 1)A] 1/2. The analysis ensemble xi a = x b + X b wi a then has the correct mean and covariance. Other choices are possible, but this choice minimizes the change (in w, or in x with P b -norm) between the background and analysis ensembles. Recall that we determine a different wi a for each grid point, using only observations from a local region near that grid point.

Weight Interpolation If the localization distance is large relative to the grid spacing, overlapping local regions make nearby analyses consistent. For a high-resolution model, the overlap causes a lot of computational redundancy. Solution: Compute the weights wi a on a coarse grid of analysis points and interpolate them to the other grid points. Can improve balance [Yang et al. 2009].

Tapering Observation Influence At a given grid point, our local analysis (as described so far) uses nearby observations at full influence, while distant observations have zero influence. The influence can be tapered more smoothly from 1 to 0 by multiplying each diagonal entry σ 2 j in R 1 by a factor 0 α j 1 that depends on the distance from the corresponding observation location to the analysis grid point. If R is not diagonal, replace R 1 by DR 1 D, where D is diagonal with entries α j.

Toy Model Results Perfect model test with Lorenz-96 system of coupled ODEs on a circular lattice: dx m /dt = (x m+1 x m 2 )x m 1 x m + 8. We used a model trajectory as the truth and add noise to simulate observations. We compared our LEKF, using only observations from within 6 grid points, with a global ensemble Kalman filter (FEKF) [Ott et al. 2004].

Global vs. Local Filter M = model size, k+1 = ensemble size, E = error

Comparison to NCEP (2004) 3D-Var We ran the U.S. National Centers for Environmental Prediction global forecast model (GSM) at reduced T62 resolution (about 500, 000 grid points). We compared our LETKF analyses with analyses from NCEP s 3D-Var system (SSI) using actual Winter 2004 non-satellite observations (about 300, 000 per 6 hours). We used a 60-member ensemble, 800 km radius local region, spatially varying inflation ρ 1.25 [Szunyogh et al. 2008].

48-hour Forecast Error (compared to radiosonde observations) + = our LETKF o = NWS 3D-Var

Computational Speed At NCEP, less than 10 minutes of every 6 hour cycle is used for data assimilation. Our implementation took about 15 minutes on a 40-processor Linux cluster. The computation time is approximately: linear in the number of observations; linear in the number of model grid points; quadratic in the number of ensemble members.

Extensions Parameter/Bias Estimation: treat parameters of M or H as state variables with time derivative zero (Baek et al. 2006, 2009; Cornick et al. 2009; Fertig et al. 2009). Nongaussian Filter: minimize nonquadratic cost function numerically in ensemble space (Harlim & Hunt 2007), as in MLEF (Zupanski 2006).

Conclusions LETKF is a relatively simple, efficient, and flexible framework for reduced-rank data assimilation that works in practice. The method is largely model-independent. It scales well to large systems. Full citations for references to our group s work are on the publications page at: http://www.weatherchaos.umd.edu Motto: Forecast globally, assimilate locally.