Forecasting and data assimilation

Similar documents
Ensemble Data Assimilation and Uncertainty Quantification

Lagrangian Data Assimilation and Its Application to Geophysical Fluid Flows

Bayesian Inverse problem, Data assimilation and Localization

DART_LAB Tutorial Section 2: How should observations impact an unobserved state variable? Multivariate assimilation.

Introduction to Ensemble Kalman Filters and the Data Assimilation Research Testbed

Local Ensemble Transform Kalman Filter

Data Assimilation Research Testbed Tutorial

Numerical Weather Prediction: Data assimilation. Steven Cavallo

A new Hierarchical Bayes approach to ensemble-variational data assimilation

Convective-scale data assimilation in the Weather Research and Forecasting model using a nonlinear ensemble filter

Ergodicity in data assimilation methods

Relative Merits of 4D-Var and Ensemble Kalman Filter

Gaussian Process Approximations of Stochastic Differential Equations

Data assimilation in high dimensions

Data assimilation in high dimensions

Ensemble Kalman Filter

The Ensemble Kalman Filter:

EnKF Review. P.L. Houtekamer 7th EnKF workshop Introduction to the EnKF. Challenges. The ultimate global EnKF algorithm

The Kalman Filter ImPr Talk

arxiv: v1 [physics.ao-ph] 23 Jan 2009

Autonomous Navigation for Flying Robots

Fundamentals of Data Assimila1on

Organization. I MCMC discussion. I project talks. I Lecture.

Fundamentals of Data Assimila1on

Some Applications of WRF/DART

Stability of Ensemble Kalman Filters

Models for models. Douglas Nychka Geophysical Statistics Project National Center for Atmospheric Research

Par$cle Filters Part I: Theory. Peter Jan van Leeuwen Data- Assimila$on Research Centre DARC University of Reading

Bayesian Statistics and Data Assimilation. Jonathan Stroud. Department of Statistics The George Washington University

The Matrix Reloaded: Computations for large spatial data sets

Data Assimilation: Finding the Initial Conditions in Large Dynamical Systems. Eric Kostelich Data Mining Seminar, Feb. 6, 2006

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Gaussian Filtering Strategies for Nonlinear Systems

A Note on the Particle Filter with Posterior Gaussian Resampling

A data-driven method for improving the correlation estimation in serial ensemble Kalman filter

Adaptive Observation Strategies for Forecast Error Minimization

LARGE-SCALE TRAFFIC STATE ESTIMATION

Bayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e.

Robust Ensemble Filtering With Improved Storm Surge Forecasting

Spectral Ensemble Kalman Filters

DATA ASSIMILATION FOR FLOOD FORECASTING

Short tutorial on data assimilation

Numerical Weather prediction at the European Centre for Medium-Range Weather Forecasts

Four-Dimensional Ensemble Kalman Filtering

Hierarchical Bayes Ensemble Kalman Filter

1. Current atmospheric DA systems 2. Coupling surface/atmospheric DA 3. Trends & ideas

(Extended) Kalman Filter

The priority program SPP1167 Quantitative Precipitation Forecast PQP and the stochastic view of weather forecasting

Demonstration and Comparison of of Sequential Approaches for Altimeter Data Assimilation in in HYCOM

Ensemble Kalman Filters for WRF-ARW. Chris Snyder MMM and IMAGe National Center for Atmospheric Research

A Non-Gaussian Ensemble Filter Update for Data Assimilation

The Local Ensemble Transform Kalman Filter (LETKF) Eric Kostelich. Main topics

Ji-Sun Kang. Pr. Eugenia Kalnay (Chair/Advisor) Pr. Ning Zeng (Co-Chair) Pr. Brian Hunt (Dean s representative) Pr. Kayo Ide Pr.

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino

Local Ensemble Transform Kalman Filter: An Efficient Scheme for Assimilating Atmospheric Data

Multi-resolution models for large data sets

Overview of data assimilation in oceanography or how best to initialize the ocean?

Uncertainty in Models of the Ocean and Atmosphere

The Matrix Reloaded: Computations for large spatial data sets

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Data Assimilation Research Testbed Tutorial

C. Gebhardt, S. Theis, R. Kohlhepp, E. Machulskaya, M. Buchhold. developers of KENDA, ICON-EPS, ICON-EDA, COSMO-D2, verification

Principles of the Global Positioning System Lecture 11

Accepted in Tellus A 2 October, *Correspondence

Smoothers: Types and Benchmarks

Lagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model. David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC

Application of the Ensemble Kalman Filter to History Matching

Relationship between Singular Vectors, Bred Vectors, 4D-Var and EnKF

The analog data assimilation: method, applications and implementation

A Global Atmospheric Model. Joe Tribbia NCAR Turbulence Summer School July 2008

State and Parameter Estimation in Stochastic Dynamical Models

An Efficient Ensemble Data Assimilation Approach To Deal With Range Limited Observation

Hybrid variational-ensemble data assimilation. Daryl T. Kleist. Kayo Ide, Dave Parrish, John Derber, Jeff Whitaker

L06. LINEAR KALMAN FILTERS. NA568 Mobile Robotics: Methods & Algorithms

Estimating Observation Impact in a Hybrid Data Assimilation System: Experiments with a Simple Model

Correcting biased observation model error in data assimilation

4. DATA ASSIMILATION FUNDAMENTALS

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance

Abstract 2. ENSEMBLE KALMAN FILTERS 1. INTRODUCTION

Satellite Retrieval Assimilation (a modified observation operator)

Lagrangian data assimilation for point vortex systems

Methods of Data Assimilation and Comparisons for Lagrangian Data

Adaptive Data Assimilation and Multi-Model Fusion

Lecture 9. Time series prediction

Data Assimilation for Dispersion Models

Ensemble square-root filters

Gaussian Mixture Filter in Hybrid Navigation

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

Particle filters, the optimal proposal and high-dimensional systems

ESTIMATING CORRELATIONS FROM A COASTAL OCEAN MODEL FOR LOCALIZING AN ENSEMBLE TRANSFORM KALMAN FILTER

Toward improved initial conditions for NCAR s real-time convection-allowing ensemble. Ryan Sobash, Glen Romine, Craig Schwartz, and Kate Fossell

2D Image Processing. Bayes filter implementation: Kalman filter

Lecture 7: Optimal Smoothing

Estimating Observation Impact in a Hybrid Data Assimilation System: Experiments with a Simple Model

9 Multi-Model State Estimation

Probabilistic Fundamentals in Robotics. DAUIN Politecnico di Torino July 2010

Weather generators for studying climate change

A Moment Matching Particle Filter for Nonlinear Non-Gaussian. Data Assimilation. and Peter Bickel

(Toward) Scale-dependent weighting and localization for the NCEP GFS hybrid 4DEnVar Scheme

Transcription:

Supported by the National Science Foundation DMS Forecasting and data assimilation Outline Numerical models Kalman Filter Ensembles Douglas Nychka, Thomas Bengtsson, Chris Snyder Geophysical Statistics Project National Center for Atmospheric Research Local-Local filter for non-gaussian distributions

Overview Data Assimilation (DA): Combining predictions made by a numerical model with observed data to estimate the state of a system, x. This is also called a filter. The statistical foundation is Bayes Theorem and the uncertainty in the state of the system is represented by a probability distribution. PRIOR for x + observations POSTERIOR for x Usually the assimilation is done at many consecutive time points and the practical implementation involves many shortcuts to approximate a posterior distribution.

Why are we doing this? Forecast the weather Assimilating data for a given geophysical model may be one of the few ways to test it. Some key ideas: Represent a continuous distribution by a random sample. Only update state variables local to the observations Use local regression to update the state variables Contribution: A Local-Local filter to handle non-gaussian data assimilation problems.

Atmospheric models and forecasting

Atmospheric models 101 A deterministic numerical model that describes the circulation of the atmosphere. State of system defined on a 3-d grid of the the atmosphere. Community climate system model (CCSM2) 128 64 30 boxes (200km) Rapid Update Cycle Model (RUC) is run on part of the earth but on a 40km grid. Evolution of the model is governed by a discretizing the nonlinear equations of motion derived from fluid dynamics, usually deterministic. x t+1 = g(x t ) g is nonlinear, complicated and fairly expensive to evaluate Making a deterministic Forecast: Given ˆx t, ˆx t+1 = g( ˆx t )

40-Dimensional Lorenz System (Lorenz, 1996) Atmospheric system describing k values of an atmospheric variable at k longitudes: x 1,..., x 40. (Subscript denotes spatial location.) Equations: for j = 1,..., 40, where F represents forcing. ẋ j = x j 1 (x j+1 x j 2 ) x j + F, The equations contain quadratic nonlinearities mimicking advection: u i u i u i x u i(u i u i )/δ. F is chosen so that phase space is bounded and the system exhibits chaotic behavior. observe z 2, z 4,..., z 40 : y j = z j + ɛ j, ɛ j N(0, 4) and δt =.40.

Observational network and forecasting Rapid Update Cycle (RUC) model: rawinsondes (balloons) and special dropwinsondes commercial aircraft wind profilers surface reporting stations and buoys Radio Acoustic Sounding System - experimental various satellite data...

The Bayes cycle p(x t ), y t Bayes p(x t y t ) g(.) p(x t+1 y t ) = p(x t+1 ), y t+1 Yesterday s posterior becomes today s prior!

Standard Kalman Filter/ conditional multivariate normal distributions This is easy in closed form if everything is multivariate normal and linear. Observation Model with Prior Kalman update for state y = Hx t + e e MN(0, R) x t MN(µ t, Σ) ˆx t = E(x t y) = µ t + H T (HΣH T + R) 1 (y Hµ t ) Kalman update for covariance V AR(x t y) = P t a = Σ H T (HΣH T + R) 1 H

Forecast mean: Assume that g is linear ˆx t+1 = Gˆx t Forecast covariance: P t+1 f,t = GP t ag t A qualifier problem: These are just results based on the conditional distributions for the multivariate normal because everything is assumed to be Gaussian or a linear transformation.

Problems x 10 6 10 7 and y 10 5 10 6 So even with closed form expressions the computations may not be feasible because the linear systems are huge. Finding from p(x t+1 y t ) = p(g(x t ) y t ) p(x t y t ) is the mother of all change of variable problems! P f can not be stored or directly propagated.

Ensembles Each distribution is represented by a random sample of the states called an ensemble. In place of propagate each ensemble member. x t,1 x t,2. x t,m π(x t y t ) π(g(x t ) y t ) g g(x t,1 ) = x t+1,1 g(x t,2 ) = x t+1,2. g(x t,m ) = x t+1,m By elementary probability: {x t,j } is a random sample from p(x t y t ) implies {x t+1,j } will be a random sample from p(x t+1 y t )

Ensemble Kalman filter (EKF) If the observations have independent errors, the observations can be assimilated sequentially to get the same result. Wherever a covariance matrix or mean vector appears replace these by the sample quantities from the ensemble. The covariance matrix is tapered to be a better estimate and inflated to make the filter stable. Sampling to get the new ensemble for posterior is computed in a very similar way as the standard update ( perturbed observation method).

Ensemble Kalman filter (EKF) (continued) The first point suggests a double loop algorithm: Assimilating at a given time: Loop over observations {y 1, y 2,...y n } Loop over ensemble members: {x 1, x 2,..., x M } Update each ensemble member x i based on y j A key aspect is that the observation only changes part of the state vector that is close to it due to the covariance tapering. Given an observation at RDU, update a Greensboro,NC grid point... but a grid point near Moscow is unchanged. It is an open question how the approximations and tuning parameters in the EKF change its statistical performance. Also given that g is nonlinear one can not expect Gaussian distributions.

4 How an Ensemble Filter Works Theory: Impact of observations can be handled sequentially Impact of observation on each state variable can be handled sequentially A. Integrate model ensemble to time at which observation becomes available. B. Observed value and observational error distribution from observing system. H H H C. Get prior ensemble sample of observation by applying H to each member of ensemble. E. (Step 2) Use linear regression to compute corresponding increments for each state variable. D. (Step 1) Find increments for prior estimate of observation. /home/jla/dart/ncar_arg_seminar/talk.fm March 31, 2003 2:52 pm

Same idea on a 3 dimensional system

Non-Gaussian distributions Represent the prior distributions as mixtures of multivariate normals p(x t ) = k i=1 p imn(µ i, P i ) The posterior distribution is also a mixture: p(x t y t ) = k i=1 p i MN(µ i, P i ) Ensembles as a mixture distribution: Each ensemble member is the center of a mixture where the covariance is the sample covariance of its nearest neighbors. From the form for the posterior mean using ensembles the posterior probabilities look like weights based on a normal kernel. The use of neighborhoods to finds the covariance results in a local linear regression. Thus in some strange way we have reinvented LOESS, local linear regression!

A non-gaussian update as a local regression Observe X2 with error and wish to update X1 and X3.

A non-gaussian example

Local-local filter Curse of dimensionality: 40 dimensions is too large a state space to apply the mixture ensemble filter directly. In 40 dimensions every state vector is far away from every other! Basic idea is to only use the mixture model for components of the state vector close to the observations. Otherwise use the usual EKF for updating components.

Summary Results We have some evidence that the practical version of the EKF actually handles nongaussian distributions better than an exact Kalman filter. The Local-Local filter clearly out performs the EKF in a simple 3-d system especially in places where g is very nonlinear. A version of the L-L filter also performs better than EKF with about 5% improvement (without any extensive tuning) for the 40 variable model. Some Future Work Adaptive estimates of tuning parameters Robust estimators Exploring more realistic test systems, e.g. primitive equation models for a dry atmosphere.