Data assimilation in high dimensions

Similar documents
Data assimilation in high dimensions

Ergodicity in data assimilation methods

What do we know about EnKF?

EnKF and filter divergence

EnKF and Catastrophic filter divergence

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications

Filtering Sparse Regular Observed Linear and Nonlinear Turbulent System

Ensemble Kalman Filter

Performance of ensemble Kalman filters with small ensembles

A Note on the Particle Filter with Posterior Gaussian Resampling

Lecture 2: From Linear Regression to Kalman Filter and Beyond

(Extended) Kalman Filter

A variance limiting Kalman filter for data assimilation: I. Sparse observational grids II. Model error

State and Parameter Estimation in Stochastic Dynamical Models

A data-driven method for improving the correlation estimation in serial ensemble Kalman filter

Bayesian Calibration of Simulators with Structured Discretization Uncertainty

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Bayesian Inverse problem, Data assimilation and Localization

A new Hierarchical Bayes approach to ensemble-variational data assimilation

The Ensemble Kalman Filter:

Lagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model. David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC

Lagrangian Data Assimilation and Its Application to Geophysical Fluid Flows

Forecasting and data assimilation

Kalman Filter and Ensemble Kalman Filter

A Stochastic Collocation based. for Data Assimilation

Local Ensemble Transform Kalman Filter

Information Flow/Transfer Review of Theory and Applications

The Ensemble Kalman Filter:

Application of the Ensemble Kalman Filter to History Matching

Methods of Data Assimilation and Comparisons for Lagrangian Data

Organization. I MCMC discussion. I project talks. I Lecture.

Superparameterization and Dynamic Stochastic Superresolution (DSS) for Filtering Sparse Geophysical Flows

Dynamic System Identification using HDMR-Bayesian Technique

Adaptive ensemble Kalman filtering of nonlinear systems

Ensemble Data Assimilation and Uncertainty Quantification

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Correcting biased observation model error in data assimilation

Gaussian Filtering Strategies for Nonlinear Systems

Data assimilation with and without a model

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

Sequential Monte Carlo Samplers for Applications in High Dimensions

Review of Covariance Localization in Ensemble Filters

An Efficient Ensemble Data Assimilation Approach To Deal With Range Limited Observation

Smoothers: Types and Benchmarks

Systematic stochastic modeling. of atmospheric variability. Christian Franzke

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

Modeling and state estimation Examples State estimation Probabilities Bayes filter Particle filter. Modeling. CSC752 Autonomous Robotic Systems

Data Assimilation for Dispersion Models

2D Image Processing. Bayes filter implementation: Kalman filter

Stochastic Analogues to Deterministic Optimizers

Short tutorial on data assimilation

Systematic strategies for real time filtering of turbulent signals in complex systems

Relative Merits of 4D-Var and Ensemble Kalman Filter

A mechanism for catastrophic filter divergence in data assimilation for sparse observation networks

Many contemporary problems in science ranging from protein

Gaussian Process Approximations of Stochastic Differential Equations

Part 1: Expectation Propagation

Efficient Data Assimilation for Spatiotemporal Chaos: a Local Ensemble Transform Kalman Filter

arxiv: v1 [physics.ao-ph] 23 Jan 2009

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Bayesian Learning. Tobias Scheffer, Niels Landwehr

Robust Ensemble Filtering With Improved Storm Surge Forecasting

Data assimilation with and without a model

Lecture : Probabilistic Machine Learning

Convective-scale data assimilation in the Weather Research and Forecasting model using a nonlinear ensemble filter

Superparameterization and Dynamic Stochastic Superresolution (DSS) for Filtering Sparse Geophysical Flows

New Fast Kalman filter method

Seminar: Data Assimilation

Particle filters, the optimal proposal and high-dimensional systems

EKF and SLAM. McGill COMP 765 Sept 18 th, 2017

Stability of Ensemble Kalman Filters

Implicit sampling for particle filters. Alexandre Chorin, Mathias Morzfeld, Xuemin Tu, Ethan Atkins

Convergence of the Ensemble Kalman Filter in Hilbert Space

Lecture 6: Bayesian Inference in SDE Models

Data assimilation as an optimal control problem and applications to UQ

Why do we care? Examples. Bayes Rule. What room am I in? Handling uncertainty over time: predicting, estimating, recognizing, learning

Fundamentals of Data Assimila1on

Bayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e.

2D Image Processing. Bayes filter implementation: Kalman filter

Der SPP 1167-PQP und die stochastische Wettervorhersage

Nonlinear Model Reduction for Uncertainty Quantification in Large-Scale Inverse Problems

Adaptive Data Assimilation and Multi-Model Fusion

Introduction to Machine Learning

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Machine Learning Basics: Maximum Likelihood Estimation

Forecasting Turbulent Modes with Nonparametric Diffusion Models: Learning from Noisy Data

Assimilation des Observations et Traitement des Incertitudes en Météorologie

A non-parametric regression model for estimation of ionospheric plasma velocity distribution from SuperDARN data

Fundamentals of Data Assimila1on

Sensitivity to Model Parameters

Sampling the posterior: An approach to non-gaussian data assimilation

Why do we care? Measurements. Handling uncertainty over time: predicting, estimating, recognizing, learning. Dealing with time

On a Data Assimilation Method coupling Kalman Filtering, MCRE Concept and PGD Model Reduction for Real-Time Updating of Structural Mechanics Model

DART_LAB Tutorial Section 2: How should observations impact an unobserved state variable? Multivariate assimilation.

A Spectral Approach to Linear Bayesian Updating

State Space Models for Wind Forecast Correction

Bayesian parameter estimation in predictive engineering

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Local Ensemble Transform Kalman Filter: An Efficient Scheme for Assimilating Atmospheric Data

Transcription:

Data assimilation in high dimensions David Kelly Courant Institute New York University New York NY www.dtbkelly.com February 12, 2015 Graduate seminar, CIMS David Kelly (CIMS) Data assimilation February 12, 2015 1 / 20

What is data assimilation? Suppose u satisfies du = F (u)dt + dw with some unknown initial condition u 0. We are most interested in geophysical models, so think high dimensional, nonlinear, stochastic. Suppose we make partial, noisy observations at times t = h, 2h,..., nh,... y n = Hu n + ξ n where H is a linear operator (think low rank projection), u n = u(nh), and ξ n N(0, Γ) iid. The aim of data assimilation is to say something about the conditional distribution of u n given the observations {y 1,..., y n } David Kelly (CIMS) Data assimilation February 12, 2015 2 / 20

Outline 1 - The basics: Bayes, Kalman etc. 2 - What to do for nonlinear models? 3 - What to do in high dimensions? David Kelly (CIMS) Data assimilation February 12, 2015 3 / 20

Illustration (Initialization) y Ψ obs x Figure: The blue circle represents our guess of u 0. Due to the uncertainty in u 0, this is a probability measure. David Kelly (CIMS) Data assimilation February 12, 2015 4 / 20

Illustration (Forecast step) y Ψ obs x Figure: Apply the time h flow map Ψ. This produces a new probability measure which is our forecasted estimate of u 1. This is called the forecast step. David Kelly (CIMS) Data assimilation February 12, 2015 4 / 20

Illustration (Make an observation) y Ψ obs x Figure: We make an observation y 1 = Hu 1 + ξ 1. In the picture, we only observe the x variable. David Kelly (CIMS) Data assimilation February 12, 2015 4 / 20

Illustration (Analysis step) y Ψ obs x Figure: Using Bayes formula we compute the conditional distribution of u 1 y 1. This new measure (called the posterior) is the new estimate of u 1. The uncertainty of the estimate is reduced by incorporating the observation. The forecast distribution steers the update from the observation. P(u 1 y 1 ) P(y 1 u 1 )P(u 1 ) David Kelly (CIMS) Data assimilation February 12, 2015 4 / 20

Bayes formula Let Y n = {y 0, y 1,..., y n }. We want to compute the conditional density P(u n+1 Y n+1 ), using P(u n Y n ) and y n+1. By Bayes formula, we have P(u n+1 Y n+1 ) = P(u n+1 Y n, y n+1 ) P(y n+1 u n+1 )P(u n+1 Y n ) In the stochastic (Markovian) case we need to compute the integral P(u n+1 Y n ) = P(u n+1 Y n, u n )P(u n Y n )du n. David Kelly (CIMS) Data assimilation February 12, 2015 5 / 20

Animation 1 Suppose the model is du = V (u)dt + σ 1/2 dw with u = (u x, u y ) and V (x, y) = 1 2 (1 x 2 y 2 ) 2. We only observe the x-variable y n = u x (nh) + γ 1/2 ξ n with ξ n N(0, 1) iid. David Kelly (CIMS) Data assimilation February 12, 2015 6 / 20

In geophysical models, we can have u R N where N = O(10 9 ). The rigorous Bayesian approach is computationally infeasible. David Kelly (CIMS) Data assimilation February 12, 2015 7 / 20

The Kalman Filter For linear models, the Bayesian integral is Gaussian and can be computed explicitly. The conditional density is characterized by its mean and covariance where m n+1 = (1 K n+1 H) m n + K n+1 y n+1 C n+1 = (I K n+1 H)Ĉ n+1, ( m n+1, Ĉ n+1) is the forecast mean and covariance. K n+1 = Ĉ n+1h T (Γ + HĈ n+1h T ) 1 is the Kalman gain. The procedure of updating (m n, C n ) (m n+1, C n+1 ) is known as the Kalman filter. David Kelly (CIMS) Data assimilation February 12, 2015 8 / 20

Extended Kalman filter Suppose we have a nonlinear model: u n+1 = Φ(u n ) + Σ 1/2 η n where Φ is a nonlinear map, η n Gaussian. The Extended Kalman filter is given by the same update formulas m n+1 = (1 K n+1 H) m n+1 + K n+1 y n+1 C n+1 = (I K n+1 H)Ĉ n+1, where m n+1 = Φ(m n ) and Ĉ n+1 = DΦ(m n )C n DΦ(m n ) T + Σ. Thus we approximate the forecast distribution with a Gaussian. David Kelly (CIMS) Data assimilation February 12, 2015 9 / 20

Back to example Suppose we approximate du = V (u)dt + σ 1/2 dw with the discrete time formulation u n+1 = Φ(u n ) + Σ 1/2 η n+1 with η n Gaussian. Then it is easy to compute the Kalman update m n+1 = Φ(m n ) + (γ + Ĉ xx) 1 (y n Φ x (m n )) and Ĉ n+1 = DΦ(m n )C n DΦ(m n ) T + Σ. [Ĉ xx Ĉ xy ] David Kelly (CIMS) Data assimilation February 12, 2015 10 / 20

Computing DΦ(x) means evaluating Φ once for each degree of freedom. We want to get away with something cheaper. David Kelly (CIMS) Data assimilation February 12, 2015 11 / 20

Ensemble Kalman filter (Evensen 94) y Ψ obs x Figure: Start with K ensemble members drawn from some distribution. Empirical representation of u 0. The ensemble members are denoted v (k) 0. Only KN numbers are stored. Better than Kalman if K < N. David Kelly (CIMS) Data assimilation February 12, 2015 12 / 20

Ensemble Kalman filter (Forecast step) y Ψ obs x Figure: Apply the dynamics Ψ to each ensemble member. David Kelly (CIMS) Data assimilation February 12, 2015 12 / 20

Ensemble Kalman filter (Make obs) y Ψ Figure: Make an observation. obs x David Kelly (CIMS) Data assimilation February 12, 2015 12 / 20

Ensemble Kalman filter (Perturb obs) y Ψ obs x Figure: Turn the observation into K artificial observations by perturbing by the same source of observational noise. y (k) 1 = y 1 + ξ (k) 1 David Kelly (CIMS) Data assimilation February 12, 2015 12 / 20

Ensemble Kalman filter (Analysis step) y Ψ obs x Figure: Update each member using the Kalman update formula. The Kalman gain K 1 is computed using the ensemble covariance. v (k) 1 = (1 K 1 H)Ψ(v (k) 0 ) + K 1Hy (k) 1 K 1 = Ĉ 1H T (Γ + HĈ 1H T ) 1 Ĉ 1 = 1 K (Ψ(v (k) 0 K 1 ) Ψ(v 0))(Ψ(v (k) 0 ) Ψ(v 0)) T David Kelly (CIMS) k=1 Data assimilation February 12, 2015 12 / 20

Ensemble Kalman filter The conditional distribution is represented empirically using an ensemble {v (k) n } K k=1. When an observation is made, it is perturbed by an iid copy of the observational noise y (k) n+1 = y n+1 + ξ (k) n+1. Each ensemble member is updated using the Kalman update formula v (k) n+1 = (1 K n+1h)ψ(v (k) n ) + K n+1 Hy (k) n+1 and the Kalman gain is computed using the ensemble covariance K n+1 = Ĉ n+1h T (Γ + HĈ n+1h T ) 1. David Kelly (CIMS) Data assimilation February 12, 2015 13 / 20

There are many good justifications for this algorithm: When the model is linear and K is large, the ensemble members are exact samples from the conditional distribution (Monte Carlo Kalman filter). But there are no great justifications... David Kelly (CIMS) Data assimilation February 12, 2015 14 / 20

Catastrophic filter divergence Lorenz-96: u j = (u j+1 u j 2 )u j 1 u j + F with j = 1,..., 40. Periodic BCs. Observe every fifth node. (Harlim-Majda 10, Gottwald-Majda 12) True solution in a bounded set, but filter blows up to machine infinity in finite time! David Kelly (CIMS) Data assimilation February 12, 2015 15 / 20

For complicated models, only heuristic arguments offered as explanation. Can we prove it for a simpler constructive model? David Kelly (CIMS) Data assimilation February 12, 2015 16 / 20

The rotate-and-lock map (K., Majda, Tong. PNAS 15.) The model Ψ : R 2 R 2 is a composition of two maps Ψ(x, y) = Ψ lock (Ψ rot (x, y)) where ( ) ( ) ρ cos θ ρ sin θ x Ψ rot (x, y) = ρ sin θ ρ cos θ y and Ψ lock rounds the input to the nearest point in the grid G = {(m, (2n + 1)ε) R 2 : m, n Z}. It is easy to show that this model has an energy dissipation principle: for α (0, 1) and β > 0. Ψ(x, y) 2 α (x, y) 2 + β David Kelly (CIMS) Data assimilation February 12, 2015 17 / 20

(a) Figure: The red square is the trajectory u n = 0. The blue dots are the positions of the forecast ensemble Ψ(v + 0 ), Ψ(v 0 ). Given the locking mechanism in Ψ, this is a natural configuration. David Kelly (CIMS) Data assimilation February 12, 2015 18 / 20

(b) Figure: We make an observation (H shown below) and perform the analysis step. The green dots are v + 1, v 1. H = ( 1 ) 0 ε 2 1 y 1 = (ξ 1,x, ξ 1,y + ε 2 ξ 1,x ) v ± 1 (ˆx, ±ε 2ˆx/(1 + 2ε2 )) David Kelly (CIMS) Data assimilation February 12, 2015 18 / 20

(c) Figure: Beginning the next assimilation step. Apply Ψ rot to the ensemble (blue dots) David Kelly (CIMS) Data assimilation February 12, 2015 18 / 20

(d) Figure: Apply Ψ lock. The blue dots are the forecast ensemble Ψ(v + 1 ), Ψ(v 1 ). Exact same as frame 1, but higher energy orbit. The cycle repeats leading to exponential growth. David Kelly (CIMS) Data assimilation February 12, 2015 18 / 20

Theorem (K.-Majda-Tong 15 PNAS) For any N > 0 and any p (0, 1) there exists a choice of parameters such that ( ) P v (k) n M n for all n N 1 p where M n is an exponentially growing sequence. ie - The filter can be made to grow exponentially for an arbitrarily long time with an arbitrarily high probability. David Kelly (CIMS) Data assimilation February 12, 2015 19 / 20

References 1 - D. Kelly, K. Law & A. Stuart. Well-Posedness And Accuracy Of The Ensemble Kalman Filter In Discrete And Continuous Time. Nonlinearity (2014). 2 - D. Kelly, A. Majda & X. Tong. Concrete ensemble Kalman filters with rigorous catastrophic filter divergence. Proc. Nat. Acad. Sci. (2015). 3 - X. Tong, A. Majda & D. Kelly. Nonlinear stability and ergodicity of ensemble based Kalman filters. Nonlinearity (2015). 4 - X. Tong, A. Majda & D. Kelly. Nonlinear stability of the ensemble Kalman filter with adaptive covariance inflation. To appear in Comm. Math. Sci. (2015). All my slides are on my website (www.dtbkelly.com) Thank you! David Kelly (CIMS) Data assimilation February 12, 2015 20 / 20