NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE

Similar documents
Parametric Problems, Stochastics, and Identification

Sampling and low-rank tensor approximation of the response surface

Uncertainty Quantification and related areas - An overview

A Spectral Approach to Linear Bayesian Updating

Inverse Problems in a Bayesian Setting

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko

Hierarchical Parallel Solution of Stochastic Systems

Non-Intrusive Solution of Stochastic and Parametric Equations

Hierarchical matrix techniques for maximum likelihood covariance estimation

Estimating functional uncertainty using polynomial chaos and adjoint equations

Numerical Approximation of Stochastic Elliptic Partial Differential Equations

Stochastic Spectral Approaches to Bayesian Inference

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications

Curriculum Vitae. Alexander Litvenenko

Quantifying Uncertainty: Modern Computational Representation of Probability and Applications

Lecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:

PC EXPANSION FOR GLOBAL SENSITIVITY ANALYSIS OF NON-SMOOTH FUNCTIONALS OF UNCERTAIN STOCHASTIC DIFFERENTIAL EQUATIONS SOLUTIONS

CERTAIN THOUGHTS ON UNCERTAINTY ANALYSIS FOR DYNAMICAL SYSTEMS

Sampling and Low-Rank Tensor Approximations

On a Data Assimilation Method coupling Kalman Filtering, MCRE Concept and PGD Model Reduction for Real-Time Updating of Structural Mechanics Model

Performance Evaluation of Generalized Polynomial Chaos

Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering

An Empirical Chaos Expansion Method for Uncertainty Quantification

Paradigms of Probabilistic Modelling

A Posteriori Adaptive Low-Rank Approximation of Probabilistic Models

Dynamic System Identification using HDMR-Bayesian Technique

Polynomial chaos expansions for sensitivity analysis

Nonlinear Model Reduction for Uncertainty Quantification in Large-Scale Inverse Problems

At A Glance. UQ16 Mobile App.

Galerkin Methods for Linear and Nonlinear Elliptic Stochastic Partial Differential Equations

Quasi-optimal and adaptive sparse grids with control variates for PDEs with random diffusion coefficient

Sparsity in system identification and data-driven control

Bayesian Inverse problem, Data assimilation and Localization

A regularized least-squares method for sparse low-rank approximation of multivariate functions

arxiv: v1 [math.na] 14 Sep 2017

Fast Numerical Methods for Stochastic Computations

Schwarz Preconditioner for the Stochastic Finite Element Method

Polynomial Chaos and Karhunen-Loeve Expansion

Uncertainty updating in the description of heterogeneous materials

Data assimilation in high dimensions

Spectral Representation of Random Processes

Numerical Solutions of ODEs by Gaussian (Kalman) Filtering

Introduction to Uncertainty Quantification in Computational Science Handout #3

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 1: Introduction

Introduction to Computational Stochastic Differential Equations

UNIVERSITY OF CALIFORNIA, SAN DIEGO. An Empirical Chaos Expansion Method for Uncertainty Quantification

Solving the Stochastic Steady-State Diffusion Problem Using Multigrid

Intro BCS/Low Rank Model Inference/Comparison Summary References. UQTk. A Flexible Python/C++ Toolkit for Uncertainty Quantification

NONLOCALITY AND STOCHASTICITY TWO EMERGENT DIRECTIONS FOR APPLIED MATHEMATICS. Max Gunzburger

Sparse polynomial chaos expansions in engineering applications

c 2004 Society for Industrial and Applied Mathematics

Polynomial chaos expansions for structural reliability analysis

Uncertainty Quantification in Computational Models

Evaluation of Non-Intrusive Approaches for Wiener-Askey Generalized Polynomial Chaos

Adaptive low-rank approximation in hierarchical tensor format using least-squares method

Partial Differential Equations with Stochastic Coefficients

Algorithms for Uncertainty Quantification

Uncertainty Propagation and Global Sensitivity Analysis in Hybrid Simulation using Polynomial Chaos Expansion

Bayesian Calibration of Simulators with Structured Discretization Uncertainty

Dinesh Kumar, Mehrdad Raisee and Chris Lacor

Seminar: Data Assimilation

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

STA414/2104 Statistical Methods for Machine Learning II

Implementation of Sparse Wavelet-Galerkin FEM for Stochastic PDEs

Introduction to asymptotic techniques for stochastic systems with multiple time-scales

Polynomial Chaos Expansion of random coefficients and the solution of stochastic partial differential equations in the Tensor Train format

Efficient Data Assimilation for Spatiotemporal Chaos: a Local Ensemble Transform Kalman Filter

Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques

A Stochastic Collocation based. for Data Assimilation

Ergodicity in data assimilation methods

Uncertainty Quantification of Radionuclide Release Models using Non-Intrusive Polynomial Chaos. Casper Hoogwerf

Uncertainty Evolution In Stochastic Dynamic Models Using Polynomial Chaos

Multilevel Sequential 2 Monte Carlo for Bayesian Inverse Problems

Data assimilation in high dimensions

Original Research. Sensitivity Analysis and Variance Reduction in a Stochastic NDT Problem

Introduction to Gaussian Processes

Bayesian parameter estimation in predictive engineering

Ensemble Data Assimilation and Uncertainty Quantification

Data assimilation with and without a model

Long Time Propagation of Stochasticity by Dynamical Polynomial Chaos Expansions. Hasan Cagan Ozen

Uncertainty analysis of large-scale systems using domain decomposition

A new Hierarchical Bayes approach to ensemble-variational data assimilation

The Bayesian approach to inverse problems

Collocation based high dimensional model representation for stochastic partial differential equations

Quantification of airfoil geometry-induced. aerodynamic uncertainties - comparison of. approaches

Gaussian Filtering Strategies for Nonlinear Systems

Bayesian Machine Learning

Organization. I MCMC discussion. I project talks. I Lecture.

Beyond Wiener Askey Expansions: Handling Arbitrary PDFs

Stochastic Finite Elements: Computational Approaches to SPDEs

arxiv: v2 [math.na] 8 Apr 2017

Mathematisches Forschungsinstitut Oberwolfach. Numerical Methods for PDE Constrained Optimization with Uncertain Data

Overview. Bayesian assimilation of experimental data into simulation (for Goland wing flutter) Why not uncertainty quantification?

Fundamentals of Data Assimila1on

Multigrid and stochastic sparse-grids techniques for PDE control problems with random coefficients

Lecture 6: Bayesian Inference in SDE Models

Least Squares Regression

Stochastic structural dynamic analysis with random damping parameters

Sparse Quadrature Algorithms for Bayesian Inverse Problems

Transcription:

tifica NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE Alexander Litvinenko 1, Hermann G. Matthies 2, Elmar Zander 2 http://sri-uq.kaust.edu.sa/ 1 Extreme Computing Research Center, KAUST, 2 Institute of Scientific Computing, TU Braunschweig, Brunswick, Germany

KAUST Figure : KAUST campus, 7 years old, approx. 7000 people (include 1700 kids), 100 nations, 36 km 2. 2 / 31

Stochastic Numerics Group at KAUST 3 / 31

4 / 31

A BIG picture 1. Computing the full Bayesian update is very expensive (MCMC is expensive) 2. Look for a cheap surrogate (linear, quadratic, cubic,... approx.) 3. Kalman filter is a particular case 4. Do Bayesian update of Polynomial Chaos Coefficients! (not probability densities!) 5. Consider non-gaussian cases 5 / 31

History 1. O. Pajonk, B. V. Rosic, A. Litvinenko, and H. G. Matthies, A Deterministic Filter for Non-Gaussian Bayesian Estimation, Physica D: Nonlinear Phenomena, Vol. 241(7), pp. 775-788, 2012. 2. B. V. Rosic, A. Litvinenko, O. Pajonk and H. G. Matthies, Sampling Free Linear Bayesian Update of Polynomial Chaos Representations, J. of Comput. Physics, Vol. 231(17), 2012, pp 5761-5787 3. A. Litvinenko and H. G. Matthies, Inverse problems and uncertainty quantification http://arxiv.org/abs/1312.5048, 2013 4. H. G. Matthies, E. Zander, B.V. Rosic, A. Litvinenko, Parameter estimation via conditional expectation - A Bayesian Inversion, accepted to AMOS-D-16-00015, 2016. 5. H. G. Matthies, E. Zander, B.V. Rosic, A. Litvinenko, Bayesian parameter estimation via filtering and functional approximation, Tehnicki vjesnik 23, 1(2016), 1-17. 6 / 31

Setting for the identification process General idea: We observe / measure a system, whose structure we know in principle. The system behaviour depends on some quantities (parameters), which we do not know uncertainty. We model (uncertainty in) our knowledge in a Bayesian setting: as a probability distribution on the parameters. We start with what we know a priori, then perform a measurement. This gives new information, to update our knowledge (identification). Update in probabilistic setting works with conditional probabilities Bayes s theorem. Repeated measurements lead to better identification. 7 / 31

Mathematical setup Consider A(u; q) = f u = S(f ; q), where S is solution operator. Operator depends on parameters q Q, hence state u U is also function of q: Measurement operator Y with values in Y: y = Y (q; u) = Y (q, S(f ; q)). tion Logo Examples Lock-up of measurements: y(ω) = D 0 u(ω, x)dx, or u in few points 8 / 31

Inverse problem For given f, measurement y is just a function of q. This function is usually not invertible ill-posed problem, measurement y does not contain enough information. In Bayesian framework state of knowledge modelled in a probabilistic way, parameters q are uncertain, and assumed as random. Bayesian setting allows updating / sharpening of information about q when measurement is performed. The problem of updating distribution state of knowledge of q becomes well-posed. Can be applied successively, each new measurement y and forcing f may also be uncertain will provide new information. 9 / 31

Conditional probability and expectation With state u a RV, the quantity to be measured y(ω) = Y (q(ω), u(ω))) is also uncertain, a random variable. Noisy data: ŷ + ɛ(ω), where ŷ is the true value and a random error ɛ. Forecast of the measurement: z(ω) = y(ω) + ɛ(ω). Classically, Bayes s theorem gives conditional probability P(I q M z ) = P(M z I q ) P(M z ) P(I q ) (or π q (q z) = p(z q) p q (q)) Z s expectation with this posterior measure is conditional expectation. Kolmogorov starts from conditional expectation E ( M z ), from this conditional probability via P(I q M z ) = E ( ) χ Iq M z. 10 / 31

Conditional expectation The conditional expectation is defined as orthogonal projection onto the closed subspace L 2 (Ω, P, σ(z)): E(q σ(z)) := P Q q = argmin q L2 (Ω,P,σ(z)) q q 2 L 2 The subspace Q := L 2 (Ω, P, σ(z)) represents the available information. The update, also called the assimilated value q a (ω) := P Q q = E(q σ(z)), tion Logo and represents Lock-up new state of knowledge after the measurement. Doob-Dynkin: Q = {ϕ Q : ϕ = φ z, φ measurable}. 11 / 31

Polynomial Chaos Expansion (Norbert Wiener, 1938) Multivariate Hermite polynomials were used to approximate random fields/stochastic processes with Gaussian random variables. According to Cameron and Martin theorem PCE expansion converges in the L 2 sense. Let Y (x, θ), θ = (θ 1,..., θ M,...), is approximated: Y (x, θ) = β J m,p H β (θ)y β (x), J m,p = (m + p)!, m!p! H β (θ) = M k=1 h β k (θ k ), Y β (x) = 1 β! Θ N q H β (θ)y (x, θ) P(dθ). Y β (x) 1 H β (θ i )Y (x, θ i )w i. β! i=1 12 / 31

Numerical computation of NLBU Look for ϕ such that q(ξ) = ϕ(z(ξ)), z(ξ) = y(ξ) + ε(ω): ϕ ϕ = α J p ϕ α Φ α (z(ξ)) (1) and minimize q(ξ) ϕ(z(ξ)) 2, where Φ α are polynomials (e.g. Hermite, Laguerre, Chebyshev or something else). Taking derivatives with respect to ϕ α : q(ξ) ϕ(z(ξ)), q(ξ) ϕ(z(ξ)) = 0 α J p (2) ϕ α Inserting representation for ϕ, obtain 13 / 31

Numerical computation of NLBU E q 2 (ξ) 2 qϕ β Φ β (z) + ϕ β ϕ γ Φ β (z)φ γ (z) ϕ α β J β,γ J = 2E qφ α (z) + ϕ β Φ β (z)φ α (z) β J = 2 E [Φ β (z)φ α (z)] ϕ β E [qφ α (z)] = 0 β J α J E [Φ β (z)φ α (z)] ϕ β = E [qφ α (z)] 14 / 31

Numerical computation of NLBU Now, rewriting the last sum in a matrix form, obtain the linear system of equations (=: A) to compute coefficients ϕ β :.......... E [Φ α (z(ξ))φ β (z(ξ))].......... where α, β J, A is of size J J.. ϕ β. =. E [q(ξ)φ α (z(ξ))]., 15 / 31

Numerical computation of NLBU Using the same quadrature rule of order q for each element of A, we can write [ ] A = E Φ Jα (z(ξ))φ Jβ (z(ξ)) T N A i=1 w A i Φ Jα (z i )Φ Jβ (z i ) T, (3) where (w A i, ξ i ) are weights and quadrature points, z i := z(ξ i ) and Φ Jα (z i ) := (...Φ α (z(ξ i ))...) T is a vector of length J α. b = E [q(ξ)φ Jα (z(ξ))] wi b q(ξ i )Φ Jα (z i ), (4) N b i=1 where Φ Jα (z(ξ i )) := (..., Φ α (z(ξ i )),...), α J α. 16 / 31

Numerical computation of NLBU We can write the Eq. 15 with the right-hand side in Eq. 4 in the compact form: [Φ A ] [diag(...wi A...)] [Φ A ] T. w ϕ β = [Φ 0 bq(ξ 0) b]... (5) w. b q(ξ N b N b) [Φ A ] R Jα NA, [diag(...wi A...)] R NA N A, [Φ b ] R Jα Nb, [w0 bq(ξ 0)...w b q(ξ N b N b)] R Nb. Solving Eq. 5, obtain vector of coefficients (...ϕ β...) T for all β. Finally, the assimilated parameter q a will be q a = q f + ϕ(ŷ) ϕ(z), (6) z(ξ) = y(ξ) + ε(ω), ϕ = β J p ϕ β Φ β (z(ξ)) 17 / 31

Example 1: ϕ does not exist in the Hermite basis Assume z(ξ) = ξ 2 and q(ξ) = ξ 3. The normalized PCE coefficients are (1, 0, 1, 0) (ξ 2 = 1 H 0 (ξ) + 0 H 1 (ξ) + 1 H 2 (ξ) + 0 H 3 (ξ)) and (0, 3, 0, 1) (ξ 3 = 0 H 0 (ξ) + 3 H 1 (ξ) + 0 H 2 (ξ) + 1 H 3 (ξ)). For such data the mapping ϕ does not exist. The matrix A is close to singular. Support of Hermite polynomials (used for Gaussian RVs) is (, ). 18 / 31

Example 2: ϕ does exist in the Laguerre basis Assume z(ξ) = ξ 2 and q(ξ) = ξ 3. The normalized gpce coefficients are (2, 4, 2, 0) and (6, 18, 18, 6). For such data the mapping mapping ϕ of order 8 and higher produces a very accurate result. Support of Laguerre polynomials (used for Gamma RVs) is [0, ). 19 / 31

Lorenz 1963 Is a system of ODEs. Has chaotic solutions for certain parameter values and initial conditions. ẋ = σ(ω)(y x) ẏ = x(ρ(ω) z) y ż = xy β(ω)z Initial state q 0 (ω) = (x 0 (ω), y 0 (ω), z 0 (ω)) are uncertain. Solving in t 0, t 1,..., t 10, Noisy Measur. UPDATE, solving in tion Logo t 11, t 12 Lock-up,..., t 20, Noisy Measur. UPDATE,... 20 / 31

21 / 31 Trajectories of x,y and z in time. After each update (new information coming) the uncertainty drops. (O. Pajonk)

Lorenz-84 Problem Figure : Partial state trajectory with uncertainty and three updates 22 / 31

Lorenz-84 Problem 0.8 0.7 x f 0.6 0.5 x a 0.4 0.3 0.2 0.1 0 10 0 10 x 0.45 0.4 0.35 0.3 y a 0.25 0.2 0.15 0.1 0.05 0 20 0 20 y y f 0.9 1 z f 0.8 0.7 z a 0.6 0.5 0.4 0.3 0.2 0.1 0 0 10 20 z Figure : NLBU: Linear measurement (x(t), y(t), z(t)): prior and posterior after one update 23 / 31

Lorenz-84 Problem 0.9 0.8 0.7 0.6 x1 x2 0.5 0.4 0.3 0.2 0.1 0 10 5 0 x 0.5 0.45 0.4 0.35 y1 y2 0.3 0.25 0.2 0.15 0.1 0.05 0 15 10 5 y 0.6 0.5 0.4 z1 z2 0.3 0.2 0.1 0 5 10 15 z Figure : Linear measurement: Comparison posterior for LBU and NLBU after second update 24 / 31

Lorenz-84 Problem 0.16 0.14 x f 0.12 0.1 x a 0.08 0.06 0.04 0.02 0 20 0 20 x 0.09 0.08 0.07 0.06 y a 0.05 0.04 0.03 0.02 0.01 0 50 0 50 y y f 0.18 0.16 0.14 0.12 z a 0.1 0.08 0.06 0.04 0.02 0 0 10 20 z Figure : Quadratic measurement (x(t) 2, y(t) 2, z(t) 2 ): Comparison of a priori and a posterior for NLBU z f 25 / 31

Example 4: 1D elliptic PDE with uncertain coeffs Taken from Stochastic Galerkin Library (sglib), by Elmar Zander (TU Braunschweig) (κ(x, ξ) u(x, ξ)) = f (x, ξ), x [0, 1] Measurements are taken at x 1 = 0.2, and x 2 = 0.8. The means are y(x 1 ) = 10, y(x 2 ) = 5 and the variances are 0.5 and 1.5 correspondingly. 26 / 31

Example 4: updating of the solution u 30 20 10 0 10 30 20 10 0 10 20 0 0.2 0.4 0.6 0.8 1 20 0 0.2 0.4 0.6 0.8 1 Figure : Original and updated solutions, mean value plus/minus 1,2,3 standard deviations See more in sglib by Elmar Zander 27 / 31

Example 4: Updating of the parameter 2 1.5 1 0.5 0 0.5 1 0 0.2 0.4 0.6 0.8 1 2 1.5 1 0.5 0 0.5 1 0 0.2 0.4 0.6 0.8 1 Figure : Original and updated parameter q. See more in sglib by Elmar Zander 28 / 31

Conclusion about NLBU + Step 1. Introduced a way to derive MMSE ϕ (as a linear, quadratic, cubic etc approximation, i. e. compute conditional expectation of q, given measurement Y. Step 2. Apply ϕ to identify parameter q + All ingredients can be given as gpc. + we apply it to solve inverse problems (ODEs and PDEs). - Stochastic dimension grows up very fast. 29 / 31

Acknowledgment I used a Matlab toolbox for stochastic Galerkin methods (sglib) https://github.com/ezander/sglib Alexander Litvinenko and his research work was supported by the King Abdullah University of Science and Technology (KAUST), SRI-UQ and ECRC centers. 30 / 31

Literature 1. Bojana Rosic, Jan Sykora, Oliver Pajonk, Anna Kucerova and Hermann G. Matthies, Comparison of Numerical Approaches to Bayesian Updating, report on www.wire.tu-bs.de, 2014 2. A. Litvinenko and H. G. Matthies, Inverse problems and uncertainty quantification http://arxiv.org/abs/1312.5048, 2013 3. L. Giraldi, A. Litvinenko, D. Liu, H. G. Matthies, A. Nouy, To be or not to be intrusive? The solution of parametric and stochastic equations - the plain vanilla Galerkin case, http://arxiv.org/abs/1309.1617, 2013 4. O. Pajonk, B. V. Rosic, A. Litvinenko, and H. G. Matthies, A Deterministic Filter for Non-Gaussian Bayesian Estimation, Physica D: Nonlinear Phenomena, Vol. 241(7), pp. 775-788, 2012. 5. B. V. Rosic, A. Litvinenko, O. Pajonk and H. G. Matthies, Sampling Free Linear Bayesian Update of Polynomial 31 / 31

Literature 1. PCE of random coefficients and the solution of stochastic partial differential equations in the Tensor Train format, S. Dolgov, B. N. Khoromskij, A. Litvinenko, H. G. Matthies, 2015/3/11, arxiv:1503.03210 2. Efficient analysis of high dimensional data in tensor formats, M. Espig, W. Hackbusch, A. Litvinenko, H.G. Matthies, E. Zander Sparse Grids and Applications, 31-56, 40, 2013 3. Application of hierarchical matrices for computing the Karhunen-Loeve expansion, B.N. Khoromskij, A. Litvinenko, H.G. Matthies, Computing 84 (1-2), 49-67, 31, 2009 4. Efficient low-rank approximation of the stochastic Galerkin matrix in tensor formats, M. Espig, W. Hackbusch, A. Litvinenko, H.G. Matthies, P. Waehnert, Comp. & Math. with Appl. 67 (4), 818-829, 2012 5. Numerical Methods for Uncertainty and Bayesian Update in Aerodynamics, A. Litvinenko, H. G. Matthies, Book Management and Minimisation of Uncertainties and Errors in Numerical Aerodynamics pp 265-282, 2013 32 / 31