Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications

Similar documents
A Stochastic Collocation Approach to Bayesian Inference in Inverse Problems

Fast Numerical Methods for Stochastic Computations

Stochastic Spectral Approaches to Bayesian Inference

Data assimilation in high dimensions

Ergodicity in data assimilation methods

Algorithms for Uncertainty Quantification

A Spectral Approach to Linear Bayesian Updating

Fast Numerical Methods for Stochastic Computations: A Review

Data assimilation in high dimensions

Benjamin L. Pence 1, Hosam K. Fathy 2, and Jeffrey L. Stein 3

Performance Evaluation of Generalized Polynomial Chaos

Beyond Wiener Askey Expansions: Handling Arbitrary PDFs

A Stochastic Collocation based. for Data Assimilation

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Uncertainty Quantification for multiscale kinetic equations with high dimensional random inputs with sparse grids

Final Report: DE-FG02-95ER25239 Spectral Representations of Uncertainty: Algorithms and Applications

Organization. I MCMC discussion. I project talks. I Lecture.

Bayesian Inverse problem, Data assimilation and Localization

Methods of Data Assimilation and Comparisons for Lagrangian Data

Convergence of the Ensemble Kalman Filter in Hilbert Space

Uncertainty Quantification in MEMS

Adaptive ensemble Kalman filtering of nonlinear systems

Simulating with uncertainty : the rough surface scattering problem

Dynamic System Identification using HDMR-Bayesian Technique

EFFICIENT STOCHASTIC GALERKIN METHODS FOR RANDOM DIFFUSION EQUATIONS

Introduction to Uncertainty Quantification in Computational Science Handout #3

Sequential Monte Carlo Samplers for Applications in High Dimensions

A Polynomial Chaos Approach to Robust Multiobjective Optimization

On a Data Assimilation Method coupling Kalman Filtering, MCRE Concept and PGD Model Reduction for Real-Time Updating of Structural Mechanics Model

Gaussian Process Approximations of Stochastic Differential Equations

EnKF and Catastrophic filter divergence

Ensemble Kalman Filter

State and Parameter Estimation in Stochastic Dynamical Models

Parametric Problems, Stochastics, and Identification

Analysis and Computation of Hyperbolic PDEs with Random Data

Local Ensemble Transform Kalman Filter

EnKF and filter divergence

Overview. Bayesian assimilation of experimental data into simulation (for Goland wing flutter) Why not uncertainty quantification?

Uncertainty Quantification in Computational Models

What do we know about EnKF?

Research Article A Pseudospectral Approach for Kirchhoff Plate Bending Problems with Uncertainties

Multilevel stochastic collocations with dimensionality reduction

Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering

A Note on the Particle Filter with Posterior Gaussian Resampling

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit

Gaussian Process Approximations of Stochastic Differential Equations

Steps in Uncertainty Quantification

Sparse polynomial chaos expansions in engineering applications

Kalman Filter and Ensemble Kalman Filter

New Fast Kalman filter method

Fully Bayesian Deep Gaussian Processes for Uncertainty Quantification

Estimating functional uncertainty using polynomial chaos and adjoint equations

Data assimilation with and without a model

Uncertainty Quantification for multiscale kinetic equations with random inputs. Shi Jin. University of Wisconsin-Madison, USA

Adaptive Collocation with Kernel Density Estimation

Quadrature for Uncertainty Analysis Stochastic Collocation. What does quadrature have to do with uncertainty?

Covariance function estimation in Gaussian process regression

The Ensemble Kalman Filter:

Stochastic structural dynamic analysis with random damping parameters

Uncertainty Quantification and hypocoercivity based sensitivity analysis for multiscale kinetic equations with random inputs.

Schwarz Preconditioner for the Stochastic Finite Element Method

NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE

Solving the Stochastic Steady-State Diffusion Problem Using Multigrid

arxiv: v2 [math.na] 6 Aug 2018

Nonlinear stochastic Galerkin and collocation methods: application to a ferromagnetic cylinder rotating at high speed

Implementation of Sparse Wavelet-Galerkin FEM for Stochastic PDEs

Uncertainty Quantification in Computational Science

Surrogate and reduced-order modeling: a comparison of approaches for large-scale statistical inverse problems

The Ensemble Kalman Filter:

Lecture 6: Bayesian Inference in SDE Models

Superparameterization and Dynamic Stochastic Superresolution (DSS) for Filtering Sparse Geophysical Flows

STOCHASTIC SAMPLING METHODS

Gaussian Filtering Strategies for Nonlinear Systems

An Empirical Chaos Expansion Method for Uncertainty Quantification

Smoothers: Types and Benchmarks

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 1: Introduction

Solving the steady state diffusion equation with uncertainty Final Presentation

Bayesian parameter estimation in predictive engineering

A new Hierarchical Bayes approach to ensemble-variational data assimilation

Data assimilation with and without a model

Nonlinear State Estimation! Particle, Sigma-Points Filters!

LARGE-SCALE TRAFFIC STATE ESTIMATION

c 2004 Society for Industrial and Applied Mathematics

Fundamentals of Data Assimila1on

Bayesian Statistics and Data Assimilation. Jonathan Stroud. Department of Statistics The George Washington University

Polynomial chaos expansions for sensitivity analysis

DATA ASSIMILATION FOR FLOOD FORECASTING

EnKF-based particle filters

Lagrangian Data Assimilation and Its Application to Geophysical Fluid Flows

Spectral Representation of Random Processes

AN EFFICIENT COMPUTATIONAL FRAMEWORK FOR UNCERTAINTY QUANTIFICATION IN MULTISCALE SYSTEMS

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Sparse Quadrature Algorithms for Bayesian Inverse Problems

Ensemble Data Assimilation and Uncertainty Quantification

Estimation of State Noise for the Ensemble Kalman filter algorithm for 2D shallow water equations.

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

Filtering Sparse Regular Observed Linear and Nonlinear Turbulent System

Numerical Solutions of ODEs by Gaussian (Kalman) Filtering

NEW ALGORITHMS FOR UNCERTAINTY QUANTIFICATION AND NONLINEAR ESTIMATION OF STOCHASTIC DYNAMICAL SYSTEMS. A Dissertation PARIKSHIT DUTTA

Robust exponential convergence of hp-fem for singularly perturbed systems of reaction-diffusion equations

Transcription:

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications Dongbin Xiu Department of Mathematics, Purdue University Support: AFOSR FA955-8-1-353 (Computational Math) SF CAREER DMS-64535 (Computational Math) DOE DE-FC52-8A28617 (PSAAP)

Overview Generalized polynomial chaos Stochastic collocation Lagrange interpolation Pseudo spectral gpc Applications Bayesian inverse problem Data assimilation

Stochastic PDE: Uncertainty Quantification via gpc u nz (, txz, ) = L( u ), (, T ] D R t B( u) =, [, T] D R u = u( x, Z), { t = } D R th -order gpc expansion: nz + u (,, ) ˆ t x Z uk(, t x) Φ k( Z), # of basis= n k = uˆ = E[ u( Z) Φ ( Z)] = u( Z) Φ ( Z ) ρ( Z ) dz,, k k k k n n z z z Orthogonal basis: E Z ( Z) ( Z) Φ Φ Φ ( Z) Φ ( Z) ρ( Z) dz = δ i j i j ij Optimality: u u = inf nz u Ψ L 2 2 ρ Z Ψ Π L ρ Z ( ) ( )

Generalized Polynomial Chaos (gpc) Basis functions: Hermite polynomials: seminal work by R. Ghanem Global orthogonal polynomials (Xiu & Karniadakis, 2) Wavelet basis (Le Maitre et al, 4) Piecewise basis (Babuska et al 4, Wan & Karniadakis, 5) Implementations: Stochastic Galerkin Stochastic collocation Properties: Rigorous mathematics High accuracy, fast covergence Curse-of-dimensionality

Stochastic Collocation Collocation: To satisfy governing equations at nodes Sampling: (solution statistics only) Random (Monte Carlo) Deterministic (lattice rule, tensor grid, cubature) Stochastic collocation: To construct polynomial approximations Lagrange interpolation Can not be constructed for any given nodes Interpolation error hard to control Pseudo spectral Utilize gpc polynomial basis Becomes multivariate integration Response surface method Multivariate interpolation Many ad hoc approaches

Stochastic Collocation Lagrange Interpolation odal set: Lagrange interpolation: Θ = Q i Q nz { Z } R 1 i= Q Q j j u ( Z) u( Z ) Lj( Z) Li( Z ) = δij, 1 i, j Q j= 1 Solution: for j=1,,q, u (, txz j, ) = L( u ), in (, T ] D, t B( u) =, [, T] D, j u = u ( x, Z ), { t = } D Tensor product: i1 n ( U U i z ) Sparse grid (Smolyak): q + 1 i q 1 q i q i i i 1 ( 1) ( U U nz ) (Xiu & Hesthaven, SIAM J. Sci. Comput., 5)

Stochastic Collocation: Pseudo Spectral Approach th -order gpc projection u (,, ) ˆ t x Z = uk(, t x) Φk( Z), uˆ = u( Z) Φ ( Z) ρ ( Z) dz. k k gpc-collocation approximation w (, t x, Z) = wˆ (, t x) Φ ( Z), k = k j j j k k k j= 1 k Q wˆ = u( t, x, Z ) Φ ( Z ) α u( Z) Φ ( Z) ρ( Z) dz k = wˆ (, t x) uˆ (, t x), Q k k Aliasing Error: ε u w Q L 2 ( Z ) ρ (Xiu, Comm. Comput Phys, vol. 2, 7)

gpc-collocation: Algorithm j j Q { Z α } j = 1 nz 1. Choose a nodal set, in R Deterministic solver 2. Solve for each j = 1,, Q, u (, txz j, ) = L( u ), in (, T ] D, t B( u) =, [, T] D, j u = u ( x, Z ), { t = } D 3. Evaluate the approximate gpc expansion coefficient Q j k k j j α k j= 1 wˆ = u( t, x, Z ) Φ ( Z ), ; th 4. Construct the -order gpc approximation w ( t, x, Z) = wˆ Φ ( Z). k = 1 k k Post-process Error bound (Xiu, 7): 2 2 2 2 ( ) ε u w ε + ε + M ε C Δ 2 L ( Z) Q Q ρ 1/2 Error Finite-term projection error + aliasing error + umerical error

Parameter Estimation: Bayesian Inverse Approach Stochastic PDE: u nz (, txz, ) = L( u ), (, T ] D R t B( u) =, [, T] D R u = u( x, Z), { t = } D R n n z z Solution: u utxz (,, ):[, T] D R n n z R Prior distribution: n z π ( z) = π ( z ) Z i i i= 1 Estimation of the prior distribution Requires direct measurements of the parameters o/not enough direct measurements? (Use experience/intuition ) How to take advantage of measurements of other variables?

Bayesian Inference Data: n d = G( Z) + e, e R d is i.i.d. Posterior distribution: d π ( d Z) π ( Z) π ( Z) π( Z d) = π( d Z) π( Z) dz Likelihood function: d LZ ( ) π( d Z) = π ( d G( Z)) n i= 1 e i i otes: Difficult to manipulate Classical sampling approaches can be time consuming (MCMC, etc) GPC (Galerkin) based approach: (Marzouk, ajm, Rahn, JCP, 7) gpc approximation: Properties: d L ( Z) π ( Z) π ( Z) = L ( Z) π ( Z) dz d L ( Z) π ( d Z) = π ( d G ( Z)) ei i, i i= 1 Allows direct sampling in term of Z with arbitrarily large samples (Virtually) no additional computational cost forward problem solver only Convergence seems natural n i

Convergence of gpc Bayesian Inference Kullback-Leibler divergence: π1( z) D( π1 π2) π1( z)log dz π ( z) 2 Observation error: 2 e (, σ I), i.i.d. ormal Theorem π d 2. If the gpc expansion G converges to G in Lπ, then the posterior density d converges to π in the sense Moreover, if G( Z) G ( Z) i i, 2 π z then for sufficiently large, L d d ( ) D π π,. D α C, 1 i n, α >, C independent of, d d 2 ( π π ) otes: Fast (exponential) convergence rate is retained Factor of 2 in the convergence rates α. d Z (Marzouk & Xiu, Comm. Comput. Phys. 8)

Parameter Estimation: Supersensitivity Example Burgers equation : Boundary conditions : 2 u u u + u = ν, x 1,1 2 t x x [ ] u( 1) = 1 + δ( Z); u(1) = 1; < δ<< 1 Deterministic results with no uncertainty (Xiu & Karniadakis, Int. J. umer. Eng., 4) 1% uncertainty in left BC

25 exact posterior gpc, p=4 gpc, p=8 δ true 2 p(δ data ) 15 1 5.1.2.3.4.5.6.7.8.9.1 Prior distribution is uniform Measurement noise: e~(,.5 2 ) δ

1 2 error 1 3 D(π π ) G G L 2 6 7 8 9 1 11 12 13 14 15 16 Factor = 2.1 (theory = 2)

Parameter Estimation: Step Function Assume the forward model is a step function Posterior distribution is discontinuous Gibb s oscillations exist Slow convergence with global gpc basis functions 1.2 1 3 2.5 exact posterior gpc posterior.8 2 G(z) or G (z).6.4 π d (z) 1.5 1.2.5 exact forward solution gpc approximation.2 1.8.6.4.2.2.4.6.8 1 z Forward model and its approximation 1.8.6.4.2.2.4.6.8 1 z Posterior distribution and its approximation

D(π π ) 1.1 G G L 2 1.2 1.3 1.4 error 1.5 1.6 1.7 1.8 1.9 2 4 6 8 1 12 14 16 18 2 p Factor = 1.99 (theory = 2)

Kalman Filter for Data Assimilation t m True state (unknown): u R, m 1 Forecast: Observation: Analysis: f du f (, tz) = F(, tu ), t (, T] dt f u (, Z) = u t m d = Hu + ε R, H : R R a f f u = u + K( d Hu ) K = P H ( HP H + R) f T f T 1 (Kalman gain matrix) P = E ( u u )( u u ) f f t f t T R E εε T = Properties: Straightforward for linear dynamic equations Extension to nonlinear equations: Extended KF (EKF) Optimal for Gaussian Explicit calculation of covariance can be costly

Ensemble Kalman Filter (EnKF) Ensemble: f f i ( u ) u (, ), = 1,, i tz i M ( d) = d+ (), ε i i = 1,, i a f f ( ) ( ) e i ( ) u = u + K ( ), i = 1,, M i i d H u i f T f T K = P H ( HP Η + R ) e e e e 1 f f P f = f f T f e ( u u )( u u ) P R εε T R e = Properties: nonlinear dynamics sampling errors Measurement. Can be eliminated by square-root filter (EnSRF) Solution states. Computational cost is of great concern

Error Analysis of the EnKF Assimilation step size: Δ T = t t n+ 1 n Lemma (local error): e M ε + ΔK ε + ΔK H ε f n+ 1 Δt Δt p ( Δ, σ ) O t α M = I KH Δ K = Ke K Theorem (global error): n E E + e exp Λ t ( ) n k n k= 1 Λ ΔT 1 ote the inverse dependence on assimilation step size (Li & Xiu, vol. 197, CMAME 8)

EnKF Example: Linear Wave Equation Model description: Linear advection equation; Periodic domain of length L=1; Wave speed = 1; grid spacing=1; time step = 1; True states are sampled from a Gaussian process, with zero mean and unit variance, and a spatial decorrelation length of 2. The dimension of random space is n z =5. Four measurements uniformly in space are made every 5 time units. Measurement variance is.1. o model error. 1 5 1 2 3 4 5 6 7 8 9 1 1 5 1 2 3 4 5 6 7 8 9 1 8 6 4 2 ( xz, ) R t=5 t=1 t=5 1 2 3 4 5 6 7 8 9 1 R 5 Long-term Wave propagation

Error Behavior of EnKF 5 5.5 T=1 T=5.15 T=1 T=5 6.1 log (Error) 6.5 7 Error.5 7.5 8 8.5 4.5 5 5.5 6 6.5 7 7.5 8 log () w.r.t. ensemble size.5.1.15.2.25.3.35.4.45.5 Standard Deviation of Measurements w.r.t. data noise level

EnKF Error Behavior 6.5 7 7.5 EnKF EnSRF qensrf 2 qensrf 3 8 log(error) 8.5 9 9.5 1 1.5 11 11.5.5 1 1.5 2 2.5 3 3.5 4 log(δ T) qensrf: EnSRF combined with deterministic sampling using optimal cubature

GPC Collocation based Kalman Filter Errors of assimilated results T=1 T=5 T=1, T=1,5 EnKF (=1) 5.3 1-3 3.4 1-3 2.3 1-3 1.7 1-3 EnKF (=1 3 ) 1.5 1-3 1. 1-3 7.4 1-4 6.2 1-4 EnKF (=1 4 ) 4.6 1-4 2.8 1-4 2.4 1-4 1.9 1-4 gpc-kf (=51) 3.5 1-4 1.6 1-4 1.1 1-4 9. 1-5 gpc-kf (=1) 1.8 1-4 7.9 1-5 5.6 1-5 4.6 1-5 5 dimensional random space (Li & Xiu, vol. 197, CMAME 8)

Accuracy Improvement of EnKF via gpc Use cubature equally weighted Use pseudo-spectral gpc Analytical expression in Z f f u (, tz ) ˆ u () t ( Z k Φk ), k = Statistics ( ) u = uˆ, P = uˆ uˆ f f f f k k < k T (Li & Xiu, J. Comput. Phys. 8)

GPC Based Ensemble Kalman Filter Ensemble: f f i ( ) k k ( ) u = u ˆ () t Φ Z, i = 1,, M, M 1 i k = Square-root update: f f f ' a a a ( ) ( ) ( ) ( ) u = u + u, u = u + u, i = 1,, M i i i i Mean state update: u a f ( f = u + K d Hu), Perturbation update: a f f ( u) ( u) K H( u) ' ' ' = +, i = 1,, M i i i ' f T f T K = P H ( HP Η + R) 1 T 1 f T f T f T = + + + ( ) ( ) K PH HPH R HPH R R 1

Example: onlinear Population Dynamics du dt f f u f f = r 1 u, u () = u A 2.8 2.7 2.6 True state Measurement Estimate 2.5 2.4 2.3 2.2 2.1 2.2.4.6.8 1 Time

Error Convergence 3 4 5 Mean Standard deviation 2 Mean Standard deviation log(error) 6 7 8 log(error) 4 6 9 8 1 1 11 1 2 3 4 5 6 7 8 9 Degree of the gpc expansion () 12 4 5 6 7 8 9 1 11 12 13 umber of quardrature points (Q) =8, Q=1 is sufficient

Comparison: gpc-kf vs EnKF 2 Mean (gpc EnSRF) Std (gpc EnSRF) Mean (EnSRF) Std (EnSRF) 4 log(error) 6 8 1 12 1 2 3 4 5 6 log(ensemble size)

onlinear System Example: Lorenz Equations dx dt dy dt dz dt = σ ( y x) = ρ x y xz = xy β z σ = 1, ρ = 28, β = 8/ 3 ( x, y, z ) = (1.5887, 1.531271, 25.4691) 2 X 2 2 4 6 8 1 12 14 16 18 2 Time 5 Y 5 2 4 6 8 1 12 14 16 18 2 Time 6 4 Z 2 2 4 6 8 1 12 14 16 18 2 Time Small deviation in initial condition (.1 in x ) causes large deviation

Qualitative Comparison: gpc EnKF vs EnKF 3 2.5 =2 (gpc EnSRF) =3 (gpc EnSRF) EnSRF Estimate True state 2 1.5 1.5 5 1 15 2 Time GPC EnKF: Q=5 3 =125 EnKF: ensemble size = 14

Summary Point selection is crucial for the efficacy of stochastic collocation GPC expansion is much more than a forward UQ method Bayesian inverse (Marzouk & Xiu, Comm. Comput. Phys, 8) Kalman filter for data assimilation (Li & Xiu, CMAME 8; JCP 8)