A Data-Driven Stochastic Method

Similar documents
An Empirical Chaos Expansion Method for Uncertainty Quantification

Hierarchical Parallel Solution of Stochastic Systems

A Data-Driven Stochastic Method for Elliptic PDEs with Random Coefficients

Fast Numerical Methods for Stochastic Computations

Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques

Polynomial Chaos and Karhunen-Loeve Expansion

Uncertainty Quantification in Computational Science

UNCERTAINTY ASSESSMENT USING STOCHASTIC REDUCED BASIS METHOD FOR FLOW IN POROUS MEDIA

Lecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:

Stochastic Spectral Approaches to Bayesian Inference

A Stochastic Collocation based. for Data Assimilation

Solving the steady state diffusion equation with uncertainty Final Presentation

Implementation of Sparse Wavelet-Galerkin FEM for Stochastic PDEs

Adaptive Collocation with Kernel Density Estimation

Multilevel accelerated quadrature for elliptic PDEs with random diffusion. Helmut Harbrecht Mathematisches Institut Universität Basel Switzerland

Estimating functional uncertainty using polynomial chaos and adjoint equations

Collocation based high dimensional model representation for stochastic partial differential equations

Efficient Solvers for Stochastic Finite Element Saddle Point Problems

Analysis IV : Assignment 3 Solutions John Toth, Winter ,...). In particular for every fixed m N the sequence (u (n)

EXPLORING THE LOCALLY LOW DIMENSIONAL STRUCTURE IN SOLVING RANDOM ELLIPTIC PDES

A Dynamically Bi-Orthogonal Method for Time-Dependent Stochastic Partial Differential Equations I: Derivation and Algorithms

Sparse polynomial chaos expansions in engineering applications

STOCHASTIC SAMPLING METHODS

Bath Institute For Complex Systems

Real Variables # 10 : Hilbert Spaces II

Monte Carlo Methods for Uncertainty Quantification

Uncertainty Quantification for multiscale kinetic equations with high dimensional random inputs with sparse grids

NONLOCALITY AND STOCHASTICITY TWO EMERGENT DIRECTIONS FOR APPLIED MATHEMATICS. Max Gunzburger

arxiv: v1 [math.na] 14 Sep 2017

Schwarz Preconditioner for the Stochastic Finite Element Method

Simple Examples on Rectangular Domains

Performance Evaluation of Generalized Polynomial Chaos

Statistical Data Analysis

Uncertainty Quantification and hypocoercivity based sensitivity analysis for multiscale kinetic equations with random inputs.

A Reduced Basis Approach for Variational Problems with Stochastic Parameters: Application to Heat Conduction with Variable Robin Coefficient

Multi-Element Probabilistic Collocation Method in High Dimensions

Uncertainty Quantification for multiscale kinetic equations with random inputs. Shi Jin. University of Wisconsin-Madison, USA

Separation of Variables in Linear PDE: One-Dimensional Problems

arxiv: v2 [math.pr] 27 Oct 2015

Introduction to Uncertainty Quantification in Computational Science Handout #3

MULTISCALE FINITE ELEMENT METHODS FOR STOCHASTIC POROUS MEDIA FLOW EQUATIONS AND APPLICATION TO UNCERTAINTY QUANTIFICATION

Beyond Wiener Askey Expansions: Handling Arbitrary PDFs

Spectral Polynomial Chaos Solutions of the Stochastic Advection Equation

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications

arxiv: v5 [math.na] 16 Nov 2017

2tdt 1 y = t2 + C y = which implies C = 1 and the solution is y = 1

Experiences with Model Reduction and Interpolation

Quasi-optimal and adaptive sparse grids with control variates for PDEs with random diffusion coefficient

STA 294: Stochastic Processes & Bayesian Nonparametrics

Downloaded 08/15/18 to Redistribution subject to SIAM license or copyright; see

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.

PARALLEL COMPUTATION OF 3D WAVE PROPAGATION BY SPECTRAL STOCHASTIC FINITE ELEMENT METHOD

conditional cdf, conditional pdf, total probability theorem?

Multigrid and stochastic sparse-grids techniques for PDE control problems with random coefficients

arxiv: v1 [math.na] 3 Apr 2019

A Polynomial Chaos Approach to Robust Multiobjective Optimization

Notes Regarding the Karhunen Loève Expansion

REAL ANALYSIS II HOMEWORK 3. Conway, Page 49

Numerical Approximation of Stochastic Elliptic Partial Differential Equations

Stochastic Solvers for the Euler Equations

Downloaded 06/11/18 to Redistribution subject to SIAM license or copyright; see

Random matrices: Distribution of the least singular value (via Property Testing)

CSCI-6971 Lecture Notes: Monte Carlo integration

Probabilistic Collocation Method for Uncertainty Analysis of Soil Infiltration in Flood Modelling

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit

A collocation method for solving some integral equations in distributions

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

Chapter 2 Spectral Expansions

Polynomial chaos expansions for sensitivity analysis

An Adaptive ANOVA-Based Data-Driven Stochastic Method for Elliptic PDE with Random Coefficients

Model Reduction, Centering, and the Karhunen-Loeve Expansion

A Vector-Space Approach for Stochastic Finite Element Analysis

The circular law. Lewis Memorial Lecture / DIMACS minicourse March 19, Terence Tao (UCLA)

PRECONDITIONING MARKOV CHAIN MONTE CARLO SIMULATIONS USING COARSE-SCALE MODELS

Solution of Stochastic Nonlinear PDEs Using Wiener-Hermite Expansion of High Orders

[2] (a) Develop and describe the piecewise linear Galerkin finite element approximation of,

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples

EFFICIENT STOCHASTIC GALERKIN METHODS FOR RANDOM DIFFUSION EQUATIONS

Galerkin Methods for Linear and Nonlinear Elliptic Stochastic Partial Differential Equations

Multilevel stochastic collocations with dimensionality reduction

Low-rank techniques applied to moment equations for the stochastic Darcy problem with lognormal permeability

Densities for the Navier Stokes equations with noise

Solving the Stochastic Steady-State Diffusion Problem Using Multigrid

Chapter 7: Bounded Operators in Hilbert Spaces

An acoustic wave equation for orthorhombic anisotropy

Kernel Method: Data Analysis with Positive Definite Kernels

LECTURE # 0 BASIC NOTATIONS AND CONCEPTS IN THE THEORY OF PARTIAL DIFFERENTIAL EQUATIONS (PDES)

Solving Boundary Value Problems (with Gaussians)

UNIVERSITY OF CALIFORNIA, SAN DIEGO. An Empirical Chaos Expansion Method for Uncertainty Quantification

Statistical signal processing

An Adaptive Multi-Element Generalized Polynomial Chaos Method for Stochastic Differential Equations

Final Exam May 4, 2016

Second Order Elliptic PDE

Long Time Propagation of Stochasticity by Dynamical Polynomial Chaos Expansions. Hasan Cagan Ozen

Lecture 7 Introduction to Statistical Decision Theory

Lecture 12: Detailed balance and Eigenfunction methods

Multilevel Monte Carlo for uncertainty quantification in structural engineering

Analysis and finite element approximations of stochastic optimal control problems constrained by stochastic elliptic partial differential equations

Gaussian processes for inference in stochastic differential equations

Transcription:

A ata-riven Stochastic Method Mulin Cheng, Thomas Y. Hou, Pengchong Yan Applied and Computational Mathematics, California Institute of Technology, Pasadena, CA 96 Abstract Stochastic partial differential equations (SPE) arise in many physical and engineering applications. Solving these SPE efficiently and accurately presents a great challenge. The Wiener Chaos Expansion method (WCE) and its variants have shown some promising features but still suffer from the slow convergence of polynomial expansion and the curse of dimensionality. The Monte Carlo method (MC) has an attractive property that its convergence is independent of the number of stochastic dimensions. However, its convergence rate is only of order O(/ N) (N is the number of realizations). Inspired by Multi-scale Finite Element methods (MsFEM) and Proper Orthogonal ecomposition methods, we propose a data-driven stochastic method, which consists of offline step and online step and uses both MC and WCE. A random basis {A i(ω)} M i= is computed in the offline step based on the Karhunen-Loeve(KL) expansion, which can be viewed as partially inversion of the random differential operator. In the online step, the polynomial basis used in Polynomial Chaos expansion is replaced by the problem-dependent stochastic basis that we constructed in the offline step. By solving a set of coupled deterministic PEs of coefficients, we obtain the numerical solutions of SPEs. The proposed framework is applied to and elliptic problems with random coefficients and a horn problem. Numerical examples show that. The stochastic basis obtained in the offline computation can be repeatedly used in the online computation for different deterministic right hand sides.. A much smaller number of bases are necessary in the online computation to achieve certain error tolerance compared to WCE. Introduction Uncertainty arises in many complex real-world problems of physical and engineering interests, such as wave, heat and pollution propagation through random media, randomly forced Burgers or Navier-Stokes equations. Additional examples can be found extensively in other branches of science and engineering, such as fluid dynamics, geophysics, quantum mechanics, statistical physics, biology and etc. Stochastic Partial differential equation(spe) is an effective mathematical tool to model these complicated problems, which manifests randomness in many aspects, such as initial/boundary conditions, model parameters, measurements, to name a few. Solving these SPEs poses a great challenge to the scientific computing community. In many applications, one is mostlt interested in statistical information of the stochastic solution. For this purpose, people have derived moment equations and probability density function (PF) methods in which effective equations such as Fokker-Planck equations, have been derived for the moments and the PF. For nonlinear problems, such methods suffer from the closure problem. Because of stability and non-intrusive nature of the Monte Carlo method(mc), it remains the most popular numerical method for SPEs. ue to the law of large number, the convergence rate of MC is merely N [?], where N is the number of realizations. There are various ways to accelerate the convergence rate of MC by using variation reduction techniques, such as anti-thetic variables, conditioning, control variates, and importance sampling. For SPEs, one can also use collocation and sparse grid methods to reduce the complexity of the method, but none of these techniques can change the nature of slow convergence of MC. While MC method represents the solution of SPE as an ensemble of realizations, another way to represent the solution is to expand it in an orthonormal basis, or polynomial chaos which was originally proposed by N. Wiener in 938[?]. The method did not receive much attention until Cameron and Martin developed an explicit

and intuitive formulation for Wiener Chaos Expansion(WCE) in their elegant work in 947. Compared to MC methods, dramatic improvements can be achieved in some problems, but WCE method has two main disadvantages. The number of terms in truncated WCE grows exponentially fast as the order of Hermite polynomials and the number of random variables increase, especially for long time integration.. The solution may depend sensitively on random dimensions, such as parameter shocks or bifurcation, which further demands more terms in WCE. To reduce dimensionality and achieve sparse representation, several generalizations of the classical WCE method have been proposed to circumvent the drawbacks, such as generalized polynomial chaos via Wiener- Askey polynomial [?], Wiener-Haar Chaos Expansion [?], Multi-element gpc [?,?,?] and etc. Improvements are observed in some problems, but not all, especially for time-dependent problems. It is important to note that the basis used in the polynomial chaos expansion is determined a prior and problem independent, which we believe is the ultimate reason of giving rise to a non-sparse representation of the stochastic solution when the solution is not smooth in the stochastic space or when the stochastic dimension is high. A typical scenario occurring frequently in practice is that we are interested in investigating what will happen when deterministic parts of the problem change and the random parts remain unchanged. For example, the properties of certain kind of composite material satisfy some stochastic processes with known parameters and we want to know its responses under different external forces. Mathematically, consider the SPE in the general form of L(ω)u(x, ω) = f(x) where x, ω Ω and L(ω) is a stochastic differential operator. Clearly, L(ω) represents the random parts of the problem while f(x) is the deterministic part of the problem. Our framework can be extended to time-dependent SPE which we should address in a separate paper. The method we proposed consists two parts: Offline: The stochastic basis {A i (ω)} is constructed based on the Karhunen-Loeve(KL) expansion of MCFEM solutions u(x, ω) for a specific right-hand side f(x) = f (x). Online: For the right-hand side f(x) of interests, the solution is projected on the constructed basis, u(x, ω) M u i (x)a i (ω) i= where A =. By Galerkin projection, a system of coupled deterministic PEs of u i (x, t) can obtained and solved numerically. The mean, variance and other statistical quantities can be recovered from the solution of u i (x, t). This paper is organized as follows. In the next two section, Offline and online algorithms will be described and analyzed in details along with Multi-Level Monte Carlo Method to accelerate offline computation. In sec.5, we applied our method to elliptic partial differential equations with random elliptic coefficients and compared the results with those by PC method to show the advantages. Some cases of physical and engineering interests, such as large variance or jumps at random locations, are also included and analyzed numerically in details. To demonstrate the potential application of our method in industrial/realistic scenario, a horn problem is considered. Some concluding remarks are given in Section 6.

Theory. Stochastic Partial ifferential Equation Consider the stochastic PE, linear or non-linear, in the general form of or L(x, ω)u(x, ω) = f (x), x, ω Ω u(x, ω) =, x, (.) L(x, ω)u(x, ω) =, x, ω Ω u(x, ω) = f (x), x Γ, (.) where is a d-dimensional physical space. Here the deterministic function f (x) exists either at the righthand-side of the PE, or as part of the boundary condition. The Monte-Carlo solution of the elliptic equation, given one sample ω, u(x, ω) is solved. The Karhunen- Loève decomposition of the PE solution gives the dominant components in the random spaces. We denote the KL decomposition as u(x, ω; f ) = ū(x; f ) + λi A i (ω)φ i (x; f ), (.3) i= where ū(x; f ) = Eu(x, ω; f ). In the current paper, we intend to derive the empirical basis {A i (ω)} i=,...,m based on a test function f (x), where M is a small truncation number. Notice that the basis {A i (ω)} i=,...,m in the KL decomposition are orthonormal basis. We don t need to store all information of A i (ω), but some moments, such as E[a(x, ω)a i (ω)a j (ω)]. Those computations can be done off-line. Then for a different function f (x), we are looking for a Galerkin projection of the solution, i.e., u(x, ω; f ) = ū(x; f ) + M A i (ω)u i (x; f ). (.4) For easy notation, we denote A (ω) = and u (x; f ) = ū(x; f ). It is important to point that if the correlation length of the solution is small, then the number of basis M may be large due to the large dimension of the stochastic solution space. Then our method is not an optimal choice to solve the problem, although it still has its big advantages over other method, such as WCE. In this paper, we will exclude this situation and consider only the case with large correlation length of the solution. In addition to that, there exists some puzzles here we need to solve here. As for the Wiener Chaos expansion, the dimension of the stochastic domain is infinite, while we only choose a PE-driven stochastic domain, which is very small (if the variance of the solution is not large). In order to validate the Galerkin method, we will i=. verify the dominant solution space is finite dimensional;. define a class of recoverable function f (x) on the basis generated off-line by f (x).. Solution space of Elliptic equation with random coefficients Let s first consider the stochastic Elliptic equation of the form div(a(x, ω) u) = f (x), x, ω Ω u(x, ω) =, x (.5) 3

where = [, ] d, d is the space dimension, and a(x, ω) is a random variable satisfying the ellipticity, < a a(x, ξ) a <. In particular, we consider the simple -d case with homogeneous irichlet boundary condition: (a(x, ω) u(x, ω)) = f(x). x x Then we derive a(x, ω) u(x, ω) = x x f(s )ds + C(ω), where C(ω) is determined by the boundary conditions. In addition, we denote b(x, ω) = (a(x, ω)). Then the solution is given by u(x, ω) = x Because of the boundary condition u(, ω) =, we have C(ω) = b(s, ω) ( s f(s )ds C(ω) ) ds. b(s, ω) s f(s )ds ds b(s, ω)ds. (.6) Also we can write the Karhunen-Loève decomposition of b(x, ω) as b(x, ω) = M b i (x)b i (ω), in which we only keep the first M leading terms. Then the solution is u(x, ω) = x i= b i (s )B i (ω) i i s f(s )ds + ( x b i (s)b i (ω)ds ) j b j(s )B j (ω) s f(s )ds ds b(s, ω)ds. Apparently, the stochastic part of first term lies in the space spanned by {B i (ω)} i=,...,m, while that of the second one lies in the space spanned by { Bi (ω)b j (ω) } b(s, ω)ds i,j with i, j =,,..., M. Again we denote B =. Then the solution of the -d Stochastic elliptic equation lies in the same stochastic space, which is independent of the right-hand-side function f(x)..3 Solution space of and higher dimensional elliptic equation with random cofficients In high-dimension case, we consider the Green s function G(x, y; ω), which satisfies div(a(x, ω) G(x, y; ω)) = δ(x y), x, y, ω Ω Again we can write out the Karhunen-Loève decomposition of G(x, y; ω) G(x, y; ω) =, x, y. (.7) G(x, y; ω) = M G i (x, y)ξ i (ω), i= 4

in which we keep the leading M terms. Then solution of the elliptic equation is u(x, ω) = G(x, y; ω)f(y)dy = M ( i= ) G i (x, y)f(y)dy ξ i (ω). So again the solution lies in a finite dimensional stochastic space spanned by {ξ i (ω)} i=,...,m, regardless of the right-hand-side f(x). Although this part is derived for Elliptic equation, the results can be generalized to a large class of stochastic PEs (e.g., those PE with a Green s function, see Helmholtz equation in Sec.5.3). 3 ata-riven Stochastic Multi-scale Method The central task of our method is to look for a ata-riven or Problem-ependent random basis under which solutions enjoy a sparse expansion. The Error-Minimizing property of Karhunen-Loeve (KL) expansion makes it a natural choices. Specifically, for u(x, ω) L ( Ω), its Karhunen-Loeve expansion is u(x, ω) = E [u] + λi ξ i (ω)φ i (x) where λ i and φ i (x) are the eigenvalue and the eigenfunction of the covariance kernel C(x, y), i.e., C(x, y)φ(y)dy = λφ(x) i= where C(x, y) = E [(u(x, ω) E [u(x, )])(u(y, ω) E [u(y, )])]. Under some general conditions, the eigenvalues in KL expansion decay exponentially fast [?], which means the first few terms, or modes, of the KL expansion can give very good approximation of u(x, ω). In other words, the dimension of input random space may be large, but the dimension of the effective output random space may be much smaller as shown in fig.. For elliptic PE, a heuristic argument based on Green s function shows the properties of solutions are mainly controlled by the properties of stochastic differential operator L(ω). Therefore, the solutions u(x, ω) for a board class of right-hand side f(x) may enjoy sparse representation under some random basis which certainly depends on the stochastic operator L(ω). In other words, a random basis computed for a specific deterministic right-hand side f(x) = f (x) may provide sparse representations for a board class of deterministic righthand side f(x). Similar philosophy has been adopted in Multi-scale Finite Element methods [?] and Proper Orthogonal ecomposition (PO) methods [?,?,?], so we name our method as ata-riven Stochastic Multiscale method. 3. Offline and Online Parts If we follow this philosophy, the framework we proposed naturally breaks into two parts, the construction of a random basis {A i (ω)} M i= and the numerical solution of a coupled system of PEs of the expansion coefficients. For a fixed stochastic differential operator L(ω), the computation of the random basis may be expansive depending on the problem and the accuracy requirements. Once the random basis are constructed, they can be used repeated for different right-hand side function f(x).r Therefore, we call the construction of random basis the offline computation part, meaning this step is a preprocess step and can be completed in an offline manner contrast to the numerical solution of a coupled system of PEs in an online manner. The following two algorithms reveal the details of each parts. The goal of offline computation is to construct the random basis based on KL expansion of the random solution. The algorithm formulated below is essentially two-pass Monte Carlo method. In the first pass, the 5

β = 3. 5 M = 5 M = M = 5 M = β =. 4 6 M = 5 M = M = 5 M = 8 5 5 5 4 5 5 4 β =. 6 M = 5 M = M = 5 M = 4 β =. 6 M = 5 M = M = 5 M = 8 8 5 5 5 5 Figure : Eigenvalues of the covariance function of solutions for elliptic PE with random coefficients a(x, ω) = M m= m β ξ m (sin(πmx + η i ) + ), where ξ i s, η i s are independent uniform random variables on [, ]. We observe that the dimensionality of input random space may be large, but effective output random space may be much smaller due the exponential convergence of eigenvalues of covariance function. mean and the covariance function of the solution are computed. In the second pass, the random basis and their relevant statistical information are computed. Various Monte Carlo acceleration techniques can be integrated into this step and reduce the offline computational cost, which is relevant, but not the central focus of this paper. Algorithm (Offline Computation).. Compute the expectation ū = E [u] and the covariance function R(x, x ) = E [(u(x, ω) ū(x))(u(x, ω) ū(x ))] by Monte Carlo Method with N realizations, i.e., ū(x) = N u(x, ω n ) N n= R(x, x ) = N u(x, ω n )u(x, ω n ) ū(x)ū(x ). Solve the first M pairs of eigenvalues and eigenfunctions λ i φ i (x) = R(x, x )φ i (x ) dx, i =,..., M 6

3. Compute the stochastic basis A i (ω) by A i (ω) = λi and statistical information of A i (ω), such as u(x, ω)φ i (x) dx E [A i A j A k ] = N A i (ω n )A j (ω n )A k (ω n ) N n= Remark 3.. Guard-lines of selections for computational parameters. The decaying property of eigenvalues can be used to select parameter M, i.e., the number of random basis M can be chosen such as λ M+ λ is small. In general, N can be chosen to be smaller than N. Remark 3.. In the online part, Galerkin projection reduces the stochastic PE to a coupled system of PEs whose coefficients involve only of statistical information of A i (ω) s, such as E [A i A j A k ], which suggests only such information be stored in the offline part. Remark 3.3. Our method is semi-non-intrusive in the sense that certified legacy codes can be used with minimum changes in offline part. This property makes it very attractive to industrial community as it reuses codes verified by possibly expansive and time-consuming experiments and avoid another life cycle of developing new programs. Remark 3.4. We can apply multi-level techniques to speed-up the construction of random basis which will be detailed in the next section. Once the random basis is constructed and stored, we are ready to apply it to cases which we really interested in. The constructed random basis spans a finite-dimensional subspace in L (Ω) and we can project the solution u(x, ω) on this subspace, and write M u(x, ω) u i (x)a i (ω) (3.) i= where A =. We include A in the random basis to make the formulation concise. The expanded random basis {A i (ω)} M i= can be verified to be orthonormal. In this paper, we mainly consider elliptic-type PE with random coefficients. To demonstrate the main idea of online part, we consider elliptic PE with random coefficients and homogeneous boundary conditions are assumed, i.e., ( a pq (x, ω) u ) (x, ω) = f(x), x R d x p x q u(x, ω) =, where p, q =,..., d, Einstein summation is assumed and L(ω) = x p x ( a pq (x, ω) x q (x, ω) Similar formulations can be obtained for other types of stochastic differential operator. Algorithm (Online Computation).. expand the solution under the random basis {A i (ω)}, substitute into the SPE, multiply both side by A j (ω) and take expectations, i.e., Galerkin projection, which gives a set of PEs for u i (x). ( E [a pq A i A j ] u ) i = f(x), x, i, j =,,..., M x p x q u i (x) =, x ) 7

. Solve the coupled system of deterministic PEs by numerical methods, such as Finite Element Method, Finite ifference Method and etc. 3. Statistical information of the solution which are of primary interests can be calculated from deterministic coefficients. For example, the centered fourth moment can be computed as E [ (u ū) 4] = E [A i A j A k A l ] u i u j u k u l i,j,k,l where E [A i A j A k A l ] is readily available in the offline part. Remark 3.5. Besides statistical information of solutions, sometimes other types of information, such as extreme events and etc, is desired, which requires distribution information of solutions. In such cases, instead of storing only the statistical information of the random basis {A i (ω)} M i=, we may exploring their sparse representation under some predefined random basis, such as polynomial chaos basis, Haar chaos basis and etc and store the offline-constructed random basis completely. This part constitutes parts of ongoing work. 3. Complexity Analysis The applicability of our method depends on the dimension of the stochastic space that the solution lies in. If the dimension is large, we have to choose more stochastic bases and hence increase the complexity of the Galerkin projection. Proposition 3.. [?] If we consider the Gaussian covariance kernels of the form then the eigenvalues λ m has the properties R(x, x ) = σ exp( x x /γ ), (/γ)m/d λ m σ +, m, Γ(.5m /d ) where x lies in a unit d dimensional domain, and Γ is the gamma function. It is clear that the larger correlation length γ, the slower decaying of the eigenvalues. Consequently in our SMM, we have to adopt more bases because of the effective dimension increasing. Hence it will worse the efficiency of our method. 4 Multi-Level Monte Carlo Accelaration Although the offline computation may be expansive, the constructed random basis can be stored and used repeatedly in online computation for different cases. Various variance reduction techniques can be used to accelerate the offline computation, among which Multi-Level Monte Carlo method (MLMC) has been proved to be very effective for stochastic differential equation [?]. Here we will demonstrate the applicability of the MLMC to SPE. We denote v(x, ω) the stochastic function given by the SPE. It should be pointed out that v(x, ω) can any of the following: a MC solution of SPE, in which v(x, ω) L ( Ω), a covariance matrix of the solution, in which v(x, ω) L ( Ω). a function of the random basis (e.g., a(x, ω)a i (ω)a j (ω), or A i (ω)a j (ω)a k (ω)), in which v(x, ω) L ( Ω) or v(ω) L (Ω). 8

In addition, we define Mean-Squared Error (MSE) of Y with respect to the mean solution v as MSE = E [ Y v ]. If v(x, [ ω) is in ] the space of L ( Ω) or L ( Ω), following the Fubini s theorem, we have MSE = E (Y v) dx. While for v(ω) L (Ω), we have MSE = E [ (Y v) ]. For easy notation, we consider the first case only, and the theorem of second case v(ω) is easy to derive. We write Y as the MLMC estimate of the mean function v and it can be decomposed as with Y k = N k N k i= i= [v k (x, ω (k) i Y = Y = N v (x, ω () i ). N L Y k, k= ) v k (x, ω (k) i )], k =,..., L; Here L + is the number of levels used in MLMC, and N k is the number of MC simulations at the k-th level. Eventually MSE becomes [ ( L MSE = E Y k (x) v(x) ) ] dx = = + = = k= [ (E L ] Y k (x) v(x) ) ( L ) dx + Var Y k (x) k= (E [ L Var N k N k k= i= ( L N k N k k= i= [v k (x, ω (k) i [v k (x, ω (k) i k= ) v k (x, ω (k) i )] + N ) v k (x, ω (k) i )] + N ( L [ v k v k ] + v v(x) ) L dx + k= (E [vl ] v ) dx + L N k k= N k k= N i= N i= Var (v k v k ) dx + N v (x, ω () i ) ] v (x, ω () i ) v(x) ) dx ) Var (v k v k ) dx + N Var (v ) dx. Var (v ) dx Apparently the first term of MSE gives the bias error introduced by the numerical discretization at the finest level L. In order to compare the performance of MLMC to the single-level MC (SLMC), we assume h L is the smallest step-size that one can solve the MC solution. Hence for SLMC simulations, we choose h L as the step size, and N as the number of simulations to achieve ɛ -MSE. Then SLMC shares the same bias error as that of MLMC. In particular, the error is γɛ /(α + γ), which will be clear after we state the following theorem. Theorem 4.. Let Y be the estimate of the mean v and h k = M l k h L be the step size of k-th level with fixed integer M >. We assume the following estimate:. ( E [u k ] ū ) dx = c h α k ;. Var (u ) dx = c ; 9

3. Var (u k u k ) dx = c 3 h β k ; 4. the computational complexity of a k-th level MC simulation is c 4 h γ k. Then the multilevel estimator has the mean-square error with bound Y = L k= Y k MSE = E [ Y ū ] ɛ with a computational cost C c 5 ɛ = c 5 ɛ γ/α C, if β > γ; C c 6 ɛ (ln ɛ) = c 6 ɛ γ/α ln ɛ C, if β = γ; c 7 ɛ (γ β)/α = c 7 ɛ β/α C, if β < γ. Here the finest step size h L is chosen as the optimal step size for the Single-Level Monte Carlo simulation, i.e., ( γɛ ) /α, h L = c (α + γ) and the optimal computational cost C, subjected to ɛ -MSE condition, is C = c 4 Nh γ L 5 Numerical Examples 5. Elliptic PE with Random Coefficients = c c γ/α c 4 (α + γ) +γ/α αγ γ/α ɛ (+γ/α). The random basis {T α (ξ(ω))} α J used in Polynomial Chaos expansion can be constructed analytically by tensor product of one-dimensional polynomial chaos, i.e., T α (ξ(ω)) = + i= H αi (ξ i (ω)) where ξ = (ξ, ξ,... ), ξ i s are independent random variables with identical distribution µ which induces a unique system of orthogonal polynomial H i (x) s [?,?,?]. The index set J contains sequences of integers whose number of non-zero entries is finite, i.e., J = {(α, α,... ) α i =,,..., α i < } i= We call {T α (ξ(ω))} α J PC basis and {A i (ω)} i= SMM basis. It is seen immediately that the PC basis is a double-infinity sequence, i.e., both the order of one-dimensional PC basis and the dimensionality of random space can go to infinity, which leads to the so-called the cursor of dimensionality. To use the PC basis, we need to truncate the expansion on both directions which, due to the nature of tensor product, will generally results in a large number of coefficients to solve. We have shown in the previous section that it is possible that the effective dimensionality of the random space where the solution resides can be much smaller than it appears to be in PC basis. Therefore we focus us on the order of polynomials used in the PC basis since polynomials of high order are known to be very

our wce our wce 3 4 4 6 8 3 4 6 8 our wce 3 our wce 3 4 6 8 3 4 6 8 Figure : Error Comparison between PC and SMM. The errors of the mean, the variance, the third and the fourth central moments are shown, respectively from the left to the right and from the top to the bottom. oscillatory and have bad reputations in global interpolation. To this end, we restrict us in the case of onedimensional random space in the following example, i.e., ξ = (ξ ). The random elliptic coefficient adopted is moderately oscillatory in space and has some random phase, a(x, ω) = (sin(3πx + ξ) +.) (5.) where ξ is a Gaussian random variable with mean and unit variance. The SMM basis is constructed for f(x) = and both online computation and PC method are applied to f(x) = sin(πx) + 5 cos(4πx) with errors shown in fig. The accuracy of our method depends on the accuracy of offline MC simulations, thus the decay of the error of the mean solution slows down noticeably after the certain number of random basis. Before this number, our method generally achieves accuracy of one or two orders higher than PC method with the same number of basis. Since the online computational cost of SMM is about the same as that of PC method for the same number of random basis. Thus our method may use much less computational resource to achieve the same accuracy requirement. For higher order (>= 3) moments, PC method suffers from numerical instability and tends to blow up as the number of basis increases which is not a surprising result considering the oscillatory nature of high order polynomial. Therefore, high order Polynomial Chaos should be avoided especially high order moments are of interests. Randomness messes up the elegant deterministic world in many possible ways. One typical scenario arising in practice is that multiple sources of randomness coexist in the system. A typical example of such scenario

.....3.4.5.6.7..4.6.8.35.3.5..5..5..4.6.8.....3.4.7.6.5.4.3.. Exact 4 Mode 6 Mode 9 Mode.5..4.6.8..4.6.8 Figure 3: Moments of the solution of one-dimensional elliptic PE with random coefficients of large variance. The mean, the variance, the third and the fourth central moments of the random solution are shown, respectively from the left to the right and from the top to the bottom. The exact solution is computed with. 7 realizations. is atmospheric turbulence in which many eddies of intermediate length scale constitute multiple sources of randomness. The variance of quantities we are interested in may be comparable to the mean value in magnitude. To demonstrate the effectiveness of SMM for this scenario, we considered a(x, ω) = + 4 i ξ i (sin(c i πx) + ) (5.) ı= where = (,,, 5), C = (.,.3, 3., 4.3) and ξ i s are iid random variables. Clearly, this elliptic coefficient has four sources of random. The results are compared with the exact solution in fig.3. Obviously, our method achieves very good approximation for the mean and moments even with only M = 4, four random basis. If the stochastic dimensionality and the highest order of polynomials in the PC basis are r and p, respectively, PC method requires solving P = (r+p)! r!p! coefficients from a coupled PE system. Further, if the coupled PE system is discretized along each direction by N elements, the size of stiffness matrix is O ( (P N d ) ), which is astronomically large even for moderate values. In this example, the dimension of random space is r = 4 and, even for p = 3, the truncated PC expansion has P = 35 terms, while our method gives very good approximations by using 4 basis and nearly indistinguishable results by using 9 basis.

3.5 3.5.5.5...3.4.5.6.7.8.9 Figure 4: Elliptic coefficient with single jump at random location. Another typical scenario arising in industrial world is that the randomness is constrained in some spatial region, and some properties may change dramatically, but the exact location of such jump is random. For example, the interfaces of composite material may be random and the properties of materials on both sides of the interface may differ dramatically. Another example is the shocks surrounding air-crafts whose onset locations are affected by many factors and thus the exact locations of shocks can be viewed random. Our method can deal with such cases without any substantial modification. To this end, we consider a(x, ω) = arctan(4(x.5 +.4(ξ.5))) +.9 (5.3) where ξ is a uniformly distributed random variable [, ]. Some typical realizations of the elliptic coefficients are shown in fig.4 with results by SMM shown in fig.5 5. Elliptic PE with Random Coefficients We further apply our method to two-dimensional elliptic problem with random coefficients. Similar to onedimensional cases, two typical scenarios are considered here, solutions of large variance and elliptic coefficients jumping at random locations. For the first case, the following isotropic elliptic coefficient is used with results by SMM shown in fig.6. a(x, y, ω) =. + i ξ i (sin(e i πx + F i πy) + ) (5.4) ı= where = (, 3), E = (.3, 3.4), F = (3.7, 5.) and ξ i s are uniform random variable on [,]. Again, the SMM random basis {A i (ω)} M i= is constructed for f(x) = and applied to the case f(x) = sin(.3πx+ 3.4πy) + 8 cos(4.3πx 3.πy) For the second case, we consider a line-type jump at random locations, i.e., a(x, y, ω) = arctan{[(x.5 +.ξ ) + (y.5 +.ξ )]} +.9 (5.5) where ξ i s are uniform random variable on [.5, +.5]. Clearly, our method can capture moments of the random solutions with a few of basis. 3

4 6 8..4.6.8 4 3.5 3.5.5.5..4.6.8 4 3.5 3.5.5.5 5 4 3 Exact 4 Mode 6 Mode 9 Mode.5..4.6.8..4.6.8 Figure 5: Moments of the solution of one-dimensional elliptic PE with coefficient jumping at a single random location. The mean, the variance, the third and the fourth central moments of the random solution are shown, respectively from the left to the right and from the top to the bottom. The exact solution is computed with. 7 realizations. 4

Expectation MC Variance MC..5..5..5.5.5.5 Expectation mode( 4) Variance mode( 4)..5...5..5.5.5.5 Expectation mode( 6) Variance mode( 6)..5...5..5.5.5.5 Figure 6: Moments of the solution of two-dimensional elliptic PE with coefficient jumping at random locations. The mean, the variance, the third and the fourth central moments of the random solution are shown, respectively from the left to the right and from the top to the bottom. The exact solution is computed with. 5 realizations. 5.3 Helmholtz equation in random media In this example, we test our method on a different PE, Helmholtz equation, in the form of with boundary conditions u(x, ω) + k ( + n (x, ω))u(x, ω) =, u n iku = on Γ, u n = on Γ, u(x, ω) = f(x) on Γ 3, where n is the unit normal of the boundary, k is the wave number, and n (x, ω) is the random reflectivity of the media. Now the deterministic function f(x) appears on part of the boundary, instead of the right-hand-side of the PE. We also call this the Horn problem because of it derivation and physical structure (see Fig.8(a)). For numerical testing, we choose the random reflectivity of the media n (x, ω) = [ ξ (ω) sin (πx) sin (πy) + ξ (ω) cos (6πx) sin (8πy) ], with ξ and ξ are i.i.d uniformly distributed random variables in [, ]. Again we pick f(x) = to construct the basis off-line, and recover the Horn solution with f(x) = sin(π) sin(πy) + on-line. For the comparison purpose, we also ran the Monte-Carlo simulations to get 5

Expectation MC Variance MC..3....5.5.5.5 Expectation mode( 4) Variance mode( 4)..3....5.5.5.5 Expectation mode( 6) Variance mode( 6)..3....5.5.5.5 Figure 7: Moments of the solution of two-dimensional elliptic PE with coefficient jumping at random locations. The mean, the variance, the third and the fourth central moments of the random solution are shown, respectively from the left to the right and from the top to the bottom. The exact solution is computed with. 5 realizations. the moments of the solution, e.g., mean and standard deviation of the Horn solution. It is important to point out that when wave number k increases, the variance of the solution increases, so do the eigenvalues of the K-L expansion. Consequently for a fixed error tolerance, the problem of high-frequency needs significantly more bases than that of low-frequency (see Fig.8(b)). For example, if we discard all eigen-directions with eigenvalues less than 4, we need as much as bases for k =, comparing to 4 bases for k =.7. In particular, if we further increase the level to 6, we may need as much as bases for high-frequency case, which makes our method impractical. 6 Conclusions In this paper, the problem we target takes the general form L(ω)u(x, ω) = f(x) where randomness only appears in the stochastic differential operator. Thus, it is reasonable, at least heuristically, to conjecture that there exists a random basis which depends only, or largely, on L(ω) under which the solution for a board class of right-hand side f(x) has some sparse representation. We have verified this conjecture by many numerical examples presented in the previous sections and are working on rigorous ananlysis. Under the framework of Monte Carlo or Polynomial Chaos expansion, the computation must be re-done from the very beginning every time the right-hand-side function f(x) is changed which happens often in the design phase. Actually, many information from previous computations may be re-usable for the future computation. However, the questoin is how! Our method constructs a random basis based on KL expansion to extract the effective subspace of the output random space in the offline part and then reuse it in the online computation. In this sense, SMM can be viewed as partially inversion of the random differential operator. Obviously, the online computation cost of SMM is only a fraction of that of Monte Carlo methods. Furthermore, because 6

8 k= 6 k= k=.7 4 3 4 6 4 8 6 8 8 6 4 4 6 8 (a) Horn problem. 4 4 6 8 4 6 8 (b) Eigenvalue distribution of K-L expansion of the Horn solution for different wave number k. Figure 8: (Left) Physical domain of the horn problem, where we have far-field boundary condition on Γ, nonflow boundary condition on Γ, and irichlet boundary condition on Γ 3. (right) isplay the leading eigenvalue distribution of K-L expansion for different wave number k. the random basis in SMM is a sparse basis, it certainly requires much fewer random modes to achieve specified accuracy and much less computational resources compared to Polynomial Chaos expansion. Our method is also semi-non-intrusive in the sense that certified legacy codes can be used with minimum changes in offline part. This property makes it very attractive to industrial community as it can reuse codes verified by possibly expansive and time-consuming experiments and avoid another life cycle of developing new programs. We should point out that SMM can be extened to the cases where the right-hand-side function involves of randomness, i.e., f(x, ω) and the SPE is time dependent, which is the ongoing research work. Offline and online computations in this framework may have different characteristics of parallelism. We intend to exploit for 3 problems these difference on appropriate parallel architectures, for example, offline on GPU and online part on distributed-memory cluster. Acknowledgements This work was in part supported by the AFOSR MURI grant FA955-9--63. Appendix ( Proof of Stability) Consider elliptic equation with random coefficients x i ( a ij (x, ω) u x j (x, ω) ) = f(x), x R d (6.) u(x, ω) = x (6.) The solution u(x, ω) can be written in terms of Green s function u(x, ω) = G(x, y, ω)f(y)dy (6.3) If the expectation of Green s function Ḡ(x, y) = E [G(x, y, ω)] (6.4) 7

exists and the covariance function Cov G (x, y, x, y ) = E [ (G(x, y, ω) Ḡ(x, y))(g(x, y, ω) Ḡ(x, y )) ] (6.5) exists and continuous, Green s function G(x, y, ω) has KL expansion G(x, y, ω) = Ḡ(x, y) + λi G i (x, y)b i (ω) (6.6) where {λ i, G i (x, y)} i= are eigenpaires of covariance function Cov G(x, y, x, y ) Cov G (x, y, x, y )G i (x, y )dx dy = λ i G i (x, y) (6.7) i= with orthogonality conditions G i (x, y)g j (x, y)dxdy = δ ij (6.8) E [B i (ω)b j (ω)] = δ ij (6.9) and B i (ω) = λi Then the solution u(x, ω) can be written as u(x, ω) = ( G(x, y, ω) Ḡ(x, y) ) G i (x, y)dxdy (6.) λi F i (x)b i (ω) (6.) i= where F i (x) = G i (x, y)f(y)dy (6.) Lemma 6.. F i (x) f(x) Proof. For each x, Integrate on both sides F i (x) = = F i (x) = G i (x, y)f(y)dy G i (x, y) dy f (y)dy G i (x, y) dy f(y) Fi (x)dx G i (x, y) dydx f(y) = f(y) The orthogonality of G i is used in the last equality. 8

Now consider KL expansion of u(x, ω) u(x, ω) = ū(x) + θi φ i (x)a i (ω) (6.3) i= where { θ i, φ i (x)} i= are eigenpaires of the covariance function Cov u(x, y), i.e., Cov u (x, y)φ i (y)dy = θ i φ i (x) (6.4) and with bi-orthogonality Combine eqn.6. and eqn.6.3, we have A i (ω) = θi u(x, ω)φ(x)dx (6.5) φ i (x)φ j (x)dx = δ ij E [A i (ω)a j (ω)] = δ ij λj F j (x)b j (ω) = j= θj φ j (x)a j (ω) j= in which the mean ū(x) has been cancelled on both sides. Multiple both sides by F i (x) and integrate over, α ij λj B j (ω) = β ij θj A j (ω) (6.6) j= j= where α ij = β ij = F i (x)f j (x)dx F i (x)φ j (x)dx The equality can be written in matrix form ( ) ( ) α α Λ (B ) ( ) ( ) β β = Θ (A ) α α Λ B β β Θ A (6.7) where α R N N, Λ = diag (λ,..., λ N ), Λ = diag (λ N+,... ), β R N N, Θ = diag (θ,..., θ N ), Θ = diag (θ N+,... ), B R N and A R N. or α Λ B + α Λ B = β Θ A + β Θ A (6.8) α Λ B + α Λ B = β Θ A + β Θ A (6.9) Before proceeding to the error analysis, we have the following bounds Lemma 6.. α ij F i (x) F j (x) f(x) β ij F i (x) f(x) 9

Suppose we can construct {A i (ω)} N i= exactly without numerical errors in offline computation and apply this basis to the case with right hand side function g(x) L (). By solving the coupled equations resulting from Galerkin projection, we have the numerical solution N ṽ(x, ω) = ṽ i (x)a i (ω) (6.) i= The solution can be written in terms of Green s function where v(x, ω) = λi H i (x)b i (ω) (6.) i= H i (x) = G i (x, y)g(y)dy. (6.) By Lemma (6.), it is clear that H i (x) g(x). In addition, the expansion of v(x, ω) on the basis A i (ω) is v(x, ω) = v i (x)a i (ω). (6.3) Thus the error is where i= E(x, ω) = v(x, ω) ṽ(x, ω) = E A (x, ω) + E B (x, ω), N E A (x, ω) = [v i (x) ṽ(x)]a i (ω), (6.4) E B (x, ω) = i= N λi H i (x)b i (ω) v i (x)a i (ω). (6.5) i= i= The first error is due to Galerkin projection, and the second is due to approximation of B i by A i. simplicity, we assume g(x) =. To bound the second error term, we recall For v j (x) = E [v(x, ω)a j (ω)] = λi H i (x)e [B i (ω)a j (ω)], (6.6) i= and plug it into E B (x, ω). We have, after changing the order of summation, E B (x, ω) = E B (x, ω) + E B (x, ω), (6.7) and N N E B (x, ω) = λi H i (x) B i (ω) E [B i (ω)a j (ω)] A j (ω), i= E B (x, ω) = j= N λi H i (x) B i (ω) E [B i (ω)a j (ω)] A j (ω). i=n + j=

Theorem 6.3. Given the condition that N satisfies λi ɛ, i=n + there exists N, such that for N N, we have [ ] E E B (x, ω) dx ɛ. Proof. Proof is given by the following inequality [ ] [ ] E E B (x, ω) dx E E B (x, ω) dx [ + E ] E B (x, ω) dx, and Lemma (6.4), (6.5). Lemma 6.4. Given the condition that N satisfies λi ɛ, we have [ E i=n + ] E B (x, ω) dx ɛ. Proof. (Lemma 6.4) Based on Cauchy-Schwarz inequality, we have [ ] N E E B (x, ω) dx = E λi H i (x) B i (ω) E [B i (ω)a j (ω)] A j (ω) dx ( The last inequality is given by and i=n + i=n + i=n + λi λi ). H i (x) dx i=n + H i (x) g(x) =, j= λi E [ B i (ω) E [B i (ω)a j (ω)] A j (ω) [ E B i (ω) ] N E [B i (ω)a j (ω)] A j (ω) = (E [B i (ω)a j (ω)]). j= ] Lemma 6.5. For a fixed N, there exists a large N such that N N we have [ ] E E B (x, ω) dx ɛ.

Proof. (Lemma 6.5) Now apply Cauchy-Schwarz inequality again on the error term E B (x, ω) and we have [ ] N N E E B (x, ω) dx = E λi H i (x) B i (ω) E [B i (ω)a j (ω)] A j (ω) dx N i= i= i= j= N [ H i (x) dx λ i E B i (ω ] E [B i (ω)a j (ω)] A j (ω)) N N N λ i ( (E [B i (ω)a j (ω)]) ). i= j= Since {A i } i=,... is a complete orthonormal basis of the space L (Ω), for any normalized random variable B(ω) we have N lim (E [B(ω)A j (ω)]) =. Hence there exists integer N (i), such that N (i) N j= (E [B i (ω)a j (ω)]) ɛ min(, N λ ). i j= Now we choose and have [ E ] E B (x, ω) dx N N = max N (i), i=,...,n N N N λ i ( (E [B i (ω)a j (ω)]) ) i= i= j= N ɛ N λ i ( ) ( min(, N λ )) i ɛ

.5.5.5.5.5.5.5.5 (a) Mean solution of horn problem..7.6.5.4.3...7.6.5.4.3.. (b) Standard deviation of the solution Figure 9: (left) Monte Carlo results; (right) Recovered results via SMM using only 4 modes. 3

x 3 8 6 4 8 6 4 x 3 8 6 4 8 6 4 Figure : Relative errors of mean solution (left) and standard deviation (right). 4