Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques

Similar documents
Efficient Solvers for Stochastic Finite Element Saddle Point Problems

AND ELISABETH ULLMANN

Lecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:

Implementation of Sparse Wavelet-Galerkin FEM for Stochastic PDEs

Solving the Stochastic Steady-State Diffusion Problem Using Multigrid

Krylov subspace projection methods

H 2 -matrices with adaptive bases

Solving the steady state diffusion equation with uncertainty Final Presentation

Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms

A refined Lanczos method for computing eigenvalues and eigenvectors of unsymmetric matrices

arxiv: v2 [math.na] 8 Apr 2017

Multilevel accelerated quadrature for elliptic PDEs with random diffusion. Helmut Harbrecht Mathematisches Institut Universität Basel Switzerland

Hierarchical Parallel Solution of Stochastic Systems

PDEs, Matrix Functions and Krylov Subspace Methods

The Lanczos and conjugate gradient algorithms

Sampling and Low-Rank Tensor Approximations

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations

Large-scale eigenvalue problems

Iterative methods for symmetric eigenvalue problems

Adaptive Collocation with Kernel Density Estimation

Proper Orthogonal Decomposition (POD) for Nonlinear Dynamical Systems. Stefan Volkwein

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

Efficient Observation of Random Phenomena

Solving an Elliptic PDE Eigenvalue Problem via Automated Multi-Level Substructuring and Hierarchical Matrices

Numerical methods for the discretization of random fields by means of the Karhunen Loève expansion

CBS Constants and Their Role in Error Estimation for Stochastic Galerkin Finite Element Methods. Crowder, Adam J and Powell, Catherine E

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Solving the stochastic steady-state diffusion problem using multigrid

MATH 115A: SAMPLE FINAL SOLUTIONS

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

Lecture 2: Linear Algebra Review

Karhunen-Loève Approximation of Random Fields by Generalized Fast Multipole Methods

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Linear Algebra. Session 12

Krylov Subspaces. Lab 1. The Arnoldi Iteration

Experiences with Model Reduction and Interpolation

Linear Algebra Massoud Malek

Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries

Hilbert Space Methods for Reduced-Rank Gaussian Process Regression

Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms

Functional Analysis Review

Arnoldi Methods in SLEPc

Second-Order Inference for Gaussian Random Curves

Linear Algebra Practice Problems

Rational Krylov methods for linear and nonlinear eigenvalue problems

Concentration Ellipsoids

Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes

PART IV Spectral Methods

Stochastic Spectral Approaches to Bayesian Inference

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

A Vector-Space Approach for Stochastic Finite Element Analysis

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

S N. hochdimensionaler Lyapunov- und Sylvestergleichungen. Peter Benner. Mathematik in Industrie und Technik Fakultät für Mathematik TU Chemnitz

Golub-Kahan iterative bidiagonalization and determining the noise level in the data

Quasi-optimal and adaptive sparse grids with control variates for PDEs with random diffusion coefficient

FINITE-DIMENSIONAL LINEAR ALGEBRA

Block-diagonal preconditioning for spectral stochastic finite-element systems

BALANCING-RELATED MODEL REDUCTION FOR DATA-SPARSE SYSTEMS

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

Space-time sparse discretization of linear parabolic equations

Krylov Subspace Methods for Large/Sparse Eigenvalue Problems

Linear Algebra Review. Vectors

Low-rank techniques applied to moment equations for the stochastic Darcy problem with lognormal permeability

QUALIFYING EXAMINATION Harvard University Department of Mathematics Tuesday September 21, 2004 (Day 1)

Collocation based high dimensional model representation for stochastic partial differential equations

Efficient Implementation of Large Scale Lyapunov and Riccati Equation Solvers

Polynomial Chaos and Karhunen-Loeve Expansion

Review (Probability & Linear Algebra)

H(div) Preconditioning for a Mixed Finite Element Formulation of the Stochastic Diffusion Problem 1

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices

arxiv: v2 [math.na] 5 Dec 2013

Mathematical foundations - linear algebra

Review problems for MA 54, Fall 2004.

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

arxiv: v2 [math.pr] 27 Oct 2015

Fast Direct Methods for Gaussian Processes

Numerical Approximation of Stochastic Elliptic Partial Differential Equations

Matrix Equations and and Bivariate Function Approximation

Lecture Summaries for Linear Algebra M51A

Hierarchical Matrices. Jon Cockayne April 18, 2017

MAT Linear Algebra Collection of sample exams

Stabilization and Acceleration of Algebraic Multigrid Method

Krylov subspace methods for linear systems with tensor product structure

Background Mathematics (2/2) 1. David Barber

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

Multi-Element Probabilistic Collocation Method in High Dimensions

1 Singular Value Decomposition and Principal Component

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

Multigrid and stochastic sparse-grids techniques for PDE control problems with random coefficients

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Iterative methods for Linear System

On solving linear systems arising from Shishkin mesh discretizations

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Algebraic Multigrid as Solvers and as Preconditioner

Krylov Subspaces. The order-n Krylov subspace of A generated by x is

Sparse polynomial chaos expansions in engineering applications

MATH 590: Meshfree Methods

DEFLATED RESTARTING FOR MATRIX FUNCTIONS

Transcription:

Institut für Numerische Mathematik und Optimierung Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Oliver Ernst Computational Methods with Applications Harrachov, CR, August 19 25, 2007

1 Collaborators Catherine Powell, David Silvester University of Manchester, School of Mathematics, Manchester, UK Ingolf Busch, Michael Eiermann, Elisabeth Ullmann, TU Bergakademie Freiberg, Freiberg, Germany Support: DAAD/British Council

2 Outline Random fields and the Karhunen-Loève expansion Discretization of the covariance operator Solution of the discrete eigenvalue problem A numerical example

3 Random Fields

4 Formally stochastic process indexed by a spatial coordinate x D R d, D bounded, i.e., measurable function a : D Ω R, where (Ω, A, P ) is a given probability space For ω Ω fixed, a(, ω) is a realization of the random field, i.e., a function D R. For x D fixed, a(x, ) is a random variable (RV) w.r.t. (Ω, A, P ).

5 Notation ξ := Ω ξ(ω) dp (ω) expected value of RV ξ : Ω R a(x) := a(x, ) Cov a (x, y) := (a(x, ) a(x))(a(y, ) a(y)) Var a (x) := Cov a (x, x) σ a (x) := Var a (x) L 2 P (Ω) := {ξ : ξ 2 < } mean of RF a at x D covariance of RF a at x, y D variance of RF a at x D standard deviation of RF a at x D RV of second order

6 A RF is of second order, if a(x, ) L 2 P (Ω) for all x D. Theorem (Karhunen-Loève expansion). Given a second order RF a = a(x, ω) with continuous covariance function c(x, y) := Cov a (x, y), denote by {(λ m, a m (x))} the eigenpairs of the (compact) integral operator C : L 2 (D) L 2 (D), (Cu)(x) = u(y) c(x, y) dy, there exists a sequence {ξ m } m N of random variables with ξ m = 0 m, ξ m ξ n = δ m,n m, n such that the Karhunen-Loève (KL) expansion D a(x, ω) = a(x) + m=1 λm a m (x) ξ m (ω) (KL) converges uniformly on D and in L 2 P.

7 Note: Covariance functions c(x, y) are continuous on D D as well as symmetric and of positive type. Therefore covariance operators C are compact, hence spectra Λ(C) consist of countably many eigenvalues accumulating at most at zero. Covariance operators are selfadjoint and positive semidefinite. Analogy Singular value expansion of integral operator A : L 2 (D) L 2 P, f(x) (Af)(ω) := A : L 2 P L 2 (D), ξ(ω) (A ξ)(x) = C = A A. D Ω f(x)a(x, ω) dx, ξ(ω)a(x, ω) dp (ω)

8 Common Covariance Models Cov a (x, y) = c(x, y) = c(ρ), ρ = x y 1 1 1 c(r) c(r) c(r) 0 1 0 1 r exponential c(r) = σ 2 e ρ/l 0 1 0 1 r Bessel c(r) = σ 2 ρ l K 1( ρ l ) 0 1 0 1 r Gaussian c(r) = σ 2 e ρ2 /l 2 l > 0 is a measure of the correlation length, here l = 0.1, 1, 2.

9 Variance For normalized eigenfunctions a m (x), D Var a (x) = c(x, x) = Var a (x) dx = m=1 m=1 λ m a m (x) 2, λ m (a m, a m ) D }{{} =1 = trace C. For constant variance (e.g., stationary RF), Var a σ 2 > 0, λ m = D σ 2. m

10 Truncated KL Expansion For computational purposes, KL expansion truncated after M terms: a (M) (x, ω) = a(x) + M m=1 λm a m (x) ξ m (ω). Truncation error a a (M) 2 D = λ m. m=m+1 Choose M such that sufficient amount of total variance of RF is retained.

11 Eigenvalue Decay Roughly: the smoother the kernel, the faster {λ m } m N 0. More precisely: if D R d, then if the kernel function c is piecewise H r : λ m c 1 m r/d piecewise smooth : λ m c 2 m r for any r > 0 piecewise analytic : λ m c 3 exp( c 4 m 1/d ) for suitable constants c 1, c 2, c 3, c 4. Note: piecewise smoothness of kernel also leads to bounds on derivatives of eigenfunctions a m in L (D). Proven e.g. in [Schwab & Todor (2006)], [Todor (2006)]

12 Galerkin Discretization T h admissible finite element triangulation of D finite dimensional subspace of piecewise polynomials V h = {φ : D R : φ T P k T T } L 2 (D). Discrete eigenvalue problem: find pairs (λ h m, a h m) such that (Ca h m, φ) = λ h m(a h m, φ) φ V h, m = 1, 2,... corresponds to generalized matrix eigenvalue problem Cx = λmx, [C ] i,j = (Cφ j, φ i ), [M ] i,j = (φ j, φ i ), i, j = 1, 2,..., N = dim V h. C large and dense, M can be made diagonal using suitable basis.

13 Discretization Error Discrete operator given by C h = P h CP h, P h the L 2 (D) orthogonal projection to V h. Discrete eigenpairs {(λ h m, a h m)} N m=1 If covariance operator is piecewise smooth, then for any r > 0 ( ) 0 λ m λ h m K r h 2(k+1) λ 1 r m + h 4(k+1) λ 2r m, (I P h )a m L2 (D) K r λ r m h k+1. [Todor (2006)]

14 Solution of Matrix Eigenvalue Problem Only fixed number of leading eigenpairs required, suggests restarted Krylov subspace technique. We use the Thick-Restart Lanczos (TRL) method [Simon & Wu (2000)]. Idea: limit dimension of Krylov space to fixed m, save some desired approximate eigenpairs, generate new Krylov space which contains these retained approximations (restart). Krylov methods require inexpensive matrix-vector product. We obtain this by replacing C by a hierarchical matrix approximation C, for which matrix vector products can be computed in O(N log N) operations [Hackbusch (1999)].

15 Thick-Restart Lanczos Cycle (1) Given Lanczos decomposition of Krylov space K m (A, v) AQ m = Q m T m + β m+1 q m+1 em, Q m = [q 1,..., q m ], Q mq m = I m, (2) compute eigenpairs T m y j = ϑ j y j, j = 1,..., m, (3) select k < m Ritz vectors to retain, Y k := [y 1,..., y k ], (4) set Q k := Q m Y k, T k := Q k T Q m k to obtain A Q k = Q k Tk + β m+1 q k+1 s with q k+1 = q m+1 and s := Y k e m, (5) extend span{ q 1,..., q m+1 } to Krylov space of order m with Lanczostype decomposition A Q m = Q m Tm + β m+1 q m+1 e m

16 After restart cycle, projection T m of A on new Krylov space in A Q m = Q m Tm + β m+1 q m+1 e m has the form T m = T k βm s β m s α k+1 βk+1. β..... k+1....... βm β m α m Note: Leading k k block is diagonal.

17 Remarks: Mathematically equivalent to implicitly restarted Lanczos method and other augmented Krylov techniques, but more efficient. Takes advantage of symmetry (ARPACK uses full recurrences). Projected matrix T k readily available (= diag(ϑ 1,..., ϑ k )). Eigenvector residual norms from coordinate calculations (like in standard symmetric Lanczos). Well-known reorthogonalization techniques can be incorporated. For covariance problem: no shift-invert techniques required. Note: Need efficient matrix-vector product.

18 Hierarchical Matrix Approximation Idea: (recall survey in Monday s plenary talk of W. Hackbusch) Partition dense matrix into square blocks of 2 types near field blocks: computed and stored as usual far field blocks: approximated by matrix of low rank UV, computed by interpolation of kernel, store factors U, V. blocks correspond to clusters of degrees of freedom, i.e., clusters of supports of Galerkin basis functions block for pair of clusters s, t in near field if admissibility condition min{diam(d s ), diam(d t )} η dist(d s, D t ) satisfied by associated domains, η is the admissibility parameter.

19 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 1 2 3 4 5 6 7 8 9 1011121314151617181920212223242526272829303132

20 Remarks: Algebraic variant of fast multipole method Admissibility parameter η scales with correlation length. Necessary smoothness requirements satisfied for all common covariance kernels. Resulting data-sparse representation of discretized integral operator can be applied to a vector in O(N log N) operations (for N DOF). Need efficient quadrature for near field. An optimal approximation must thus balance the errors due to truncation of the KL series, Galerkin error in approximation a h m a m, λ h m λ m Lanczos approximation of discrete eigenpairs hierarchical matrix approximation C C

21 Numerical Example Bessel covariance kernel c(x, y) = x y l ( ) x y K 1, x, y D = [ 1, 1] 2. l Discretization: piecewise constant functions w.r.t. triangular mesh on D Hierarchical matrix parameters: interpolation polynomial degree : 4 admissibility constant : η = 1/l minimal block size : 62 Computations: MATLAB R2007a, Intel Xeon 5160, 3 GHz, 16 GB RAM calls to HLib-1.3 library (MPI Leipzig) via MEX

22 Some modes (l = 0.5) 1 1 0.8 0.8 0.6 0.6 0.4 0.4 0.2 0.2 mode 1 0 0 mode 4 0.2 0.2 0.4 0.4 0.6 0.6 0.8 0.8 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 1 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 1 0.8 0.8 0.6 0.6 0.4 0.4 0.2 0.2 mode 7 0 0 mode 8 0.2 0.2 0.4 0.4 0.6 0.6 0.8 0.8 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 1 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1

23 Performance of TRL N # evs % variance m restarts l = 0.5 402 36 94.99 44 5 1608 36 95.66 44 6 6432 36 95.87 44 5 25728 36 95.88 44 5 102912 36 95.51 44 5 l = 1 402 10 95.30 14 8 1608 10 95.46 14 8 6432 10 95.50 14 8 25728 10 95.51 14 9 102912 10 95.51 14 9 l = 2 402 4 95.30 7 8 1608 4 96.06 7 7 6432 4 96.10 7 7 25728 4 96.10 7 7 102912 4 96.11 7 7

24 Timings 10 7 10 6 N log(n) l=0.5 l=1 l=2 10 8 10 6 N log(n) l=0.5 l=1 l=2 10 5 tgen [s] 10 4 10 3 tlanczos [s] 10 4 10 2 10 2 10 1 10 0 10 2 10 3 10 4 10 5 10 6 N generation of hierarchical matrix approximation 10 0 10 2 10 2 10 3 10 4 10 5 10 6 N eigenvalue calculation

25 Conclusions Covariance eigenvalue problem challenging due to its size Can exploit regularity of covariance kernels Lanczos combined with hierarchical matrix approximation promising Becomes intractable for very small correlation lengths (too many relevant modes) Ongoing Work more careful tuning of hierarchical matrix approximation parameters multiple eigenvalues (symmetries in the domain) extend optimal quadrature techniques to 3D higher order FE approximation