Sampling and Low-Rank Tensor Approximations

Size: px
Start display at page:

Download "Sampling and Low-Rank Tensor Approximations"

Transcription

1 Sampling and Low-Rank Tensor Approximations Hermann G. Matthies Alexander Litvinenko, Tarek A. El-Moshely +, Brunswick, Germany + MIT, Cambridge, MA, USA wire@tu-bs.de $Id: 2_Sydney-MCQMC.tex,v.3 22/2/2 6:52:28 hgm Exp $

2 Overview 2. Functionals of SPDE solutions 2. Computing the simulation 3. Parametric problems 4. Tensor products and other factorisations 5. Functional approximation 6. Emulation approximation 7. Examples and conclusion

3 Problem statement 3 We want to compute J k = E (Ψ k (, u e ( ))) = Ψ k (ω, u e (ω)) P(dω), where P is a probability measure on Ω, and u e is the solution of a PDE depending on the parameter ω Ω. A[ω](u e (ω)) = f(ω) a.s. in ω Ω, u e (ω) is a U-valued random variable (RV). To compute an approximation u M (ω) to u e (ω) via simulation is expensive, even for one value of ω, let alone for N J k Ψ k (ω n, u M (ω n )) w n n= Not all Ψ k of interest are known from the outset. Ω

4 Example: stochastic diffusion 4 Geometry 2 Aquifer D Model Simple stationary model of groundwater flow with stochastic data κ, f (κ(x, ω) u(x, ω)) = f(x, ω) x D R d & b.c. Solution is in tensor space S U =: W, e.g. W = L 2 (Ω, P) H (D) leads after Galerkin discretisation with U M = span{v m } M m= U to A[ω](u M (ω)) = f(ω) a.s. in ω Ω, where u M (ω) = M m= u m(ω)v m S U M..5

5 Realisation of κ(x, ω) 5

6 Solution example 6 Geometry 2.5 flow = flow out.5 2 Sources Dirichlet b.c Realization of κ Realization of solution Mean of solution Variance of solution x Pr{u(x) > 8}.5 y.5

7 Computing the simulation 7 To simulate u M one needs samples of the random field (RF) κ, which depends on infinitely many random variables (RVs). This has to be reduced / transformed Ξ : Ω [, ] s to a finite number s of RVs ξ = (ξ,..., ξ s ), with µ = Ξ P the push-forward measure: J k = Ψ k (ω, u e (ω)) P(dω) ˆΨk (ξ, u M (ξ)) µ(dξ). Ω [,] s This is a product measure for independent RVs (ξ,..., ξ s ). Approximate expensive simulation u M (ξ) by cheaper emulation. Both tasks are related by viewing u M : ξ u M (ξ), or κ : x κ(x, ) (RF indexed by x), or κ 2 : ω κ(, ω) (function valued RV), maps from a set of parameters into a vector space.

8 Parametric problems and RKHS 8 For each p in a parameter set P, let r(p) be an object in a Hilbert space V (for simplicity). With r : P V, denote U = span r(p) = span im r, then to each function r : P U corresponds a linear map R : U ˆR: R : U v r( ) v V ˆR = im R R P. (sometimes called a weak distribution) By construction R is injective. Use this to make ˆR a pre-hilbert space: φ, ψ ˆR : φ ψ R := R φ R ψ U. R is unitary on completion R which is a RKHS reproducing kernel Hilbert space with kernel ρ(p, p 2 ) = r(p ) r(p 2 ) U. Functions in R are in one-to-one correspondence with elements of U.

9 Covariance 9 If Q R P is Hilbert with inner product Q ; e.g. Q = L 2 (P, ν), define in U a positive self-adjoint map the covariance C = R R Cu v U = Ru Rv Q, has spectrum σ(c) R +, with spectral projectors E λ : C = λ de λ Similarly, define Ĉ : Q Q for φ, ψ Q such that Ĉ = RR by Ĉφ ψ Q = R φ R ψ U has same spectrum as C : σ(ĉ) = σ(c), and unitarily equivalent projectors Êλ = W E λ W : Ĉ = λ dêλ. Spectrum and projectors (σ(c), E λ ) are essence of r(p). Specifically, for φ, ψ L 2 (P, ν) we have R φ R ψ U = φ(p )ρ(p, p 2 )ψ(p 2 ) ν(dp ) ν(dp 2 ). P P

10 Covariance operator and SVD Cv = Spectral decomposition with projectors E λ λ de λ v = λ j e j v U e j + λ j σ p (C) R + \σ p (C) λ de λ v. C unitarily equivalent to multiplication operator M k with non-negative k: C = U M k U = (U M /2 k )(M /2 k U), with M /2 k = M k. This connects to the singular value decomposition (SVD) of R = V M /2 k U, with a (partial) isometry V. Often C has a pure point spectrum (e.g. C compact) last integral vanishes. In general to show tensors we have to invoke generalised eigenvectors and Gelfand triplets (rigged Hilbert spaces) for the continuous spectrum.

11 SVD, Karhunen-Loève-expansion, and tensors For sake of simplicity assume σ(c) = σ p (C). C = λ j e j U e j = λ j e j e j. j j (Rv)(p) = r(p) v U = j λj e j v U s j (p) with s j := Re j with R = j λj (s j e j ), or R = λj (e j s j ), r(p) = λj s j (p)e j, r S U. j j The singular value decomposition, a.k.a. Karhunen-Loève-expansion. A sum of rank- operators / tensors. In general C = R + λ e λ, e λ ϱ(dλ) with generalised eigenvectors e λ.

12 Examples and interpretations 2 If V is a space of centred random variables (RVs), r is a random field or stochastic process indexed by P, then Ĉ represented by the kernel ρ(p, p 2 ) is the covariance function. If in this case P = R d and moreover ρ(p, p 2 ) = c(p p 2 ) (stationary process / homogeneous field), then the diagonalisation U is effected by the Fourier transform, and the point spectrum is typically empty. If ν is a probability measure (ν(p) = ), and r is a V-valued RV, then C is the covariance operator. If P = {, 2,..., n} and R = R n, then ρ is the Gram matrix of the vectors r,..., r n. If n < dim V, the map R can be seen as a model reduction projector.

13 Factorisations / re-parametrisations 3 R serves as representation for Karhunen-Loève expansion. This is a factorisation of C. Some other possible ones: C = R R = (V M /2 k )(V M /2 k ) = C /2 C /2 = B B, where C = B B is an arbitrary one. Each factorisation leads to a representation all unitarily equivalent. (When C is a matrix, a favourite is Cholesky: C = LL ). Assume that C = B B and B : U H r U H. Select a orthonormal basis {e k } in H. Unitary Q : l 2 a = (a, a 2,...) k a ke k H. Approximation possible by injection Ps : R s l 2. Let r(a) := B Qa := R a (linear in a), i.e. R : l 2 U. Then R R = (B Q)(Q B) = B B = C.

14 Representations 4 Several representions for object r(p) U in a simpler space. The RKHS The Karhunen-Loève expansion based on spectral decomposition of C. The multiplicative spectral decomposition, as V M /2 k maps into U. Arbitrary factorisations C = B B. Analogous: consider Ĉ instead of C. If Q = L 2(P, ν) this leads to integral transforms, the kernel decompositions. These can all be used for model reduction, choosing a smaller subspace. Applied to RF κ(x, ω), and hence to u M (ω), yielding u M (ξ). Can again be applied to u M (ξ).

15 Functional approximation 5 Emulation replace expensive simulation u M (ξ) by inexpensive approximation / emulation u E (ξ) u M (ξ) ( alias response surfaces, proxy / surrogate models, etc.) Choose subspace S B S with basis {X β } B β=, make ansatz for each u m (ξ) β uβ mx β (ξ), giving u E (ξ) = m,β u β mx β (ξ)v m = m,β u β mx β (ξ) v m. Set U = (u β m) (M B). Sampling, we generate matrix / tensor U = [u M (ξ ),..., u M (ξ N )] = (u m (ξ n )) n m (M N).

16 Tensor product structure 6 Story does not end here as one may choose S = k S k, approximated by S B = K k= S B k, with S Bk S k. Solution represented( as a tensor of grade K + K ) in W B,N = k= S B k U N. For higher grade tensor product structure, more reduction is possible, but that is a story for another talk, here we stay with K =. With orthonormal X β one has N u β m = X β (ξ)u m (ξ) µ(dξ) w n X β (ξ n )u m (ξ n ). [,] s n= Let W = diag (w n ) (N N), X = (X β (ξ n )) (B N), hence U = U(W X T ). For B = N this is just a basis change.

17 Low-rank approximation 7 Focus on array of numbers U := [u m (ξ n )], view as matrix / tensor: N M U m,n e m M e n N, with unit vectors e n N R N, e m M R M. n= m= The sum has M N terms, the number of entries in U. Rank-R representation is approximation with R terms N M R U = U m,n e m M(e n N) T a l b T l = AB T, n= m= with A = [a,..., a R ] (M R) and B = [b,..., b R ] (N R). It contains only R (M + N) M N numbers. We will use updated, truncated SVD. This gives for coefficients U = U(W X T ) AB T (W X T ) = A(XW B) T =: A B T l=

18 Emulation instead of simulation 8 Let x(ξ) := [X (ξ),..., X B (ξ)] T. Emulator and low-rank emulator is u E (ξ) = Ux(ξ), and u L (ξ) := A B T x(ξ). Computing A, B: start with z samples U z = [u M (ξ ),..., u M (ξ z )]. Compute truncated, error controled SVD: M z U z M R W R R Σ ( z R V ) T ; then set A = W Σ /2, B = V Σ /2 B. For each n = z +,..., 2z, emulate u L (ξ n ) and evaluate residuum r n := r(ξ n ) := f(ξ n ) A[ξ n ](u L (ξ n )). If r n is small, accept u n A = u L(ξ n ), otherwise solve for u M (ξ n ) and set u n A = u M(ξ n ). Set U z2 = [u z+ A,..., u2z A ], compute updated SVD of [U z, U z2 ], A 2, B 2. Repeat for each batch of z samples.

19 Emulator in integration 9 J k = Ω To evaluate Ψ k (ω, u e (ω)) P(dω) ˆΨk (ξ, u M (ξ)) µ(dξ), [,] s we compute N J k w n ˆΨk (ξ n, u L (ξ n )). n= If we are lucky, we need much fewer than N samples to find the low-rank representation A, B for u L. This is cheap to compute from samples, and uses only little storage. In the integral the integrand is cheap to evaluate, and the low-rank representation can be re-used if a new (J k, Ψ k ) has to be evaluated.

20 Use in MC sampling solution sample 2 Example: Compressible RANS-flow around RAE air-foil. Sample solution turbulent kinetic energy pressure

21 Use in MC sampling solution storage 2 Inflow and air-foil shape uncertain. Data compression achieved by updated SVD: Made from 6 MC Simulations, SVD is updated every samples. M = 26, N = 6 Updated SVD: Relative errors, memory requirements: rank R pressure turb. kin. energy memory [MB].9e-2 4.e e-2 5.9e e-3.5e-4 4 Dense matrix R 26 6 costs 25 MB storage.

22 Use in QMC sampling mean 22 Trans-sonic flow with shock with N = 26 samples. Relative error for the density mean for rank R = 5,, 3, 5.

23 Use in QMC sampling variance 23 Trans-sonic flow with shock with N = 26 samples. Relative error for the density variance for rank R = 5,, 3, 5.

24 Conclusion 24 Random field discretisation and sampling can be seen as weak distribution with associated covariance. Analysis of associated linear map reveals essential structure. Factorisations of covariance lead to SVD (Karhunen-Loève expansion) and tensor products. Functional approximation to construct emulator. Sparse and inexpensive emulation.

Non-Intrusive Solution of Stochastic and Parametric Equations

Non-Intrusive Solution of Stochastic and Parametric Equations Non-Intrusive Solution of Stochastic and Parametric Equations Hermann G. Matthies a Loïc Giraldi b, Alexander Litvinenko c, Dishi Liu d, and Anthony Nouy b a,, Brunswick, Germany b École Centrale de Nantes,

More information

Sampling and low-rank tensor approximation of the response surface

Sampling and low-rank tensor approximation of the response surface Sampling and low-rank tensor approximation of the response surface tifica Alexander Litvinenko 1,2 (joint work with Hermann G. Matthies 3 ) 1 Group of Raul Tempone, SRI UQ, and 2 Group of David Keyes,

More information

Parametric Problems, Stochastics, and Identification

Parametric Problems, Stochastics, and Identification Parametric Problems, Stochastics, and Identification Hermann G. Matthies a B. Rosić ab, O. Pajonk ac, A. Litvinenko a a, b University of Kragujevac c SPT Group, Hamburg wire@tu-bs.de http://www.wire.tu-bs.de

More information

Quantifying Uncertainty: Modern Computational Representation of Probability and Applications

Quantifying Uncertainty: Modern Computational Representation of Probability and Applications Quantifying Uncertainty: Modern Computational Representation of Probability and Applications Hermann G. Matthies with Andreas Keese Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de

More information

Numerical Approximation of Stochastic Elliptic Partial Differential Equations

Numerical Approximation of Stochastic Elliptic Partial Differential Equations Numerical Approximation of Stochastic Elliptic Partial Differential Equations Hermann G. Matthies, Andreas Keese Institut für Wissenschaftliches Rechnen Technische Universität Braunschweig wire@tu-bs.de

More information

Hierarchical Parallel Solution of Stochastic Systems

Hierarchical Parallel Solution of Stochastic Systems Hierarchical Parallel Solution of Stochastic Systems Second M.I.T. Conference on Computational Fluid and Solid Mechanics Contents: Simple Model of Stochastic Flow Stochastic Galerkin Scheme Resulting Equations

More information

Paradigms of Probabilistic Modelling

Paradigms of Probabilistic Modelling Paradigms of Probabilistic Modelling Hermann G. Matthies Brunswick, Germany wire@tu-bs.de http://www.wire.tu-bs.de abstract RV-measure.tex,v 4.5 2017/07/06 01:56:46 hgm Exp Overview 2 1. Motivation challenges

More information

Lecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:

Lecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion: tifica Lecture 1: Computation of Karhunen-Loeve Expansion: Alexander Litvinenko http://sri-uq.kaust.edu.sa/ Stochastic PDEs We consider div(κ(x, ω) u) = f (x, ω) in G, u = 0 on G, with stochastic coefficients

More information

Implementation of Sparse Wavelet-Galerkin FEM for Stochastic PDEs

Implementation of Sparse Wavelet-Galerkin FEM for Stochastic PDEs Implementation of Sparse Wavelet-Galerkin FEM for Stochastic PDEs Roman Andreev ETH ZÜRICH / 29 JAN 29 TOC of the Talk Motivation & Set-Up Model Problem Stochastic Galerkin FEM Conclusions & Outlook Motivation

More information

Polynomial Chaos and Karhunen-Loeve Expansion

Polynomial Chaos and Karhunen-Loeve Expansion Polynomial Chaos and Karhunen-Loeve Expansion 1) Random Variables Consider a system that is modeled by R = M(x, t, X) where X is a random variable. We are interested in determining the probability of the

More information

Galerkin Methods for Linear and Nonlinear Elliptic Stochastic Partial Differential Equations

Galerkin Methods for Linear and Nonlinear Elliptic Stochastic Partial Differential Equations ScientifiComputing Galerkin Methods for Linear and Nonlinear Elliptic Stochastic Partial Differential Equations Hermann G. Matthies, Andreas Keese Institute of Scientific Computing Technical University

More information

Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques

Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Institut für Numerische Mathematik und Optimierung Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Oliver Ernst Computational Methods with Applications Harrachov, CR,

More information

Math Linear Algebra II. 1. Inner Products and Norms

Math Linear Algebra II. 1. Inner Products and Norms Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,

More information

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice 3. Eigenvalues and Eigenvectors, Spectral Representation 3.. Eigenvalues and Eigenvectors A vector ' is eigenvector of a matrix K, if K' is parallel to ' and ' 6, i.e., K' k' k is the eigenvalue. If is

More information

Partial Differential Equations with Stochastic Coefficients

Partial Differential Equations with Stochastic Coefficients Partial Differential Equations with Stochastic Coefficients Hermann G. Matthies gemeinsam mit Andreas Keese Institut für Wissenschaftliches Rechnen Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de

More information

Efficient Solvers for Stochastic Finite Element Saddle Point Problems

Efficient Solvers for Stochastic Finite Element Saddle Point Problems Efficient Solvers for Stochastic Finite Element Saddle Point Problems Catherine E. Powell c.powell@manchester.ac.uk School of Mathematics University of Manchester, UK Efficient Solvers for Stochastic Finite

More information

NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE

NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE tifica NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE Alexander Litvinenko 1, Hermann G. Matthies 2, Elmar Zander 2 http://sri-uq.kaust.edu.sa/ 1 Extreme Computing Research Center, KAUST, 2 Institute of Scientific

More information

Solving the Stochastic Steady-State Diffusion Problem Using Multigrid

Solving the Stochastic Steady-State Diffusion Problem Using Multigrid Solving the Stochastic Steady-State Diffusion Problem Using Multigrid Tengfei Su Applied Mathematics and Scientific Computing Advisor: Howard Elman Department of Computer Science Sept. 29, 2015 Tengfei

More information

Platzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen

Platzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen Platzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen PARAMetric UNCertainties, Budapest STOCHASTIC PROCESSES AND FIELDS Noémi Friedman Institut für Wissenschaftliches Rechnen, wire@tu-bs.de

More information

Stochastic Spectral Approaches to Bayesian Inference

Stochastic Spectral Approaches to Bayesian Inference Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to

More information

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko tifica Lecture 1: Introduction to low-rank tensor representation/approximation Alexander Litvinenko http://sri-uq.kaust.edu.sa/ KAUST Figure : KAUST campus, 5 years old, approx. 7000 people (include 1400

More information

3. Some tools for the analysis of sequential strategies based on a Gaussian process prior

3. Some tools for the analysis of sequential strategies based on a Gaussian process prior 3. Some tools for the analysis of sequential strategies based on a Gaussian process prior E. Vazquez Computer experiments June 21-22, 2010, Paris 21 / 34 Function approximation with a Gaussian prior Aim:

More information

A Posteriori Adaptive Low-Rank Approximation of Probabilistic Models

A Posteriori Adaptive Low-Rank Approximation of Probabilistic Models A Posteriori Adaptive Low-Rank Approximation of Probabilistic Models Rainer Niekamp and Martin Krosche. Institute for Scientific Computing TU Braunschweig ILAS: 22.08.2011 A Posteriori Adaptive Low-Rank

More information

Linear Algebra 2 Final Exam, December 7, 2015 SOLUTIONS. a + 2b = x a + 3b = y. This solves to a = 3x 2y, b = y x. Thus

Linear Algebra 2 Final Exam, December 7, 2015 SOLUTIONS. a + 2b = x a + 3b = y. This solves to a = 3x 2y, b = y x. Thus Linear Algebra 2 Final Exam, December 7, 2015 SOLUTIONS 1. (5.5 points) Let T : R 2 R 4 be a linear mapping satisfying T (1, 1) = ( 1, 0, 2, 3), T (2, 3) = (2, 3, 0, 0). Determine T (x, y) for (x, y) R

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

Karhunen-Loève decomposition of Gaussian measures on Banach spaces

Karhunen-Loève decomposition of Gaussian measures on Banach spaces Karhunen-Loève decomposition of Gaussian measures on Banach spaces Jean-Charles Croix GT APSSE - April 2017, the 13th joint work with Xavier Bay. 1 / 29 Sommaire 1 Preliminaries on Gaussian processes 2

More information

Kernel Method: Data Analysis with Positive Definite Kernels

Kernel Method: Data Analysis with Positive Definite Kernels Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University

More information

Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering

Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering Hermann G. Matthies Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de

More information

Linear Algebra using Dirac Notation: Pt. 2

Linear Algebra using Dirac Notation: Pt. 2 Linear Algebra using Dirac Notation: Pt. 2 PHYS 476Q - Southern Illinois University February 6, 2018 PHYS 476Q - Southern Illinois University Linear Algebra using Dirac Notation: Pt. 2 February 6, 2018

More information

Uncertainty Quantification and related areas - An overview

Uncertainty Quantification and related areas - An overview Uncertainty Quantification and related areas - An overview Alexander Litvinenko, talk given at Department of Mathematical Sciences, Durham University Bayesian Computational Statistics & Modeling, KAUST

More information

Chapter 8 Integral Operators

Chapter 8 Integral Operators Chapter 8 Integral Operators In our development of metrics, norms, inner products, and operator theory in Chapters 1 7 we only tangentially considered topics that involved the use of Lebesgue measure,

More information

Linear Algebra and Dirac Notation, Pt. 3

Linear Algebra and Dirac Notation, Pt. 3 Linear Algebra and Dirac Notation, Pt. 3 PHYS 500 - Southern Illinois University February 1, 2017 PHYS 500 - Southern Illinois University Linear Algebra and Dirac Notation, Pt. 3 February 1, 2017 1 / 16

More information

Analysis Preliminary Exam Workshop: Hilbert Spaces

Analysis Preliminary Exam Workshop: Hilbert Spaces Analysis Preliminary Exam Workshop: Hilbert Spaces 1. Hilbert spaces A Hilbert space H is a complete real or complex inner product space. Consider complex Hilbert spaces for definiteness. If (, ) : H H

More information

Characterization of heterogeneous hydraulic conductivity field via Karhunen-Loève expansions and a measure-theoretic computational method

Characterization of heterogeneous hydraulic conductivity field via Karhunen-Loève expansions and a measure-theoretic computational method Characterization of heterogeneous hydraulic conductivity field via Karhunen-Loève expansions and a measure-theoretic computational method Jiachuan He University of Texas at Austin April 15, 2016 Jiachuan

More information

Linear Algebra in Actuarial Science: Slides to the lecture

Linear Algebra in Actuarial Science: Slides to the lecture Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

Summary of Week 9 B = then A A =

Summary of Week 9 B = then A A = Summary of Week 9 Finding the square root of a positive operator Last time we saw that positive operators have a unique positive square root We now briefly look at how one would go about calculating the

More information

(VII.E) The Singular Value Decomposition (SVD)

(VII.E) The Singular Value Decomposition (SVD) (VII.E) The Singular Value Decomposition (SVD) In this section we describe a generalization of the Spectral Theorem to non-normal operators, and even to transformations between different vector spaces.

More information

1. The Polar Decomposition

1. The Polar Decomposition A PERSONAL INTERVIEW WITH THE SINGULAR VALUE DECOMPOSITION MATAN GAVISH Part. Theory. The Polar Decomposition In what follows, F denotes either R or C. The vector space F n is an inner product space with

More information

arxiv: v2 [math.na] 8 Apr 2017

arxiv: v2 [math.na] 8 Apr 2017 A LOW-RANK MULTIGRID METHOD FOR THE STOCHASTIC STEADY-STATE DIFFUSION PROBLEM HOWARD C. ELMAN AND TENGFEI SU arxiv:1612.05496v2 [math.na] 8 Apr 2017 Abstract. We study a multigrid method for solving large

More information

Non-square diagonal matrices

Non-square diagonal matrices Non-square diagonal matrices Aim lecture: We apply the spectral theory of hermitian operators to look at linear maps T : V W which are not necessarily endomorphisms. This gives useful factorisations of

More information

Schwarz Preconditioner for the Stochastic Finite Element Method

Schwarz Preconditioner for the Stochastic Finite Element Method Schwarz Preconditioner for the Stochastic Finite Element Method Waad Subber 1 and Sébastien Loisel 2 Preprint submitted to DD22 conference 1 Introduction The intrusive polynomial chaos approach for uncertainty

More information

EE731 Lecture Notes: Matrix Computations for Signal Processing

EE731 Lecture Notes: Matrix Computations for Signal Processing EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University October 17, 005 Lecture 3 3 he Singular Value Decomposition

More information

Math 671: Tensor Train decomposition methods II

Math 671: Tensor Train decomposition methods II Math 671: Tensor Train decomposition methods II Eduardo Corona 1 1 University of Michigan at Ann Arbor December 13, 2016 Table of Contents 1 What we ve talked about so far: 2 The Tensor Train decomposition

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Monte Carlo Methods for Uncertainty Quantification

Monte Carlo Methods for Uncertainty Quantification Monte Carlo Methods for Uncertainty Quantification Mike Giles Mathematical Institute, University of Oxford Contemporary Numerical Techniques Mike Giles (Oxford) Monte Carlo methods 1 / 23 Lecture outline

More information

LECTURE 7. k=1 (, v k)u k. Moreover r

LECTURE 7. k=1 (, v k)u k. Moreover r LECTURE 7 Finite rank operators Definition. T is said to be of rank r (r < ) if dim T(H) = r. The class of operators of rank r is denoted by K r and K := r K r. Theorem 1. T K r iff T K r. Proof. Let T

More information

Collocation based high dimensional model representation for stochastic partial differential equations

Collocation based high dimensional model representation for stochastic partial differential equations Collocation based high dimensional model representation for stochastic partial differential equations S Adhikari 1 1 Swansea University, UK ECCM 2010: IV European Conference on Computational Mechanics,

More information

A New Scheme for the Tensor Representation

A New Scheme for the Tensor Representation J Fourier Anal Appl (2009) 15: 706 722 DOI 10.1007/s00041-009-9094-9 A New Scheme for the Tensor Representation W. Hackbusch S. Kühn Received: 18 December 2008 / Revised: 29 June 2009 / Published online:

More information

COMPACT OPERATORS. 1. Definitions

COMPACT OPERATORS. 1. Definitions COMPACT OPERATORS. Definitions S:defi An operator M : X Y, X, Y Banach, is compact if M(B X (0, )) is relatively compact, i.e. it has compact closure. We denote { E:kk (.) K(X, Y ) = M L(X, Y ), M compact

More information

Experiences with Model Reduction and Interpolation

Experiences with Model Reduction and Interpolation Experiences with Model Reduction and Interpolation Paul Constantine Stanford University, Sandia National Laboratories Qiqi Wang (MIT) David Gleich (Purdue) Emory University March 7, 2012 Joe Ruthruff (SNL)

More information

Stochastic Finite Elements: Computational Approaches to SPDEs

Stochastic Finite Elements: Computational Approaches to SPDEs Stochastic Finite Elements: Computational Approaches to SPDEs Hermann G. Matthies Andreas Keese, Elmar Zander, Alexander Litvinenko, Bojana Rosić Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de

More information

An Empirical Chaos Expansion Method for Uncertainty Quantification

An Empirical Chaos Expansion Method for Uncertainty Quantification An Empirical Chaos Expansion Method for Uncertainty Quantification Melvin Leok and Gautam Wilkins Abstract. Uncertainty quantification seeks to provide a quantitative means to understand complex systems

More information

Fermionic coherent states in infinite dimensions

Fermionic coherent states in infinite dimensions Fermionic coherent states in infinite dimensions Robert Oeckl Centro de Ciencias Matemáticas Universidad Nacional Autónoma de México Morelia, Mexico Coherent States and their Applications CIRM, Marseille,

More information

Matrix decompositions

Matrix decompositions Matrix decompositions Zdeněk Dvořák May 19, 2015 Lemma 1 (Schur decomposition). If A is a symmetric real matrix, then there exists an orthogonal matrix Q and a diagonal matrix D such that A = QDQ T. The

More information

Multilevel accelerated quadrature for elliptic PDEs with random diffusion. Helmut Harbrecht Mathematisches Institut Universität Basel Switzerland

Multilevel accelerated quadrature for elliptic PDEs with random diffusion. Helmut Harbrecht Mathematisches Institut Universität Basel Switzerland Multilevel accelerated quadrature for elliptic PDEs with random diffusion Mathematisches Institut Universität Basel Switzerland Overview Computation of the Karhunen-Loéve expansion Elliptic PDE with uniformly

More information

Math 113 Final Exam: Solutions

Math 113 Final Exam: Solutions Math 113 Final Exam: Solutions Thursday, June 11, 2013, 3.30-6.30pm. 1. (25 points total) Let P 2 (R) denote the real vector space of polynomials of degree 2. Consider the following inner product on P

More information

Direct Learning: Linear Classification. Donglin Zeng, Department of Biostatistics, University of North Carolina

Direct Learning: Linear Classification. Donglin Zeng, Department of Biostatistics, University of North Carolina Direct Learning: Linear Classification Logistic regression models for classification problem We consider two class problem: Y {0, 1}. The Bayes rule for the classification is I(P(Y = 1 X = x) > 1/2) so

More information

Dimensionality reduction of parameter-dependent problems through proper orthogonal decomposition

Dimensionality reduction of parameter-dependent problems through proper orthogonal decomposition MATHICSE Mathematics Institute of Computational Science and Engineering School of Basic Sciences - Section of Mathematics MATHICSE Technical Report Nr. 01.2016 January 2016 (New 25.05.2016) Dimensionality

More information

A Randomized Algorithm for the Approximation of Matrices

A Randomized Algorithm for the Approximation of Matrices A Randomized Algorithm for the Approximation of Matrices Per-Gunnar Martinsson, Vladimir Rokhlin, and Mark Tygert Technical Report YALEU/DCS/TR-36 June 29, 2006 Abstract Given an m n matrix A and a positive

More information

Karhunen-Loève decomposition of Gaussian measures on Banach spaces

Karhunen-Loève decomposition of Gaussian measures on Banach spaces Karhunen-Loève decomposition of Gaussian measures on Banach spaces Jean-Charles Croix jean-charles.croix@emse.fr Génie Mathématique et Industriel (GMI) First workshop on Gaussian processes at Saint-Etienne

More information

The QR Factorization

The QR Factorization The QR Factorization How to Make Matrices Nicer Radu Trîmbiţaş Babeş-Bolyai University March 11, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) The QR Factorization March 11, 2009 1 / 25 Projectors A projector

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson Name: TA Name and section: NO CALCULATORS, SHOW ALL WORK, NO OTHER PAPERS ON DESK. There is very little actual work to be done on this exam if

More information

ECE 275A Homework #3 Solutions

ECE 275A Homework #3 Solutions ECE 75A Homework #3 Solutions. Proof of (a). Obviously Ax = 0 y, Ax = 0 for all y. To show sufficiency, note that if y, Ax = 0 for all y, then it must certainly be true for the particular value of y =

More information

CBS Constants and Their Role in Error Estimation for Stochastic Galerkin Finite Element Methods. Crowder, Adam J and Powell, Catherine E

CBS Constants and Their Role in Error Estimation for Stochastic Galerkin Finite Element Methods. Crowder, Adam J and Powell, Catherine E CBS Constants and Their Role in Error Estimation for Stochastic Galerkin Finite Element Methods Crowder, Adam J and Powell, Catherine E 2017 MIMS EPrint: 2017.18 Manchester Institute for Mathematical Sciences

More information

Pseudoinverse & Orthogonal Projection Operators

Pseudoinverse & Orthogonal Projection Operators Pseudoinverse & Orthogonal Projection Operators ECE 174 Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 48 The Four

More information

Principal Component Analysis

Principal Component Analysis Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used

More information

MP463 QUANTUM MECHANICS

MP463 QUANTUM MECHANICS MP463 QUANTUM MECHANICS Introduction Quantum theory of angular momentum Quantum theory of a particle in a central potential - Hydrogen atom - Three-dimensional isotropic harmonic oscillator (a model of

More information

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax = . (5 points) (a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? dim N(A), since rank(a) 3. (b) If we also know that Ax = has no solution, what do we know about the rank of A? C(A)

More information

1 Math 241A-B Homework Problem List for F2015 and W2016

1 Math 241A-B Homework Problem List for F2015 and W2016 1 Math 241A-B Homework Problem List for F2015 W2016 1.1 Homework 1. Due Wednesday, October 7, 2015 Notation 1.1 Let U be any set, g be a positive function on U, Y be a normed space. For any f : U Y let

More information

PART IV Spectral Methods

PART IV Spectral Methods PART IV Spectral Methods Additional References: R. Peyret, Spectral methods for incompressible viscous flow, Springer (2002), B. Mercier, An introduction to the numerical analysis of spectral methods,

More information

A Spectral Approach to Linear Bayesian Updating

A Spectral Approach to Linear Bayesian Updating A Spectral Approach to Linear Bayesian Updating Oliver Pajonk 1,2, Bojana V. Rosic 1, Alexander Litvinenko 1, and Hermann G. Matthies 1 1 Institute of Scientific Computing, TU Braunschweig, Germany 2 SPT

More information

REAL ANALYSIS II HOMEWORK 3. Conway, Page 49

REAL ANALYSIS II HOMEWORK 3. Conway, Page 49 REAL ANALYSIS II HOMEWORK 3 CİHAN BAHRAN Conway, Page 49 3. Let K and k be as in Proposition 4.7 and suppose that k(x, y) k(y, x). Show that K is self-adjoint and if {µ n } are the eigenvalues of K, each

More information

RKHS, Mercer s theorem, Unbounded domains, Frames and Wavelets Class 22, 2004 Tomaso Poggio and Sayan Mukherjee

RKHS, Mercer s theorem, Unbounded domains, Frames and Wavelets Class 22, 2004 Tomaso Poggio and Sayan Mukherjee RKHS, Mercer s theorem, Unbounded domains, Frames and Wavelets 9.520 Class 22, 2004 Tomaso Poggio and Sayan Mukherjee About this class Goal To introduce an alternate perspective of RKHS via integral operators

More information

Determinant lines and determinant line bundles

Determinant lines and determinant line bundles CHAPTER Determinant lines and determinant line bundles This appendix is an exposition of G. Segal s work sketched in [?] on determinant line bundles over the moduli spaces of Riemann surfaces with parametrized

More information

A Vector-Space Approach for Stochastic Finite Element Analysis

A Vector-Space Approach for Stochastic Finite Element Analysis A Vector-Space Approach for Stochastic Finite Element Analysis S Adhikari 1 1 Swansea University, UK CST2010: Valencia, Spain Adhikari (Swansea) Vector-Space Approach for SFEM 14-17 September, 2010 1 /

More information

Lecture 9: Numerical Linear Algebra Primer (February 11st)

Lecture 9: Numerical Linear Algebra Primer (February 11st) 10-725/36-725: Convex Optimization Spring 2015 Lecture 9: Numerical Linear Algebra Primer (February 11st) Lecturer: Ryan Tibshirani Scribes: Avinash Siravuru, Guofan Wu, Maosheng Liu Note: LaTeX template

More information

Scientific Computing WS 2017/2018. Lecture 18. Jürgen Fuhrmann Lecture 18 Slide 1

Scientific Computing WS 2017/2018. Lecture 18. Jürgen Fuhrmann Lecture 18 Slide 1 Scientific Computing WS 2017/2018 Lecture 18 Jürgen Fuhrmann juergen.fuhrmann@wias-berlin.de Lecture 18 Slide 1 Lecture 18 Slide 2 Weak formulation of homogeneous Dirichlet problem Search u H0 1 (Ω) (here,

More information

Separation of Variables in Linear PDE: One-Dimensional Problems

Separation of Variables in Linear PDE: One-Dimensional Problems Separation of Variables in Linear PDE: One-Dimensional Problems Now we apply the theory of Hilbert spaces to linear differential equations with partial derivatives (PDE). We start with a particular example,

More information

Linear Algebra and Dirac Notation, Pt. 2

Linear Algebra and Dirac Notation, Pt. 2 Linear Algebra and Dirac Notation, Pt. 2 PHYS 500 - Southern Illinois University February 1, 2017 PHYS 500 - Southern Illinois University Linear Algebra and Dirac Notation, Pt. 2 February 1, 2017 1 / 14

More information

Quantum Computing Lecture 2. Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces

More information

MATH36001 Generalized Inverses and the SVD 2015

MATH36001 Generalized Inverses and the SVD 2015 MATH36001 Generalized Inverses and the SVD 201 1 Generalized Inverses of Matrices A matrix has an inverse only if it is square and nonsingular. However there are theoretical and practical applications

More information

Maximal vectors in Hilbert space and quantum entanglement

Maximal vectors in Hilbert space and quantum entanglement Maximal vectors in Hilbert space and quantum entanglement William Arveson arveson@math.berkeley.edu UC Berkeley Summer 2008 arxiv:0712.4163 arxiv:0801.2531 arxiv:0804.1140 Overview Quantum Information

More information

Approximate Kernel PCA with Random Features

Approximate Kernel PCA with Random Features Approximate Kernel PCA with Random Features (Computational vs. Statistical Tradeoff) Bharath K. Sriperumbudur Department of Statistics, Pennsylvania State University Journées de Statistique Paris May 28,

More information

Proposition 42. Let M be an m n matrix. Then (32) N (M M)=N (M) (33) R(MM )=R(M)

Proposition 42. Let M be an m n matrix. Then (32) N (M M)=N (M) (33) R(MM )=R(M) RODICA D. COSTIN. Singular Value Decomposition.1. Rectangular matrices. For rectangular matrices M the notions of eigenvalue/vector cannot be defined. However, the products MM and/or M M (which are square,

More information

Low Rank Approximation Lecture 7. Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL

Low Rank Approximation Lecture 7. Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL Low Rank Approximation Lecture 7 Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL daniel.kressner@epfl.ch 1 Alternating least-squares / linear scheme General setting:

More information

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3. 1. GAUSSIAN PROCESSES A Gaussian process on a set T is a collection of random variables X =(X t ) t T on a common probability space such that for any n 1 and any t 1,...,t n T, the vector (X(t 1 ),...,X(t

More information

Review and problem list for Applied Math I

Review and problem list for Applied Math I Review and problem list for Applied Math I (This is a first version of a serious review sheet; it may contain errors and it certainly omits a number of topic which were covered in the course. Let me know

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

Introduction to Index Theory. Elmar Schrohe Institut für Analysis

Introduction to Index Theory. Elmar Schrohe Institut für Analysis Introduction to Index Theory Elmar Schrohe Institut für Analysis Basics Background In analysis and pde, you want to solve equations. In good cases: Linearize, end up with Au = f, where A L(E, F ) is a

More information

Second Order Elliptic PDE

Second Order Elliptic PDE Second Order Elliptic PDE T. Muthukumar tmk@iitk.ac.in December 16, 2014 Contents 1 A Quick Introduction to PDE 1 2 Classification of Second Order PDE 3 3 Linear Second Order Elliptic Operators 4 4 Periodic

More information

«Random Vectors» Lecture #2: Introduction Andreas Polydoros

«Random Vectors» Lecture #2: Introduction Andreas Polydoros «Random Vectors» Lecture #2: Introduction Andreas Polydoros Introduction Contents: Definitions: Correlation and Covariance matrix Linear transformations: Spectral shaping and factorization he whitening

More information

Mathematics Department Stanford University Math 61CM/DM Inner products

Mathematics Department Stanford University Math 61CM/DM Inner products Mathematics Department Stanford University Math 61CM/DM Inner products Recall the definition of an inner product space; see Appendix A.8 of the textbook. Definition 1 An inner product space V is a vector

More information

Matrix assembly by low rank tensor approximation

Matrix assembly by low rank tensor approximation Matrix assembly by low rank tensor approximation Felix Scholz 13.02.2017 References Angelos Mantzaflaris, Bert Juettler, Boris Khoromskij, and Ulrich Langer. Matrix generation in isogeometric analysis

More information

Sample ECE275A Midterm Exam Questions

Sample ECE275A Midterm Exam Questions Sample ECE275A Midterm Exam Questions The questions given below are actual problems taken from exams given in in the past few years. Solutions to these problems will NOT be provided. These problems and

More information

EE731 Lecture Notes: Matrix Computations for Signal Processing

EE731 Lecture Notes: Matrix Computations for Signal Processing EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten

More information

Dominant feature extraction

Dominant feature extraction Dominant feature extraction Francqui Lecture 7-5-200 Paul Van Dooren Université catholique de Louvain CESAME, Louvain-la-Neuve, Belgium Goal of this lecture Develop basic ideas for large scale dense matrices

More information