Sampling and Low-Rank Tensor Approximations
|
|
- Roderick Hodge
- 5 years ago
- Views:
Transcription
1 Sampling and Low-Rank Tensor Approximations Hermann G. Matthies Alexander Litvinenko, Tarek A. El-Moshely +, Brunswick, Germany + MIT, Cambridge, MA, USA wire@tu-bs.de $Id: 2_Sydney-MCQMC.tex,v.3 22/2/2 6:52:28 hgm Exp $
2 Overview 2. Functionals of SPDE solutions 2. Computing the simulation 3. Parametric problems 4. Tensor products and other factorisations 5. Functional approximation 6. Emulation approximation 7. Examples and conclusion
3 Problem statement 3 We want to compute J k = E (Ψ k (, u e ( ))) = Ψ k (ω, u e (ω)) P(dω), where P is a probability measure on Ω, and u e is the solution of a PDE depending on the parameter ω Ω. A[ω](u e (ω)) = f(ω) a.s. in ω Ω, u e (ω) is a U-valued random variable (RV). To compute an approximation u M (ω) to u e (ω) via simulation is expensive, even for one value of ω, let alone for N J k Ψ k (ω n, u M (ω n )) w n n= Not all Ψ k of interest are known from the outset. Ω
4 Example: stochastic diffusion 4 Geometry 2 Aquifer D Model Simple stationary model of groundwater flow with stochastic data κ, f (κ(x, ω) u(x, ω)) = f(x, ω) x D R d & b.c. Solution is in tensor space S U =: W, e.g. W = L 2 (Ω, P) H (D) leads after Galerkin discretisation with U M = span{v m } M m= U to A[ω](u M (ω)) = f(ω) a.s. in ω Ω, where u M (ω) = M m= u m(ω)v m S U M..5
5 Realisation of κ(x, ω) 5
6 Solution example 6 Geometry 2.5 flow = flow out.5 2 Sources Dirichlet b.c Realization of κ Realization of solution Mean of solution Variance of solution x Pr{u(x) > 8}.5 y.5
7 Computing the simulation 7 To simulate u M one needs samples of the random field (RF) κ, which depends on infinitely many random variables (RVs). This has to be reduced / transformed Ξ : Ω [, ] s to a finite number s of RVs ξ = (ξ,..., ξ s ), with µ = Ξ P the push-forward measure: J k = Ψ k (ω, u e (ω)) P(dω) ˆΨk (ξ, u M (ξ)) µ(dξ). Ω [,] s This is a product measure for independent RVs (ξ,..., ξ s ). Approximate expensive simulation u M (ξ) by cheaper emulation. Both tasks are related by viewing u M : ξ u M (ξ), or κ : x κ(x, ) (RF indexed by x), or κ 2 : ω κ(, ω) (function valued RV), maps from a set of parameters into a vector space.
8 Parametric problems and RKHS 8 For each p in a parameter set P, let r(p) be an object in a Hilbert space V (for simplicity). With r : P V, denote U = span r(p) = span im r, then to each function r : P U corresponds a linear map R : U ˆR: R : U v r( ) v V ˆR = im R R P. (sometimes called a weak distribution) By construction R is injective. Use this to make ˆR a pre-hilbert space: φ, ψ ˆR : φ ψ R := R φ R ψ U. R is unitary on completion R which is a RKHS reproducing kernel Hilbert space with kernel ρ(p, p 2 ) = r(p ) r(p 2 ) U. Functions in R are in one-to-one correspondence with elements of U.
9 Covariance 9 If Q R P is Hilbert with inner product Q ; e.g. Q = L 2 (P, ν), define in U a positive self-adjoint map the covariance C = R R Cu v U = Ru Rv Q, has spectrum σ(c) R +, with spectral projectors E λ : C = λ de λ Similarly, define Ĉ : Q Q for φ, ψ Q such that Ĉ = RR by Ĉφ ψ Q = R φ R ψ U has same spectrum as C : σ(ĉ) = σ(c), and unitarily equivalent projectors Êλ = W E λ W : Ĉ = λ dêλ. Spectrum and projectors (σ(c), E λ ) are essence of r(p). Specifically, for φ, ψ L 2 (P, ν) we have R φ R ψ U = φ(p )ρ(p, p 2 )ψ(p 2 ) ν(dp ) ν(dp 2 ). P P
10 Covariance operator and SVD Cv = Spectral decomposition with projectors E λ λ de λ v = λ j e j v U e j + λ j σ p (C) R + \σ p (C) λ de λ v. C unitarily equivalent to multiplication operator M k with non-negative k: C = U M k U = (U M /2 k )(M /2 k U), with M /2 k = M k. This connects to the singular value decomposition (SVD) of R = V M /2 k U, with a (partial) isometry V. Often C has a pure point spectrum (e.g. C compact) last integral vanishes. In general to show tensors we have to invoke generalised eigenvectors and Gelfand triplets (rigged Hilbert spaces) for the continuous spectrum.
11 SVD, Karhunen-Loève-expansion, and tensors For sake of simplicity assume σ(c) = σ p (C). C = λ j e j U e j = λ j e j e j. j j (Rv)(p) = r(p) v U = j λj e j v U s j (p) with s j := Re j with R = j λj (s j e j ), or R = λj (e j s j ), r(p) = λj s j (p)e j, r S U. j j The singular value decomposition, a.k.a. Karhunen-Loève-expansion. A sum of rank- operators / tensors. In general C = R + λ e λ, e λ ϱ(dλ) with generalised eigenvectors e λ.
12 Examples and interpretations 2 If V is a space of centred random variables (RVs), r is a random field or stochastic process indexed by P, then Ĉ represented by the kernel ρ(p, p 2 ) is the covariance function. If in this case P = R d and moreover ρ(p, p 2 ) = c(p p 2 ) (stationary process / homogeneous field), then the diagonalisation U is effected by the Fourier transform, and the point spectrum is typically empty. If ν is a probability measure (ν(p) = ), and r is a V-valued RV, then C is the covariance operator. If P = {, 2,..., n} and R = R n, then ρ is the Gram matrix of the vectors r,..., r n. If n < dim V, the map R can be seen as a model reduction projector.
13 Factorisations / re-parametrisations 3 R serves as representation for Karhunen-Loève expansion. This is a factorisation of C. Some other possible ones: C = R R = (V M /2 k )(V M /2 k ) = C /2 C /2 = B B, where C = B B is an arbitrary one. Each factorisation leads to a representation all unitarily equivalent. (When C is a matrix, a favourite is Cholesky: C = LL ). Assume that C = B B and B : U H r U H. Select a orthonormal basis {e k } in H. Unitary Q : l 2 a = (a, a 2,...) k a ke k H. Approximation possible by injection Ps : R s l 2. Let r(a) := B Qa := R a (linear in a), i.e. R : l 2 U. Then R R = (B Q)(Q B) = B B = C.
14 Representations 4 Several representions for object r(p) U in a simpler space. The RKHS The Karhunen-Loève expansion based on spectral decomposition of C. The multiplicative spectral decomposition, as V M /2 k maps into U. Arbitrary factorisations C = B B. Analogous: consider Ĉ instead of C. If Q = L 2(P, ν) this leads to integral transforms, the kernel decompositions. These can all be used for model reduction, choosing a smaller subspace. Applied to RF κ(x, ω), and hence to u M (ω), yielding u M (ξ). Can again be applied to u M (ξ).
15 Functional approximation 5 Emulation replace expensive simulation u M (ξ) by inexpensive approximation / emulation u E (ξ) u M (ξ) ( alias response surfaces, proxy / surrogate models, etc.) Choose subspace S B S with basis {X β } B β=, make ansatz for each u m (ξ) β uβ mx β (ξ), giving u E (ξ) = m,β u β mx β (ξ)v m = m,β u β mx β (ξ) v m. Set U = (u β m) (M B). Sampling, we generate matrix / tensor U = [u M (ξ ),..., u M (ξ N )] = (u m (ξ n )) n m (M N).
16 Tensor product structure 6 Story does not end here as one may choose S = k S k, approximated by S B = K k= S B k, with S Bk S k. Solution represented( as a tensor of grade K + K ) in W B,N = k= S B k U N. For higher grade tensor product structure, more reduction is possible, but that is a story for another talk, here we stay with K =. With orthonormal X β one has N u β m = X β (ξ)u m (ξ) µ(dξ) w n X β (ξ n )u m (ξ n ). [,] s n= Let W = diag (w n ) (N N), X = (X β (ξ n )) (B N), hence U = U(W X T ). For B = N this is just a basis change.
17 Low-rank approximation 7 Focus on array of numbers U := [u m (ξ n )], view as matrix / tensor: N M U m,n e m M e n N, with unit vectors e n N R N, e m M R M. n= m= The sum has M N terms, the number of entries in U. Rank-R representation is approximation with R terms N M R U = U m,n e m M(e n N) T a l b T l = AB T, n= m= with A = [a,..., a R ] (M R) and B = [b,..., b R ] (N R). It contains only R (M + N) M N numbers. We will use updated, truncated SVD. This gives for coefficients U = U(W X T ) AB T (W X T ) = A(XW B) T =: A B T l=
18 Emulation instead of simulation 8 Let x(ξ) := [X (ξ),..., X B (ξ)] T. Emulator and low-rank emulator is u E (ξ) = Ux(ξ), and u L (ξ) := A B T x(ξ). Computing A, B: start with z samples U z = [u M (ξ ),..., u M (ξ z )]. Compute truncated, error controled SVD: M z U z M R W R R Σ ( z R V ) T ; then set A = W Σ /2, B = V Σ /2 B. For each n = z +,..., 2z, emulate u L (ξ n ) and evaluate residuum r n := r(ξ n ) := f(ξ n ) A[ξ n ](u L (ξ n )). If r n is small, accept u n A = u L(ξ n ), otherwise solve for u M (ξ n ) and set u n A = u M(ξ n ). Set U z2 = [u z+ A,..., u2z A ], compute updated SVD of [U z, U z2 ], A 2, B 2. Repeat for each batch of z samples.
19 Emulator in integration 9 J k = Ω To evaluate Ψ k (ω, u e (ω)) P(dω) ˆΨk (ξ, u M (ξ)) µ(dξ), [,] s we compute N J k w n ˆΨk (ξ n, u L (ξ n )). n= If we are lucky, we need much fewer than N samples to find the low-rank representation A, B for u L. This is cheap to compute from samples, and uses only little storage. In the integral the integrand is cheap to evaluate, and the low-rank representation can be re-used if a new (J k, Ψ k ) has to be evaluated.
20 Use in MC sampling solution sample 2 Example: Compressible RANS-flow around RAE air-foil. Sample solution turbulent kinetic energy pressure
21 Use in MC sampling solution storage 2 Inflow and air-foil shape uncertain. Data compression achieved by updated SVD: Made from 6 MC Simulations, SVD is updated every samples. M = 26, N = 6 Updated SVD: Relative errors, memory requirements: rank R pressure turb. kin. energy memory [MB].9e-2 4.e e-2 5.9e e-3.5e-4 4 Dense matrix R 26 6 costs 25 MB storage.
22 Use in QMC sampling mean 22 Trans-sonic flow with shock with N = 26 samples. Relative error for the density mean for rank R = 5,, 3, 5.
23 Use in QMC sampling variance 23 Trans-sonic flow with shock with N = 26 samples. Relative error for the density variance for rank R = 5,, 3, 5.
24 Conclusion 24 Random field discretisation and sampling can be seen as weak distribution with associated covariance. Analysis of associated linear map reveals essential structure. Factorisations of covariance lead to SVD (Karhunen-Loève expansion) and tensor products. Functional approximation to construct emulator. Sparse and inexpensive emulation.
Non-Intrusive Solution of Stochastic and Parametric Equations
Non-Intrusive Solution of Stochastic and Parametric Equations Hermann G. Matthies a Loïc Giraldi b, Alexander Litvinenko c, Dishi Liu d, and Anthony Nouy b a,, Brunswick, Germany b École Centrale de Nantes,
More informationSampling and low-rank tensor approximation of the response surface
Sampling and low-rank tensor approximation of the response surface tifica Alexander Litvinenko 1,2 (joint work with Hermann G. Matthies 3 ) 1 Group of Raul Tempone, SRI UQ, and 2 Group of David Keyes,
More informationParametric Problems, Stochastics, and Identification
Parametric Problems, Stochastics, and Identification Hermann G. Matthies a B. Rosić ab, O. Pajonk ac, A. Litvinenko a a, b University of Kragujevac c SPT Group, Hamburg wire@tu-bs.de http://www.wire.tu-bs.de
More informationQuantifying Uncertainty: Modern Computational Representation of Probability and Applications
Quantifying Uncertainty: Modern Computational Representation of Probability and Applications Hermann G. Matthies with Andreas Keese Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de
More informationNumerical Approximation of Stochastic Elliptic Partial Differential Equations
Numerical Approximation of Stochastic Elliptic Partial Differential Equations Hermann G. Matthies, Andreas Keese Institut für Wissenschaftliches Rechnen Technische Universität Braunschweig wire@tu-bs.de
More informationHierarchical Parallel Solution of Stochastic Systems
Hierarchical Parallel Solution of Stochastic Systems Second M.I.T. Conference on Computational Fluid and Solid Mechanics Contents: Simple Model of Stochastic Flow Stochastic Galerkin Scheme Resulting Equations
More informationParadigms of Probabilistic Modelling
Paradigms of Probabilistic Modelling Hermann G. Matthies Brunswick, Germany wire@tu-bs.de http://www.wire.tu-bs.de abstract RV-measure.tex,v 4.5 2017/07/06 01:56:46 hgm Exp Overview 2 1. Motivation challenges
More informationLecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:
tifica Lecture 1: Computation of Karhunen-Loeve Expansion: Alexander Litvinenko http://sri-uq.kaust.edu.sa/ Stochastic PDEs We consider div(κ(x, ω) u) = f (x, ω) in G, u = 0 on G, with stochastic coefficients
More informationImplementation of Sparse Wavelet-Galerkin FEM for Stochastic PDEs
Implementation of Sparse Wavelet-Galerkin FEM for Stochastic PDEs Roman Andreev ETH ZÜRICH / 29 JAN 29 TOC of the Talk Motivation & Set-Up Model Problem Stochastic Galerkin FEM Conclusions & Outlook Motivation
More informationPolynomial Chaos and Karhunen-Loeve Expansion
Polynomial Chaos and Karhunen-Loeve Expansion 1) Random Variables Consider a system that is modeled by R = M(x, t, X) where X is a random variable. We are interested in determining the probability of the
More informationGalerkin Methods for Linear and Nonlinear Elliptic Stochastic Partial Differential Equations
ScientifiComputing Galerkin Methods for Linear and Nonlinear Elliptic Stochastic Partial Differential Equations Hermann G. Matthies, Andreas Keese Institute of Scientific Computing Technical University
More informationKarhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques
Institut für Numerische Mathematik und Optimierung Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Oliver Ernst Computational Methods with Applications Harrachov, CR,
More informationMath Linear Algebra II. 1. Inner Products and Norms
Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,
More informationlinearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice
3. Eigenvalues and Eigenvectors, Spectral Representation 3.. Eigenvalues and Eigenvectors A vector ' is eigenvector of a matrix K, if K' is parallel to ' and ' 6, i.e., K' k' k is the eigenvalue. If is
More informationPartial Differential Equations with Stochastic Coefficients
Partial Differential Equations with Stochastic Coefficients Hermann G. Matthies gemeinsam mit Andreas Keese Institut für Wissenschaftliches Rechnen Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de
More informationEfficient Solvers for Stochastic Finite Element Saddle Point Problems
Efficient Solvers for Stochastic Finite Element Saddle Point Problems Catherine E. Powell c.powell@manchester.ac.uk School of Mathematics University of Manchester, UK Efficient Solvers for Stochastic Finite
More informationNON-LINEAR APPROXIMATION OF BAYESIAN UPDATE
tifica NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE Alexander Litvinenko 1, Hermann G. Matthies 2, Elmar Zander 2 http://sri-uq.kaust.edu.sa/ 1 Extreme Computing Research Center, KAUST, 2 Institute of Scientific
More informationSolving the Stochastic Steady-State Diffusion Problem Using Multigrid
Solving the Stochastic Steady-State Diffusion Problem Using Multigrid Tengfei Su Applied Mathematics and Scientific Computing Advisor: Howard Elman Department of Computer Science Sept. 29, 2015 Tengfei
More informationPlatzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen
Platzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen PARAMetric UNCertainties, Budapest STOCHASTIC PROCESSES AND FIELDS Noémi Friedman Institut für Wissenschaftliches Rechnen, wire@tu-bs.de
More informationStochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More informationLecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko
tifica Lecture 1: Introduction to low-rank tensor representation/approximation Alexander Litvinenko http://sri-uq.kaust.edu.sa/ KAUST Figure : KAUST campus, 5 years old, approx. 7000 people (include 1400
More information3. Some tools for the analysis of sequential strategies based on a Gaussian process prior
3. Some tools for the analysis of sequential strategies based on a Gaussian process prior E. Vazquez Computer experiments June 21-22, 2010, Paris 21 / 34 Function approximation with a Gaussian prior Aim:
More informationA Posteriori Adaptive Low-Rank Approximation of Probabilistic Models
A Posteriori Adaptive Low-Rank Approximation of Probabilistic Models Rainer Niekamp and Martin Krosche. Institute for Scientific Computing TU Braunschweig ILAS: 22.08.2011 A Posteriori Adaptive Low-Rank
More informationLinear Algebra 2 Final Exam, December 7, 2015 SOLUTIONS. a + 2b = x a + 3b = y. This solves to a = 3x 2y, b = y x. Thus
Linear Algebra 2 Final Exam, December 7, 2015 SOLUTIONS 1. (5.5 points) Let T : R 2 R 4 be a linear mapping satisfying T (1, 1) = ( 1, 0, 2, 3), T (2, 3) = (2, 3, 0, 0). Determine T (x, y) for (x, y) R
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More informationLecture notes: Applied linear algebra Part 1. Version 2
Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and
More informationKarhunen-Loève decomposition of Gaussian measures on Banach spaces
Karhunen-Loève decomposition of Gaussian measures on Banach spaces Jean-Charles Croix GT APSSE - April 2017, the 13th joint work with Xavier Bay. 1 / 29 Sommaire 1 Preliminaries on Gaussian processes 2
More informationKernel Method: Data Analysis with Positive Definite Kernels
Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University
More informationOptimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering
Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering Hermann G. Matthies Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de
More informationLinear Algebra using Dirac Notation: Pt. 2
Linear Algebra using Dirac Notation: Pt. 2 PHYS 476Q - Southern Illinois University February 6, 2018 PHYS 476Q - Southern Illinois University Linear Algebra using Dirac Notation: Pt. 2 February 6, 2018
More informationUncertainty Quantification and related areas - An overview
Uncertainty Quantification and related areas - An overview Alexander Litvinenko, talk given at Department of Mathematical Sciences, Durham University Bayesian Computational Statistics & Modeling, KAUST
More informationChapter 8 Integral Operators
Chapter 8 Integral Operators In our development of metrics, norms, inner products, and operator theory in Chapters 1 7 we only tangentially considered topics that involved the use of Lebesgue measure,
More informationLinear Algebra and Dirac Notation, Pt. 3
Linear Algebra and Dirac Notation, Pt. 3 PHYS 500 - Southern Illinois University February 1, 2017 PHYS 500 - Southern Illinois University Linear Algebra and Dirac Notation, Pt. 3 February 1, 2017 1 / 16
More informationAnalysis Preliminary Exam Workshop: Hilbert Spaces
Analysis Preliminary Exam Workshop: Hilbert Spaces 1. Hilbert spaces A Hilbert space H is a complete real or complex inner product space. Consider complex Hilbert spaces for definiteness. If (, ) : H H
More informationCharacterization of heterogeneous hydraulic conductivity field via Karhunen-Loève expansions and a measure-theoretic computational method
Characterization of heterogeneous hydraulic conductivity field via Karhunen-Loève expansions and a measure-theoretic computational method Jiachuan He University of Texas at Austin April 15, 2016 Jiachuan
More informationLinear Algebra in Actuarial Science: Slides to the lecture
Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations
More informationElementary linear algebra
Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The
More informationSummary of Week 9 B = then A A =
Summary of Week 9 Finding the square root of a positive operator Last time we saw that positive operators have a unique positive square root We now briefly look at how one would go about calculating the
More information(VII.E) The Singular Value Decomposition (SVD)
(VII.E) The Singular Value Decomposition (SVD) In this section we describe a generalization of the Spectral Theorem to non-normal operators, and even to transformations between different vector spaces.
More information1. The Polar Decomposition
A PERSONAL INTERVIEW WITH THE SINGULAR VALUE DECOMPOSITION MATAN GAVISH Part. Theory. The Polar Decomposition In what follows, F denotes either R or C. The vector space F n is an inner product space with
More informationarxiv: v2 [math.na] 8 Apr 2017
A LOW-RANK MULTIGRID METHOD FOR THE STOCHASTIC STEADY-STATE DIFFUSION PROBLEM HOWARD C. ELMAN AND TENGFEI SU arxiv:1612.05496v2 [math.na] 8 Apr 2017 Abstract. We study a multigrid method for solving large
More informationNon-square diagonal matrices
Non-square diagonal matrices Aim lecture: We apply the spectral theory of hermitian operators to look at linear maps T : V W which are not necessarily endomorphisms. This gives useful factorisations of
More informationSchwarz Preconditioner for the Stochastic Finite Element Method
Schwarz Preconditioner for the Stochastic Finite Element Method Waad Subber 1 and Sébastien Loisel 2 Preprint submitted to DD22 conference 1 Introduction The intrusive polynomial chaos approach for uncertainty
More informationEE731 Lecture Notes: Matrix Computations for Signal Processing
EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University October 17, 005 Lecture 3 3 he Singular Value Decomposition
More informationMath 671: Tensor Train decomposition methods II
Math 671: Tensor Train decomposition methods II Eduardo Corona 1 1 University of Michigan at Ann Arbor December 13, 2016 Table of Contents 1 What we ve talked about so far: 2 The Tensor Train decomposition
More informationThe following definition is fundamental.
1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationMonte Carlo Methods for Uncertainty Quantification
Monte Carlo Methods for Uncertainty Quantification Mike Giles Mathematical Institute, University of Oxford Contemporary Numerical Techniques Mike Giles (Oxford) Monte Carlo methods 1 / 23 Lecture outline
More informationLECTURE 7. k=1 (, v k)u k. Moreover r
LECTURE 7 Finite rank operators Definition. T is said to be of rank r (r < ) if dim T(H) = r. The class of operators of rank r is denoted by K r and K := r K r. Theorem 1. T K r iff T K r. Proof. Let T
More informationCollocation based high dimensional model representation for stochastic partial differential equations
Collocation based high dimensional model representation for stochastic partial differential equations S Adhikari 1 1 Swansea University, UK ECCM 2010: IV European Conference on Computational Mechanics,
More informationA New Scheme for the Tensor Representation
J Fourier Anal Appl (2009) 15: 706 722 DOI 10.1007/s00041-009-9094-9 A New Scheme for the Tensor Representation W. Hackbusch S. Kühn Received: 18 December 2008 / Revised: 29 June 2009 / Published online:
More informationCOMPACT OPERATORS. 1. Definitions
COMPACT OPERATORS. Definitions S:defi An operator M : X Y, X, Y Banach, is compact if M(B X (0, )) is relatively compact, i.e. it has compact closure. We denote { E:kk (.) K(X, Y ) = M L(X, Y ), M compact
More informationExperiences with Model Reduction and Interpolation
Experiences with Model Reduction and Interpolation Paul Constantine Stanford University, Sandia National Laboratories Qiqi Wang (MIT) David Gleich (Purdue) Emory University March 7, 2012 Joe Ruthruff (SNL)
More informationStochastic Finite Elements: Computational Approaches to SPDEs
Stochastic Finite Elements: Computational Approaches to SPDEs Hermann G. Matthies Andreas Keese, Elmar Zander, Alexander Litvinenko, Bojana Rosić Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de
More informationAn Empirical Chaos Expansion Method for Uncertainty Quantification
An Empirical Chaos Expansion Method for Uncertainty Quantification Melvin Leok and Gautam Wilkins Abstract. Uncertainty quantification seeks to provide a quantitative means to understand complex systems
More informationFermionic coherent states in infinite dimensions
Fermionic coherent states in infinite dimensions Robert Oeckl Centro de Ciencias Matemáticas Universidad Nacional Autónoma de México Morelia, Mexico Coherent States and their Applications CIRM, Marseille,
More informationMatrix decompositions
Matrix decompositions Zdeněk Dvořák May 19, 2015 Lemma 1 (Schur decomposition). If A is a symmetric real matrix, then there exists an orthogonal matrix Q and a diagonal matrix D such that A = QDQ T. The
More informationMultilevel accelerated quadrature for elliptic PDEs with random diffusion. Helmut Harbrecht Mathematisches Institut Universität Basel Switzerland
Multilevel accelerated quadrature for elliptic PDEs with random diffusion Mathematisches Institut Universität Basel Switzerland Overview Computation of the Karhunen-Loéve expansion Elliptic PDE with uniformly
More informationMath 113 Final Exam: Solutions
Math 113 Final Exam: Solutions Thursday, June 11, 2013, 3.30-6.30pm. 1. (25 points total) Let P 2 (R) denote the real vector space of polynomials of degree 2. Consider the following inner product on P
More informationDirect Learning: Linear Classification. Donglin Zeng, Department of Biostatistics, University of North Carolina
Direct Learning: Linear Classification Logistic regression models for classification problem We consider two class problem: Y {0, 1}. The Bayes rule for the classification is I(P(Y = 1 X = x) > 1/2) so
More informationDimensionality reduction of parameter-dependent problems through proper orthogonal decomposition
MATHICSE Mathematics Institute of Computational Science and Engineering School of Basic Sciences - Section of Mathematics MATHICSE Technical Report Nr. 01.2016 January 2016 (New 25.05.2016) Dimensionality
More informationA Randomized Algorithm for the Approximation of Matrices
A Randomized Algorithm for the Approximation of Matrices Per-Gunnar Martinsson, Vladimir Rokhlin, and Mark Tygert Technical Report YALEU/DCS/TR-36 June 29, 2006 Abstract Given an m n matrix A and a positive
More informationKarhunen-Loève decomposition of Gaussian measures on Banach spaces
Karhunen-Loève decomposition of Gaussian measures on Banach spaces Jean-Charles Croix jean-charles.croix@emse.fr Génie Mathématique et Industriel (GMI) First workshop on Gaussian processes at Saint-Etienne
More informationThe QR Factorization
The QR Factorization How to Make Matrices Nicer Radu Trîmbiţaş Babeş-Bolyai University March 11, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) The QR Factorization March 11, 2009 1 / 25 Projectors A projector
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationFinal Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson
Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson Name: TA Name and section: NO CALCULATORS, SHOW ALL WORK, NO OTHER PAPERS ON DESK. There is very little actual work to be done on this exam if
More informationECE 275A Homework #3 Solutions
ECE 75A Homework #3 Solutions. Proof of (a). Obviously Ax = 0 y, Ax = 0 for all y. To show sufficiency, note that if y, Ax = 0 for all y, then it must certainly be true for the particular value of y =
More informationCBS Constants and Their Role in Error Estimation for Stochastic Galerkin Finite Element Methods. Crowder, Adam J and Powell, Catherine E
CBS Constants and Their Role in Error Estimation for Stochastic Galerkin Finite Element Methods Crowder, Adam J and Powell, Catherine E 2017 MIMS EPrint: 2017.18 Manchester Institute for Mathematical Sciences
More informationPseudoinverse & Orthogonal Projection Operators
Pseudoinverse & Orthogonal Projection Operators ECE 174 Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 48 The Four
More informationPrincipal Component Analysis
Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used
More informationMP463 QUANTUM MECHANICS
MP463 QUANTUM MECHANICS Introduction Quantum theory of angular momentum Quantum theory of a particle in a central potential - Hydrogen atom - Three-dimensional isotropic harmonic oscillator (a model of
More information(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =
. (5 points) (a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? dim N(A), since rank(a) 3. (b) If we also know that Ax = has no solution, what do we know about the rank of A? C(A)
More information1 Math 241A-B Homework Problem List for F2015 and W2016
1 Math 241A-B Homework Problem List for F2015 W2016 1.1 Homework 1. Due Wednesday, October 7, 2015 Notation 1.1 Let U be any set, g be a positive function on U, Y be a normed space. For any f : U Y let
More informationPART IV Spectral Methods
PART IV Spectral Methods Additional References: R. Peyret, Spectral methods for incompressible viscous flow, Springer (2002), B. Mercier, An introduction to the numerical analysis of spectral methods,
More informationA Spectral Approach to Linear Bayesian Updating
A Spectral Approach to Linear Bayesian Updating Oliver Pajonk 1,2, Bojana V. Rosic 1, Alexander Litvinenko 1, and Hermann G. Matthies 1 1 Institute of Scientific Computing, TU Braunschweig, Germany 2 SPT
More informationREAL ANALYSIS II HOMEWORK 3. Conway, Page 49
REAL ANALYSIS II HOMEWORK 3 CİHAN BAHRAN Conway, Page 49 3. Let K and k be as in Proposition 4.7 and suppose that k(x, y) k(y, x). Show that K is self-adjoint and if {µ n } are the eigenvalues of K, each
More informationRKHS, Mercer s theorem, Unbounded domains, Frames and Wavelets Class 22, 2004 Tomaso Poggio and Sayan Mukherjee
RKHS, Mercer s theorem, Unbounded domains, Frames and Wavelets 9.520 Class 22, 2004 Tomaso Poggio and Sayan Mukherjee About this class Goal To introduce an alternate perspective of RKHS via integral operators
More informationDeterminant lines and determinant line bundles
CHAPTER Determinant lines and determinant line bundles This appendix is an exposition of G. Segal s work sketched in [?] on determinant line bundles over the moduli spaces of Riemann surfaces with parametrized
More informationA Vector-Space Approach for Stochastic Finite Element Analysis
A Vector-Space Approach for Stochastic Finite Element Analysis S Adhikari 1 1 Swansea University, UK CST2010: Valencia, Spain Adhikari (Swansea) Vector-Space Approach for SFEM 14-17 September, 2010 1 /
More informationLecture 9: Numerical Linear Algebra Primer (February 11st)
10-725/36-725: Convex Optimization Spring 2015 Lecture 9: Numerical Linear Algebra Primer (February 11st) Lecturer: Ryan Tibshirani Scribes: Avinash Siravuru, Guofan Wu, Maosheng Liu Note: LaTeX template
More informationScientific Computing WS 2017/2018. Lecture 18. Jürgen Fuhrmann Lecture 18 Slide 1
Scientific Computing WS 2017/2018 Lecture 18 Jürgen Fuhrmann juergen.fuhrmann@wias-berlin.de Lecture 18 Slide 1 Lecture 18 Slide 2 Weak formulation of homogeneous Dirichlet problem Search u H0 1 (Ω) (here,
More informationSeparation of Variables in Linear PDE: One-Dimensional Problems
Separation of Variables in Linear PDE: One-Dimensional Problems Now we apply the theory of Hilbert spaces to linear differential equations with partial derivatives (PDE). We start with a particular example,
More informationLinear Algebra and Dirac Notation, Pt. 2
Linear Algebra and Dirac Notation, Pt. 2 PHYS 500 - Southern Illinois University February 1, 2017 PHYS 500 - Southern Illinois University Linear Algebra and Dirac Notation, Pt. 2 February 1, 2017 1 / 14
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationMATH36001 Generalized Inverses and the SVD 2015
MATH36001 Generalized Inverses and the SVD 201 1 Generalized Inverses of Matrices A matrix has an inverse only if it is square and nonsingular. However there are theoretical and practical applications
More informationMaximal vectors in Hilbert space and quantum entanglement
Maximal vectors in Hilbert space and quantum entanglement William Arveson arveson@math.berkeley.edu UC Berkeley Summer 2008 arxiv:0712.4163 arxiv:0801.2531 arxiv:0804.1140 Overview Quantum Information
More informationApproximate Kernel PCA with Random Features
Approximate Kernel PCA with Random Features (Computational vs. Statistical Tradeoff) Bharath K. Sriperumbudur Department of Statistics, Pennsylvania State University Journées de Statistique Paris May 28,
More informationProposition 42. Let M be an m n matrix. Then (32) N (M M)=N (M) (33) R(MM )=R(M)
RODICA D. COSTIN. Singular Value Decomposition.1. Rectangular matrices. For rectangular matrices M the notions of eigenvalue/vector cannot be defined. However, the products MM and/or M M (which are square,
More informationLow Rank Approximation Lecture 7. Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL
Low Rank Approximation Lecture 7 Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL daniel.kressner@epfl.ch 1 Alternating least-squares / linear scheme General setting:
More informationIn terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.
1. GAUSSIAN PROCESSES A Gaussian process on a set T is a collection of random variables X =(X t ) t T on a common probability space such that for any n 1 and any t 1,...,t n T, the vector (X(t 1 ),...,X(t
More informationReview and problem list for Applied Math I
Review and problem list for Applied Math I (This is a first version of a serious review sheet; it may contain errors and it certainly omits a number of topic which were covered in the course. Let me know
More informationFinite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product
Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )
More informationIntroduction to Index Theory. Elmar Schrohe Institut für Analysis
Introduction to Index Theory Elmar Schrohe Institut für Analysis Basics Background In analysis and pde, you want to solve equations. In good cases: Linearize, end up with Au = f, where A L(E, F ) is a
More informationSecond Order Elliptic PDE
Second Order Elliptic PDE T. Muthukumar tmk@iitk.ac.in December 16, 2014 Contents 1 A Quick Introduction to PDE 1 2 Classification of Second Order PDE 3 3 Linear Second Order Elliptic Operators 4 4 Periodic
More information«Random Vectors» Lecture #2: Introduction Andreas Polydoros
«Random Vectors» Lecture #2: Introduction Andreas Polydoros Introduction Contents: Definitions: Correlation and Covariance matrix Linear transformations: Spectral shaping and factorization he whitening
More informationMathematics Department Stanford University Math 61CM/DM Inner products
Mathematics Department Stanford University Math 61CM/DM Inner products Recall the definition of an inner product space; see Appendix A.8 of the textbook. Definition 1 An inner product space V is a vector
More informationMatrix assembly by low rank tensor approximation
Matrix assembly by low rank tensor approximation Felix Scholz 13.02.2017 References Angelos Mantzaflaris, Bert Juettler, Boris Khoromskij, and Ulrich Langer. Matrix generation in isogeometric analysis
More informationSample ECE275A Midterm Exam Questions
Sample ECE275A Midterm Exam Questions The questions given below are actual problems taken from exams given in in the past few years. Solutions to these problems will NOT be provided. These problems and
More informationEE731 Lecture Notes: Matrix Computations for Signal Processing
EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten
More informationDominant feature extraction
Dominant feature extraction Francqui Lecture 7-5-200 Paul Van Dooren Université catholique de Louvain CESAME, Louvain-la-Neuve, Belgium Goal of this lecture Develop basic ideas for large scale dense matrices
More information