Polynomial Chaos and Karhunen-Loeve Expansion
|
|
- John Lucas
- 5 years ago
- Views:
Transcription
1 Polynomial Chaos and Karhunen-Loeve Expansion 1) Random Variables Consider a system that is modeled by R = M(x, t, X) where X is a random variable. We are interested in determining the probability of the response, p(r). If there is only one random variable, we write p(r) = p(x) dr/dx (1) If there are two random variables, X 1, X 2, then the determination of p(r) is more difficult, but in principle possible. For more than two random variables, or if p(x) is complicated, it it unlikely that we can obtain an analytic expression for p(r). 2) Monte Carlo and Surrogate Models Instead of analytic solutions, which are rarely possible, we will make use of the Monte Carlo approach in which we draw samples of X i from their distributions, compute R and from these obtain approximate distributions p(r). Unfortunately MC requires a very large number of samples to converge and if the model is complex and computationally difficult, this approach is usually not a realistic possibility. In this case we use surrogate models, that is we sample R for a small number of values of X and assume that behavior of the samples R(X) is a reasonable representation of the complete response for all values of X. One approach is to represent M(x, t, X) by a reduced order form, for example using Proper Orthogonal Decomposition (POD) and then sampling this reduced form using Latin Hypercube or a similar approach. 3) Polynomial Chaos, PC A Polynomial Chaos expression (PCE) is a way of representing an arbitrary random variable as a function of another random variable with a given distribution, and representing that function as a polynomial expansion. For example we might choose Ξ to have a uniform distribution, or a normal distribution. We will represent X in much the same way that a time series, f(t), can be represented by a Fourier Series of sines and cosines. Let X be the random variable and Ξ be the random variable in terms of which X will be expressed. In PC theory Ξ is often called the germ. In other words we write X = f(ξ) (2) and we seek an appropriate function f(.) such that given that Ξ has the desired germ distribution, that X will have its required distribution. 1
2 Polynomial Chaos expands f(ξ) as a polynomial series in terms of basis functions. For example if a sample from p(ξ) is denoted by ξ, we might use f(ξ) = a 0 + a 1 ξ +, a n ξ n. The specific form of this series depends upon the distribution of Ξ. While there is considerable freedom in choosing the basis functions of the series, it is most convenient to use orthogonal functions, just like sines and cosines are convenient. Let the polynomial expression be X = f(ξ) = (ψ 0 = 1, ψ 1, ψ 2, ) (3) where ψ j (ξ) is a polynomial or order j in terms of ξ which represents a sample (also called a realization) from the distribution p(ξ). With orthogonal functions we have the orthogonality relationship < ψ i, ψ j >= ψ i (ξ)ψ j (ξ)p(ξ)dξ = 0 (4) where the operation < a, b > is defined by Eq. 4. Since ψ i (ξ) are orthogonal to all other ψ j (ξ) including to ψ 0, all ψ j (ξ) for j 1 have zero mean and because of orthogonality they are uncorrelated, i.e., independent. Now using the orthogonal basis functions we write X = f(ξ) = x j ψ j (Ξ) (5) The expansion in the form of Eq. 5 is called a Polynomial Chaos Expansion (PCE). Given f(ξ) and ψ j, and making use of orthogonality we have In practice, the series in Eq. 5 is truncated x j =< f, ψ j > / < ψ j, ψ j > (6) X = f M (Ξ) x j ψ j (Ξ) (7) Thus, given a random variable X and knowing its distribution p(x), we can express a realization of X by Eq. 7. While in principle it is possible to go from X Ξ, in practice, one posits a polynomial with specific values of x i and creates the corresponding X, see Sec ) Examples Some common PC expansions are 1. If the basis functions are to represent uniformly distributed distributions of ξ in [-1,1], they are the Legendre functions 2. If the basis functions are to represent normal distributions of ξ, they are the Hermite polynomials, ψ 0 (ξ) = 1, ψ 1 (ξ) = ξ, ψ 2 (ξ) = ξ 2 1 2
3 3. If they represent an exponential random variable in [0, ) they are the Laguerre polynomials A common case is when X is assumed to be normally distributed. Then we write X = x 0 ψ 0 + x 1 ψ 1 (ξ) = µ + σξ (8) where µ and σ are the mean and standard deviation of X. 3.2) Multivariate PC Theory Both X and Ξ can be vectors, If Ξ is a vector, then the polynomial ψ j (Ξ) is multivariate, i.e., a vector, and is usually written as Ψ j (Ξ). When the components of Ξ are independent (not correlated), then Ψ j is a tensor product of the polynomial bases for each Ξ i and we write 3.3) Random Fields Ψ j (ξ) = m ψ j (ξ i ) (9) Consider a random process (also called a field), X(t). We then write X(t) 3.4) The output R = M(x, t, X) x j (t)ψ j (Ξ) (10) Since X is uncertain, so also is R. In practice both X and R are generally vectors, or may even be processes. We then write the output as or in practice R = g(ξ) = r j ψ j (Ξ) R g M (Ξ) = r j ψ j (Ξ) Although not necessary, it is usual to use the same base functions (i.e., the same germ) for g as for f and the same order of truncation, M. The final result is then r j ψ j (Ξ) M(x, t, x j (t)ψ j (Ξ)) (11) 3
4 4) Karhunen-Loeve Expansions, K-L The Karhunen-Loeve expansion is written as X(t) = E[X(t)] + λj φ j (t)ξ j (12) Comparing with Eq. 5 we see the the K-L expansion is a PCE because it expresses X(t) as a function of Ξ and looks just like the multi-variate form, Eq. 9, but only up to polynomial order 1. The importance of the K-L expansion because the λ i and φ j (t) are the eigenvalues and eigenvectors of the covariance of X(t), that is we can include correlations between the different random variables. 4.1) Isotropic Correlation Consider the case where the function R(.) depends on two random variables, X and Y. Suppose that X, Y are correlated and that the correlation depends only on the distance between the two arguments, i.e., on x y for scalars or x y for vector arguments. Then we are saying that the correlation j=1 correlation(x, Y )) = E[(X E(X])(Y E[Y ]) σ(x)σ(y) (13a) = c( x y ) (13b) Such a correlation is said to be isotropic. Given some reference correlation length ρ 0, and defining ρ = x y, ˆρ = ρ/ρ o, some popular correlation functions are Table 1: Isotropic Correlation Functions Constant c(ρ)=1 everywhere Exponential c(ρ) = e ˆρ Gaussian c(ρ) = e ˆρ2 Linear c(ρ) = max(1 ˆρ, 0) White Noise c(ρ) = 1 if ρ = 0, 0 otherwise 4.2) SVD Every real m by n matrix A has a singular value decomposition A = USV T (14) where 4
5 1. U is an m by m orthogonal matrix (U T U = I) 2. S is a m by n diagonal matrix with non-negative entries ordered in decreasing values 3. V is an n by n orthogonal matrix where r A = λi u i vi T λ i = eigenvalue of AA T r = rank of A (15a) (15b) (15c) If A k is the sum that extends only to k terms, then it can be shown that A A k 2 = λ max of (A A k ) T (A A k ) A A k 2 2 = λ k+1 (16a) (16b) In other words, A k is the best possible representation of A in a least squares sense. Comparing Eq. 15a with Eq. 12 we see that the SVD is the discrete version of the K-L expansion. 5) How does it work? The most common application of PCE for engineering problems uses the K-L expansion where the material property is assumed to vary spatially. The K-L treats the random property as having a Gaussian distribution, that is P(x) = E[P(x)] + λj φ j (t)ξ j (17) where ξ i is a sample from a Gaussian distribution. Consider a finite element program that solves Ku = f where K is called the stiffness matrix (conductance matrix for thermal problems). For linear problems K can be represented as j=1 Ku = (K 0 + K ξ )R = K 0 R 0 + ξ i K ξ R ξ = f (18) where K 0 is based upon the average value E[P(x)] and K ξ is based on the random part of Eq. 17. Thus we have M + 1 problems to solve. If M is large, this may be expensive, hence the frequent use of reduced order methods. 5
Stochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More informationAn Empirical Chaos Expansion Method for Uncertainty Quantification
An Empirical Chaos Expansion Method for Uncertainty Quantification Melvin Leok and Gautam Wilkins Abstract. Uncertainty quantification seeks to provide a quantitative means to understand complex systems
More informationA Vector-Space Approach for Stochastic Finite Element Analysis
A Vector-Space Approach for Stochastic Finite Element Analysis S Adhikari 1 1 Swansea University, UK CST2010: Valencia, Spain Adhikari (Swansea) Vector-Space Approach for SFEM 14-17 September, 2010 1 /
More informationRobustness of Principal Components
PCA for Clustering An objective of principal components analysis is to identify linear combinations of the original variables that are useful in accounting for the variation in those original variables.
More information5 Linear Algebra and Inverse Problem
5 Linear Algebra and Inverse Problem 5.1 Introduction Direct problem ( Forward problem) is to find field quantities satisfying Governing equations, Boundary conditions, Initial conditions. The direct problem
More informationCollocation based high dimensional model representation for stochastic partial differential equations
Collocation based high dimensional model representation for stochastic partial differential equations S Adhikari 1 1 Swansea University, UK ECCM 2010: IV European Conference on Computational Mechanics,
More informationBackground Mathematics (2/2) 1. David Barber
Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and
More informationKernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.
SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University
More informationSolving the Stochastic Steady-State Diffusion Problem Using Multigrid
Solving the Stochastic Steady-State Diffusion Problem Using Multigrid Tengfei Su Applied Mathematics and Scientific Computing Advisor: Howard Elman Department of Computer Science Sept. 29, 2015 Tengfei
More informationPARALLEL COMPUTATION OF 3D WAVE PROPAGATION BY SPECTRAL STOCHASTIC FINITE ELEMENT METHOD
13 th World Conference on Earthquake Engineering Vancouver, B.C., Canada August 1-6, 24 Paper No. 569 PARALLEL COMPUTATION OF 3D WAVE PROPAGATION BY SPECTRAL STOCHASTIC FINITE ELEMENT METHOD Riki Honda
More informationStructure in Data. A major objective in data analysis is to identify interesting features or structure in the data.
Structure in Data A major objective in data analysis is to identify interesting features or structure in the data. The graphical methods are very useful in discovering structure. There are basically two
More informationPolynomial chaos expansions for sensitivity analysis
c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty
More informationPreliminary Examination, Numerical Analysis, August 2016
Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any
More informationSchwarz Preconditioner for the Stochastic Finite Element Method
Schwarz Preconditioner for the Stochastic Finite Element Method Waad Subber 1 and Sébastien Loisel 2 Preprint submitted to DD22 conference 1 Introduction The intrusive polynomial chaos approach for uncertainty
More informationStochastic Solvers for the Euler Equations
43rd AIAA Aerospace Sciences Meeting and Exhibit 1-13 January 5, Reno, Nevada 5-873 Stochastic Solvers for the Euler Equations G. Lin, C.-H. Su and G.E. Karniadakis Division of Applied Mathematics Brown
More informationRisk Assessment for CO 2 Sequestration- Uncertainty Quantification based on Surrogate Models of Detailed Simulations
Risk Assessment for CO 2 Sequestration- Uncertainty Quantification based on Surrogate Models of Detailed Simulations Yan Zhang Advisor: Nick Sahinidis Department of Chemical Engineering Carnegie Mellon
More informationSparse polynomial chaos expansions in engineering applications
DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Sparse polynomial chaos expansions in engineering applications B. Sudret G. Blatman (EDF R&D,
More informationSampling and Low-Rank Tensor Approximations
Sampling and Low-Rank Tensor Approximations Hermann G. Matthies Alexander Litvinenko, Tarek A. El-Moshely +, Brunswick, Germany + MIT, Cambridge, MA, USA wire@tu-bs.de http://www.wire.tu-bs.de $Id: 2_Sydney-MCQMC.tex,v.3
More informationRemark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.
Sec 5 Eigenvectors and Eigenvalues In this chapter, vector means column vector Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called
More informationSection 5.2 Series Solution Near Ordinary Point
DE Section 5.2 Series Solution Near Ordinary Point Page 1 of 5 Section 5.2 Series Solution Near Ordinary Point We are interested in second order homogeneous linear differential equations with variable
More informationJordan Normal Form and Singular Decomposition
University of Debrecen Diagonalization and eigenvalues Diagonalization We have seen that if A is an n n square matrix, then A is diagonalizable if and only if for all λ eigenvalues of A we have dim(u λ
More informationKeywords: Sonic boom analysis, Atmospheric uncertainties, Uncertainty quantification, Monte Carlo method, Polynomial chaos method.
Blucher Mechanical Engineering Proceedings May 2014, vol. 1, num. 1 www.proceedings.blucher.com.br/evento/10wccm SONIC BOOM ANALYSIS UNDER ATMOSPHERIC UNCERTAINTIES BY A NON-INTRUSIVE POLYNOMIAL CHAOS
More informationDOA Estimation using MUSIC and Root MUSIC Methods
DOA Estimation using MUSIC and Root MUSIC Methods EE602 Statistical signal Processing 4/13/2009 Presented By: Chhavipreet Singh(Y515) Siddharth Sahoo(Y5827447) 2 Table of Contents 1 Introduction... 3 2
More informationEstimating functional uncertainty using polynomial chaos and adjoint equations
0. Estimating functional uncertainty using polynomial chaos and adjoint equations February 24, 2011 1 Florida State University, Tallahassee, Florida, Usa 2 Moscow Institute of Physics and Technology, Moscow,
More informationPART IV Spectral Methods
PART IV Spectral Methods Additional References: R. Peyret, Spectral methods for incompressible viscous flow, Springer (2002), B. Mercier, An introduction to the numerical analysis of spectral methods,
More informationRemark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.
Sec 6 Eigenvalues and Eigenvectors Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called an eigenvalue of A if there is a nontrivial
More informationEfficient Solvers for Stochastic Finite Element Saddle Point Problems
Efficient Solvers for Stochastic Finite Element Saddle Point Problems Catherine E. Powell c.powell@manchester.ac.uk School of Mathematics University of Manchester, UK Efficient Solvers for Stochastic Finite
More informationEE731 Lecture Notes: Matrix Computations for Signal Processing
EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten
More informationPrincipal Component Analysis and Linear Discriminant Analysis
Principal Component Analysis and Linear Discriminant Analysis Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1/29
More informationApplied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More informationBlock-diagonal preconditioning for spectral stochastic finite-element systems
IMA Journal of Numerical Analysis (2009) 29, 350 375 doi:0.093/imanum/drn04 Advance Access publication on April 4, 2008 Block-diagonal preconditioning for spectral stochastic finite-element systems CATHERINE
More informationPrincipal Component Analysis CS498
Principal Component Analysis CS498 Today s lecture Adaptive Feature Extraction Principal Component Analysis How, why, when, which A dual goal Find a good representation The features part Reduce redundancy
More informationVAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:
VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector
More informationSampling and low-rank tensor approximation of the response surface
Sampling and low-rank tensor approximation of the response surface tifica Alexander Litvinenko 1,2 (joint work with Hermann G. Matthies 3 ) 1 Group of Raul Tempone, SRI UQ, and 2 Group of David Keyes,
More informationLecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko
tifica Lecture 1: Introduction to low-rank tensor representation/approximation Alexander Litvinenko http://sri-uq.kaust.edu.sa/ KAUST Figure : KAUST campus, 5 years old, approx. 7000 people (include 1400
More informationHierarchical Parallel Solution of Stochastic Systems
Hierarchical Parallel Solution of Stochastic Systems Second M.I.T. Conference on Computational Fluid and Solid Mechanics Contents: Simple Model of Stochastic Flow Stochastic Galerkin Scheme Resulting Equations
More informationCME 345: MODEL REDUCTION
CME 345: MODEL REDUCTION Proper Orthogonal Decomposition (POD) Charbel Farhat & David Amsallem Stanford University cfarhat@stanford.edu 1 / 43 Outline 1 Time-continuous Formulation 2 Method of Snapshots
More informationUncertainty Propagation and Global Sensitivity Analysis in Hybrid Simulation using Polynomial Chaos Expansion
Uncertainty Propagation and Global Sensitivity Analysis in Hybrid Simulation using Polynomial Chaos Expansion EU-US-Asia workshop on hybrid testing Ispra, 5-6 October 2015 G. Abbiati, S. Marelli, O.S.
More informationStrain and stress computations in stochastic finite element. methods
INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN ENGINEERING Int. J. Numer. Meth. Engng 2007; 00:1 6 [Version: 2002/09/18 v2.02] Strain and stress computations in stochastic finite element methods Debraj
More informationLecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:
tifica Lecture 1: Computation of Karhunen-Loeve Expansion: Alexander Litvinenko http://sri-uq.kaust.edu.sa/ Stochastic PDEs We consider div(κ(x, ω) u) = f (x, ω) in G, u = 0 on G, with stochastic coefficients
More informationE = UV W (9.1) = I Q > V W
91 9. EOFs, SVD A common statistical tool in oceanography, meteorology and climate research are the so-called empirical orthogonal functions (EOFs). Anyone, in any scientific field, working with large
More informationUncertainty Evolution In Stochastic Dynamic Models Using Polynomial Chaos
Noname manuscript No. (will be inserted by the editor) Uncertainty Evolution In Stochastic Dynamic Models Using Polynomial Chaos Umamaheswara Konda Puneet Singla Tarunraj Singh Peter Scott Received: date
More informationConcentration Ellipsoids
Concentration Ellipsoids ECE275A Lecture Supplement Fall 2008 Kenneth Kreutz Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California, San Diego VERSION LSECE275CE
More informationA Polynomial Chaos Approach to Robust Multiobjective Optimization
A Polynomial Chaos Approach to Robust Multiobjective Optimization Silvia Poles 1, Alberto Lovison 2 1 EnginSoft S.p.A., Optimization Consulting Via Giambellino, 7 35129 Padova, Italy s.poles@enginsoft.it
More informationIV. Matrix Approximation using Least-Squares
IV. Matrix Approximation using Least-Squares The SVD and Matrix Approximation We begin with the following fundamental question. Let A be an M N matrix with rank R. What is the closest matrix to A that
More informationA reduced-order stochastic finite element analysis for structures with uncertainties
A reduced-order stochastic finite element analysis for structures with uncertainties Ji Yang 1, Béatrice Faverjon 1,2, Herwig Peters 1, icole Kessissoglou 1 1 School of Mechanical and Manufacturing Engineering,
More information11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.
C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a
More informationGaussian process regression for Sensitivity analysis
Gaussian process regression for Sensitivity analysis GPSS Workshop on UQ, Sheffield, September 2016 Nicolas Durrande, Mines St-Étienne, durrande@emse.fr GPSS workshop on UQ GPs for sensitivity analysis
More informationElectromagnetic Relaxation Time Distribution Inverse Problems in the Time-domain
Electromagnetic Relaxation Time Distribution Inverse Problems in the Time-domain Prof Nathan L Gibson Department of Mathematics Joint Math Meeting Jan 9, 2011 Prof Gibson (OSU) Inverse Problems for Distributions
More informationRandom Eigenvalue Problems Revisited
Random Eigenvalue Problems Revisited S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. Email: S.Adhikari@bristol.ac.uk URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html
More informationFoundations of Computer Vision
Foundations of Computer Vision Wesley. E. Snyder North Carolina State University Hairong Qi University of Tennessee, Knoxville Last Edited February 8, 2017 1 3.2. A BRIEF REVIEW OF LINEAR ALGEBRA Apply
More informationLecture 3: Review of Linear Algebra
ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters, transforms,
More informationChapter 2 Spectral Expansions
Chapter 2 Spectral Expansions In this chapter, we discuss fundamental and practical aspects of spectral expansions of random model data and of model solutions. We focus on a specific class of random process
More informationPrincipal Component Analysis
Principal Component Analysis November 24, 2015 From data to operators Given is data set X consisting of N vectors x n R D. Without loss of generality, assume x n = 0 (subtract mean). Let P be D N matrix
More informationSTOCHASTIC SAMPLING METHODS
STOCHASTIC SAMPLING METHODS APPROXIMATING QUANTITIES OF INTEREST USING SAMPLING METHODS Recall that quantities of interest often require the evaluation of stochastic integrals of functions of the solutions
More informationSobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis
Sobol-Hoeffding decomposition Application to Global SA Computation of the SI Sobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis Olivier Le Maître with Colleague & Friend Omar
More informationEigenvalues and diagonalization
Eigenvalues and diagonalization Patrick Breheny November 15 Patrick Breheny BST 764: Applied Statistical Modeling 1/20 Introduction The next topic in our course, principal components analysis, revolves
More informationApplied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in
More informationLet A an n n real nonsymmetric matrix. The eigenvalue problem: λ 1 = 1 with eigenvector u 1 = ( ) λ 2 = 2 with eigenvector u 2 = ( 1
Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Au = λu Example: ( ) 2 0 A = 2 1 λ 1 = 1 with eigenvector
More informationLinear Algebra- Final Exam Review
Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.
More informationEvaluation of Non-Intrusive Approaches for Wiener-Askey Generalized Polynomial Chaos
49th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference 6t 7 - April 28, Schaumburg, IL AIAA 28-892 Evaluation of Non-Intrusive Approaches for Wiener-Askey Generalized
More informationIntroduction to Uncertainty Quantification in Computational Science Handout #3
Introduction to Uncertainty Quantification in Computational Science Handout #3 Gianluca Iaccarino Department of Mechanical Engineering Stanford University June 29 - July 1, 2009 Scuola di Dottorato di
More informationConcepts in Global Sensitivity Analysis IMA UQ Short Course, June 23, 2015
Concepts in Global Sensitivity Analysis IMA UQ Short Course, June 23, 2015 A good reference is Global Sensitivity Analysis: The Primer. Saltelli, et al. (2008) Paul Constantine Colorado School of Mines
More informationLECTURE 16: PCA AND SVD
Instructor: Sael Lee CS549 Computational Biology LECTURE 16: PCA AND SVD Resource: PCA Slide by Iyad Batal Chapter 12 of PRML Shlens, J. (2003). A tutorial on principal component analysis. CONTENT Principal
More informationIntroduction to Mathematical Physics
Introduction to Mathematical Physics Methods and Concepts Second Edition Chun Wa Wong Department of Physics and Astronomy University of California Los Angeles OXFORD UNIVERSITY PRESS Contents 1 Vectors
More informationQuantifying Uncertainty: Modern Computational Representation of Probability and Applications
Quantifying Uncertainty: Modern Computational Representation of Probability and Applications Hermann G. Matthies with Andreas Keese Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de
More informationUnsupervised Learning: Dimensionality Reduction
Unsupervised Learning: Dimensionality Reduction CMPSCI 689 Fall 2015 Sridhar Mahadevan Lecture 3 Outline In this lecture, we set about to solve the problem posed in the previous lecture Given a dataset,
More informationOrganization. I MCMC discussion. I project talks. I Lecture.
Organization I MCMC discussion I project talks. I Lecture. Content I Uncertainty Propagation Overview I Forward-Backward with an Ensemble I Model Reduction (Intro) Uncertainty Propagation in Causal Systems
More informationModule 9: Stationary Processes
Module 9: Stationary Processes Lecture 1 Stationary Processes 1 Introduction A stationary process is a stochastic process whose joint probability distribution does not change when shifted in time or space.
More informationPreviously Monte Carlo Integration
Previously Simulation, sampling Monte Carlo Simulations Inverse cdf method Rejection sampling Today: sampling cont., Bayesian inference via sampling Eigenvalues and Eigenvectors Markov processes, PageRank
More informationUNIT 6: The singular value decomposition.
UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationDimensionality reduction of parameter-dependent problems through proper orthogonal decomposition
MATHICSE Mathematics Institute of Computational Science and Engineering School of Basic Sciences - Section of Mathematics MATHICSE Technical Report Nr. 01.2016 January 2016 (New 25.05.2016) Dimensionality
More informationPCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani
PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given
More informationStochastic Dimension Reduction
Stochastic Dimension Reduction Roger Ghanem University of Southern California Los Angeles, CA, USA Computational and Theoretical Challenges in Interdisciplinary Predictive Modeling Over Random Fields 12th
More informationEIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4]
EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: Au = λu λ C : eigenvalue u C n : eigenvector Example:
More informationSolving the steady state diffusion equation with uncertainty Final Presentation
Solving the steady state diffusion equation with uncertainty Final Presentation Virginia Forstall vhfors@gmail.com Advisor: Howard Elman elman@cs.umd.edu Department of Computer Science May 6, 2012 Problem
More informationApplications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices
Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Vahid Dehdari and Clayton V. Deutsch Geostatistical modeling involves many variables and many locations.
More informationComputational Methods. Eigenvalues and Singular Values
Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations
More informationPolynomial chaos expansions for structural reliability analysis
DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for structural reliability analysis B. Sudret & S. Marelli Incl.
More informationLecture: Face Recognition and Feature Reduction
Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed
More informationECE 598: Representation Learning: Algorithms and Models Fall 2017
ECE 598: Representation Learning: Algorithms and Models Fall 2017 Lecture 1: Tensor Methods in Machine Learning Lecturer: Pramod Viswanathan Scribe: Bharath V Raghavan, Oct 3, 2017 11 Introduction Tensors
More informationLinear Algebra. and
Instructions Please answer the six problems on your own paper. These are essay questions: you should write in complete sentences. 1. Are the two matrices 1 2 2 1 3 5 2 7 and 1 1 1 4 4 2 5 5 2 row equivalent?
More informationLinear Algebra - Part II
Linear Algebra - Part II Projection, Eigendecomposition, SVD (Adapted from Sargur Srihari s slides) Brief Review from Part 1 Symmetric Matrix: A = A T Orthogonal Matrix: A T A = AA T = I and A 1 = A T
More information12.4 Known Channel (Water-Filling Solution)
ECEn 665: Antennas and Propagation for Wireless Communications 54 2.4 Known Channel (Water-Filling Solution) The channel scenarios we have looed at above represent special cases for which the capacity
More informationReview of Some Concepts from Linear Algebra: Part 2
Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More informationSENSITIVITY ANALYSIS IN NUMERICAL SIMULATION OF MULTIPHASE FLOW FOR CO 2 STORAGE IN SALINE AQUIFERS USING THE PROBABILISTIC COLLOCATION APPROACH
XIX International Conference on Water Resources CMWR 2012 University of Illinois at Urbana-Champaign June 17-22,2012 SENSITIVITY ANALYSIS IN NUMERICAL SIMULATION OF MULTIPHASE FLOW FOR CO 2 STORAGE IN
More informationLinear Algebra review Powers of a diagonalizable matrix Spectral decomposition
Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2016 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing
More informationUNIVERSITY OF CALIFORNIA, SAN DIEGO. An Empirical Chaos Expansion Method for Uncertainty Quantification
UNIVERSITY OF CALIFORNIA, SAN DIEGO An Empirical Chaos Expansion Method for Uncertainty Quantification A Dissertation submitted in partial satisfaction of the requirements for the degree Doctor of Philosophy
More informationFast Numerical Methods for Stochastic Computations
Fast AreviewbyDongbinXiu May 16 th,2013 Outline Motivation 1 Motivation 2 3 4 5 Example: Burgers Equation Let us consider the Burger s equation: u t + uu x = νu xx, x [ 1, 1] u( 1) =1 u(1) = 1 Example:
More informationKernel principal component analysis for stochastic input model generation
Kernel principal component analysis for stochastic input model generation Xiang Ma, Nicholas Zabaras Materials Process Design and Control Laboratory, Sibley School of Mechanical and Aerospace Engineering,
More informationDynamic response of structures with uncertain properties
Dynamic response of structures with uncertain properties S. Adhikari 1 1 Chair of Aerospace Engineering, College of Engineering, Swansea University, Bay Campus, Fabian Way, Swansea, SA1 8EN, UK International
More informationEfficient Observation of Random Phenomena
Lecture 9 Efficient Observation of Random Phenomena Tokyo Polytechnic University The 1st Century Center of Excellence Program Yukio Tamura POD Proper Orthogonal Decomposition Stochastic Representation
More informationLinear Algebra. Session 12
Linear Algebra. Session 12 Dr. Marco A Roque Sol 08/01/2017 Example 12.1 Find the constant function that is the least squares fit to the following data x 0 1 2 3 f(x) 1 0 1 2 Solution c = 1 c = 0 f (x)
More informationLinear Algebra Review (Course Notes for Math 308H - Spring 2016)
Linear Algebra Review (Course Notes for Math 308H - Spring 2016) Dr. Michael S. Pilant February 12, 2016 1 Background: We begin with one of the most fundamental notions in R 2, distance. Letting (x 1,
More information5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors
EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we
More informationMLCC 2015 Dimensionality Reduction and PCA
MLCC 2015 Dimensionality Reduction and PCA Lorenzo Rosasco UNIGE-MIT-IIT June 25, 2015 Outline PCA & Reconstruction PCA and Maximum Variance PCA and Associated Eigenproblem Beyond the First Principal Component
More informationDinesh Kumar, Mehrdad Raisee and Chris Lacor
Dinesh Kumar, Mehrdad Raisee and Chris Lacor Fluid Mechanics and Thermodynamics Research Group Vrije Universiteit Brussel, BELGIUM dkumar@vub.ac.be; m_raisee@yahoo.com; chris.lacor@vub.ac.be October, 2014
More information