Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics
|
|
- Lora McDaniel
- 5 years ago
- Views:
Transcription
1 Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. URL: Random Matrix Eigenvalue Problems p.1/36
2 Outline of the Presentation Random eigenvalue problem Existing methods Exact methods Perturbation methods Asymptotic analysis of multidimensional integrals Joint moments and pdf of the natural frequencies Numerical examples & results Conclusions Random Matrix Eigenvalue Problems p.2/36
3 Random Eigenvalue Problem The random eigenvalue problem of undamped or proportionally damped linear systems: K(x)φ j = ω 2 jm(x)φ j (1) ω j natural frequencies; φ j eigenvectors; M(x) R N N mass matrix and K(x) R N N stiffness matrix. x R m is random parameter vector with pdf p x (x) = e L(x) L(x) is the log-likelihood function. Random Matrix Eigenvalue Problems p.3/36
4 The Objectives The aim is to obtain the joint probability density function of the natural frequencies and the eigenvectors in this work we look at the joint statistics of the eigenvalues while several papers are available on the distribution of individual eigenvalues, only first-order perturbation results are available for the joint pdf of the eigenvalues Random Matrix Eigenvalue Problems p.4/36
5 Exact Joint pdf Without any loss of generality the original eigenvalue problem can be expressed by where H(x)ψ j = ω 2 jψ j (2) and H(x) = M 1/2 (x)k(x)m 1/2 (x) R N N ψ j = M 1/2 φ j Random Matrix Eigenvalue Problems p.5/36
6 Exact Joint pdf The joint probability (following Muirhead, 1982) density function of the natural frequencies of an N-dimensional linear positive definite dynamic system is given by p Ω (ω 1, ω 2,, ω N ) = πn2 /2 ( ω 2 j ω 2 i ) Γ(N/2) i<j N ( p H ΨΩ 2 Ψ T) (dψ) (3) O(N) where H = M 1/2 KM 1/2 & p H (H) is the pdf of H. Random Matrix Eigenvalue Problems p.6/36
7 Limitations of the Exact Method the multidimensional integral over the orthogonal group O(N) is difficult to carry out in practice and exact closed-form results can be derived only for few special cases the derivation of an expression of the joint pdf of the system matrix p H (H) is non-trivial even if the joint pdf of the random system parameters x is known Random Matrix Eigenvalue Problems p.7/36
8 Limitations of the Exact Method even one can overcome the previous two problems, the joint pdf of the natural frequencies given by Eq. (3) is too much information to be useful for practical problems because it is not easy to visualize the joint pdf in the space of N natural frequencies, and the derivation of the marginal density functions of the natural frequencies from Eq. (3) is not straightforward, especially when N is large. Random Matrix Eigenvalue Problems p.8/36
9 Eigenvalues of GOE Matrices Suppose the system matrix H is from a Gaussian orthogonal ensemble (GOE). The pdf of H: p H (H) = exp ( θ 2 Trace ( H 2) + θ 1 Trace (H) + θ 0 ) The joint pdf of the natural frequencies: p Ω (ω 1, ω 2,, ω N ) = exp [ ( N j=1 θ 2 ω 4 j θ 1 ω 2 j θ 0 )] i<j ω 2 j ω 2 i Random Matrix Eigenvalue Problems p.9/36
10 Perturbation Method Taylor series expansion of ω j (x) about the mean x = µ ω j (x) ω j (µ) + d T ω j (µ) (x µ) (x µ)t D ωj (µ) (x µ) Here d ωj (µ) R m and D ωj (µ) R m m are respectively the gradient vector and the Hessian matrix of ω j (x) evaluated at x = µ. Random Matrix Eigenvalue Problems p.10/36
11 Joint Statistics Joint statistics of the natural frequencies can be obtained provided it is assumed that the x is Gaussian. Assuming x N(µ,Σ), first few cumulants can be obtained as κ (1,0) jk = E [ω j ] = ω j Trace( D ωj Σ ), κ (0,1) jk κ (1,1) jk = E [ω k ] = ω k Trace (D ω k Σ), = Cov (ω j, ω k ) = 1 2 Trace(( D ωj Σ ) (D ωk Σ) ) + d T ω j Σd ωk Random Matrix Eigenvalue Problems p.11/36
12 Multidimensional Integrals We want to evaluate an m-dimensional integral over the unbounded domain R m : J = R e f(x) dx m Assume f(x) is smooth and at least twice differentiable The maximum contribution to this integral comes from the neighborhood where f(x) reaches its global minimum, say θ R m Random Matrix Eigenvalue Problems p.12/36
13 Multidimensional Integrals Therefore, at x = θ f(x) x k = 0, k or d f (θ) = 0 Expand f(x) in a Taylor series about θ: { J = R e f(θ)+ 1 2(x θ) T D } f(θ)(x θ)+ε(x,θ) dx m = e f (θ) 2(x θ) T D f(θ)(x θ) ε(x,θ) dx R m e 1 Random Matrix Eigenvalue Problems p.13/36
14 Multidimensional Integrals The error ε(x, θ) depends on higher derivatives of f(x) at x = θ. If they are small compared to f (θ) their contribution will negligible to the value of the integral. So we assume that f(θ) is large so that 1 f (θ) D(j) (f (θ)) 0 for j > 2 where D (j) (f (θ)) is jth order derivative of f(x) evaluated at x = θ. Under such assumptions ε(x, θ) 0. Random Matrix Eigenvalue Problems p.14/36
15 Multidimensional Integrals Use the coordinate transformation: ξ = (x θ)d 1/2 f (θ) The Jacobian: J = D f (θ) 1/2 The integral becomes: J e f (θ) R D m f (θ) 1/2 e 1 2 ( ξ T or J (2π) m/2 e f (θ) Df (θ) 1/2 ξ ) dξ Random Matrix Eigenvalue Problems p.15/36
16 Moments of Single Eigenvalues An arbitrary rth order moment of the natural frequencies can be obtained from µ (r) j = E [ ωj(x) ] r = R ωr j(x)p x (x)dx m = R e (L(x) r lnωj(x)) dx, r = 1, 2, 3 m Previous result can be used by choosing f(x) = L(x) r ln ω j (x) Random Matrix Eigenvalue Problems p.16/36
17 Moments of Single Eigenvalues After some simplifications µ (r) j (2π) m/2 ωj(θ)e r L (θ) D L(θ) + 1 r d L(θ)d L (θ) T r ω j (θ) D ω j (θ) 1/2 r = 1, 2, 3, θ is obtained from: d ωj (θ)r = ω j (θ)d L (θ) Random Matrix Eigenvalue Problems p.17/36
18 Maximum Entropy pdf Constraints for u [0, ]: 0 0 p ωj (u)du = 1 u r p ωj (u)du = µ (r) j, r = 1, 2, 3,, n Maximizing Shannon s measure of entropy S = 0 p ωj (u) lnp ωj (u)du, the pdf of ω j is p ωj (u) = e {ρ 0 + n i=1 ρ iu i } = e ρ 0 e n i=1 ρ iu i, u 0 Random Matrix Eigenvalue Problems p.18/36
19 Maximum Entropy pdf Taking first two moments, the resulting pdf is a truncated Gaussian density function { } 1 p ωj (u) = 2πσj Φ ( ω j /σ j ) exp (u ω j) 2 2σj 2 where σ 2 j = µ(2) j ω 2 j Ensures that the probability of any natural frequencies becoming negative is zero Random Matrix Eigenvalue Problems p.19/36
20 Joint Moments of Two Eigenvalues Arbitrary r s-th order joint moment of two natural frequencies µ (rs) jl = E [ ωj(x)ω r l s (x) ] = R exp { (L(x) r ln ω j(x) s ln ω l (x))} dx, m r = 1, 2, 3 Choose f(x) = L(x) r ln ω j (x) s ln ω l (x) Random Matrix Eigenvalue Problems p.20/36
21 Joint Moments of Two Eigenvalues After some simplifications µ (rs) jl (2π) m/2 ω r j(θ)ω s l (θ) exp { L (θ)} D f (θ) 1/2 where θ is obtained from: d L (θ) = r ω j (θ) d ω j (θ) + s ω l (θ) d ω l (θ) and D f (θ) = D L (θ) + r ω 2 j(θ) d ω j (θ)d ωj (θ) T r ω j(θ) D ω j (θ) + s ω 2 l(θ) d ω l (θ)d ωl (θ) T s ω l(θ) D ω l (θ) Random Matrix Eigenvalue Problems p.21/36
22 Joint Moments of Multiple Eigenvalues We want to obtain µ (r 1r 2 r n ) { j 1 j 2 j n = ω r 1 R m j 1 (x)ω r 2 j 2 (x) ω r n j n (x) } p x (x)dx It can be shown that µ (r 1r 2 r n ) j 1 j 2 j n (2π) m/2 { ω r 1 j 1 (θ)ω r 2 j 2 (θ) ω r n j n (θ) } exp { L (θ)} D f (θ) 1/2 Random Matrix Eigenvalue Problems p.22/36
23 Joint Moments of Multiple Eigenvalues Here θ is obtained from d L (θ) = r 1 ω j1 (θ) d ω j1 (θ)+ r 2 ω j1 (θ) d ω j2 (θ)+ and the Hessian matrix is given by r n ω jn (θ) d ω j n (θ) D f (θ) = D L (θ)+ j n,r n j = j 1, j 2, r = r 1, r 2, r ω 2 j (θ)d ω j (θ)d ωj (θ) T r ω j (θ) D ω j (θ) Random Matrix Eigenvalue Problems p.23/36
24 Example System Undamped three degree-of-freedom random system: k 6 k 1 m 1 k 4 k 5 m 3 k 3 k 2 m 2 m i = 1.0 kg for i = 1, 2, 3; k i = 1.0 N/m for i = 1,, 5 and k 6 = 3.0 N/m Random Matrix Eigenvalue Problems p.24/36
25 Example System m i = m i (1 + ǫ m x i ), i = 1, 2, 3 k i = k i (1 + ǫ k x i+3 ), i = 1,, 6 Vector of random variables: x = {x 1,, x 9 } T R 9 x is standard Gaussian, µ = 0 and Σ = I Strength parameters ǫ m = 0.15 and ǫ k = 0.20 Random Matrix Eigenvalue Problems p.25/36
26 Computational Methods Following four methods are compared 1. First-order perturbation 2. Second-order perturbation 3. Asymptotic method 4. Monte Carlo Simulation (15K samples) - can be considered as benchmark. The percentage error: Error = ( ) ( ) MCS ( ) MCS 100 Random Matrix Eigenvalue Problems p.26/36
27 Scatter of the Eigenvalues ω 1 ω 2 ω Samples Eigenvalues, ω j (rad/s) Statistical scatter of the natural frequencies ω 1 = 1, ω 2 = 2, and ω 3 = 3 Random Matrix Eigenvalue Problems p.27/36
28 Error in the Mean Values st order perturbation 2nd order perturbation Asymptotic method 0.7 Percentage error wrt MCS Eigenvalue number j Error in the mean values Random Matrix Eigenvalue Problems p.28/36
29 Error in Covariance Matrix st order perturbation 2nd order perturbation Asymptotic method Percentage error wrt MCS (1,1) (1,2) (1,3) (2,2) (2,3) (3,3) Elements of the covariance matrix Error in the elements of the covariance matrix Random Matrix Eigenvalue Problems p.29/36
30 Mean and Covariance Using the asymptotic method, the mean and covariance matrix of the natural frequencies are obtained as µ Ω = {0.9962, , } T and Σ Ω = Individual pdf and joint pdf of the natural frequencies are computed using these values. Random Matrix Eigenvalue Problems p.30/36
31 Individual pdf 6 5 Asymptotic method Monte Carlo Simulation pdf of the natural frequencies Natural frequencies (rad/s) Individual pdf of the natural frequencies Random Matrix Eigenvalue Problems p.31/36
32 Analytical Joint pdf Joint pdf using asymptotic method Random Matrix Eigenvalue Problems p.32/36
33 Joint pdf from MCS Joint pdf from Monte Carlo Simulation Random Matrix Eigenvalue Problems p.33/36
34 Contours of the joint pdf 4 (ω 1,ω 3 ) 3.5 ω k (rad/s) (ω 2,ω 3 ) 2 (ω 1,ω 2 ) Asymptotic method Monte Carlo Simulation ω j (rad/s) Contours of the joint pdf Random Matrix Eigenvalue Problems p.34/36
35 Conclusions Statistics of the natural frequencies of linear stochastic dynamic systems has been considered usual assumption of small randomness is not employed in this study. a general expression of the joint pdf of the natural frequencies of linear stochastic systems has been given Random Matrix Eigenvalue Problems p.35/36
36 Conclusions a closed-form expression is obtained for the general order joint moments of the eigenvalues it was observed that the natural frequencies are not jointly Gaussian even they are so individually future studies will consider joint statistics of the eigenvalues and eigenvectors and dynamic response analysis using eigensolution distributions Random Matrix Eigenvalue Problems p.36/36
37 References Muirhead, R. J. (1982), Aspects of Multivariate Statistical Theory, John Wiely and Sons, New York, USA. 36-1
Random Eigenvalue Problems Revisited
Random Eigenvalue Problems Revisited S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. Email: S.Adhikari@bristol.ac.uk URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html
More informationRandom Eigenvalue Problems in Structural Dynamics
45th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics & Materials Confer 19-22 April 24, Palm Springs, California AIAA 24-1747 Random Eigenvalue Problems in Structural Dynamics S. Adhiakri and M.
More informationAn Efficient Computational Solution Scheme of the Random Eigenvalue Problems
50th AIAA SDM Conference, 4-7 May 2009 An Efficient Computational Solution Scheme of the Random Eigenvalue Problems Rajib Chowdhury & Sondipon Adhikari School of Engineering Swansea University Swansea,
More informationOn the Nature of Random System Matrices in Structural Dynamics
On the Nature of Random System Matrices in Structural Dynamics S. ADHIKARI AND R. S. LANGLEY Cambridge University Engineering Department Cambridge, U.K. Nature of Random System Matrices p.1/20 Outline
More informationDynamic System Identification using HDMR-Bayesian Technique
Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in
More informationDynamic Response of Structures With Frequency Dependent Damping
Dynamic Response of Structures With Frequency Dependent Damping Blanca Pascual & S Adhikari School of Engineering, Swansea University, Swansea, UK Email: S.Adhikari@swansea.ac.uk URL: http://engweb.swan.ac.uk/
More informationRandom Eigenvalue Problems in Structural Dynamics: An Experimental Investigation
Random Eigenvalue Problems in Structural Dynamics: An Experimental Investigation S. Adhikari, A. Srikantha Phani and D. A. Pape School of Engineering, Swansea University, Swansea, UK Email: S.Adhikari@swansea.ac.uk
More informationProbabilistic Structural Dynamics: Parametric vs. Nonparametric Approach
Probabilistic Structural Dynamics: Parametric vs. Nonparametric Approach S Adhikari School of Engineering, Swansea University, Swansea, UK Email: S.Adhikari@swansea.ac.uk URL: http://engweb.swan.ac.uk/
More informationThe quantum mechanics approach to uncertainty modeling in structural dynamics
p. 1/3 The quantum mechanics approach to uncertainty modeling in structural dynamics Andreas Kyprianou Department of Mechanical and Manufacturing Engineering, University of Cyprus Outline Introduction
More information11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.
C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a
More informationDynamic response of structures with uncertain properties
Dynamic response of structures with uncertain properties S. Adhikari 1 1 Chair of Aerospace Engineering, College of Engineering, Swansea University, Bay Campus, Fabian Way, Swansea, SA1 8EN, UK International
More informationRayleigh s Classical Damping Revisited
Rayleigh s Classical Damping Revisited S. Adhikari and A. Srikantha Phani Department of Aerospace Engineering, University of Bristol, Bristol, U.K. Email: S.Adhikari@bristol.ac.uk URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html
More informationEcon 508B: Lecture 5
Econ 508B: Lecture 5 Expectation, MGF and CGF Hongyi Liu Washington University in St. Louis July 31, 2017 Hongyi Liu (Washington University in St. Louis) Math Camp 2017 Stats July 31, 2017 1 / 23 Outline
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationCOMPLEX MODES IN LINEAR STOCHASTIC SYSTEMS
Proceedings of VETOMAC-I October 25-27, 2, Bangalore, INDIA COMPLEX MODES IN LINEAR STOCHASTIC SYSTEMS S. Adhikari Cambridge University Engineering Department Trumpington Street Cambridge CB2 1PZ (U.K.)
More informationGaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008
Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:
More informationEEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as
L30-1 EEL 5544 Noise in Linear Systems Lecture 30 OTHER TRANSFORMS For a continuous, nonnegative RV X, the Laplace transform of X is X (s) = E [ e sx] = 0 f X (x)e sx dx. For a nonnegative RV, the Laplace
More informationDAMPING MODELLING AND IDENTIFICATION USING GENERALIZED PROPORTIONAL DAMPING
DAMPING MODELLING AND IDENTIFICATION USING GENERALIZED PROPORTIONAL DAMPING S. Adhikari Department of Aerospace Engineering, University of Bristol, Queens Building, University Walk, Bristol BS8 1TR (U.K.)
More informationAn introduction to Birkhoff normal form
An introduction to Birkhoff normal form Dario Bambusi Dipartimento di Matematica, Universitá di Milano via Saldini 50, 0133 Milano (Italy) 19.11.14 1 Introduction The aim of this note is to present an
More informationFast Numerical Methods for Stochastic Computations
Fast AreviewbyDongbinXiu May 16 th,2013 Outline Motivation 1 Motivation 2 3 4 5 Example: Burgers Equation Let us consider the Burger s equation: u t + uu x = νu xx, x [ 1, 1] u( 1) =1 u(1) = 1 Example:
More informationExam 2. Jeremy Morris. March 23, 2006
Exam Jeremy Morris March 3, 006 4. Consider a bivariate normal population with µ 0, µ, σ, σ and ρ.5. a Write out the bivariate normal density. The multivariate normal density is defined by the following
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationMultiple Random Variables
Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An
More informationOn the entropy flows to disorder
CHAOS 2009 Charnia, Crete 1-5 June 2009 On the entropy flows to disorder C.T.J. Dodson School of Mathematics University of Manchester, UK Abstract Gamma distributions, which contain the exponential as
More informationSystem Identification, Lecture 4
System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2012 F, FRI Uppsala University, Information Technology 30 Januari 2012 SI-2012 K. Pelckmans
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationPhysics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester
Physics 403 Propagation of Uncertainties Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals
More informationSystem Identification, Lecture 4
System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2016 F, FRI Uppsala University, Information Technology 13 April 2016 SI-2016 K. Pelckmans
More informationwhere r n = dn+1 x(t)
Random Variables Overview Probability Random variables Transforms of pdfs Moments and cumulants Useful distributions Random vectors Linear transformations of random vectors The multivariate normal distribution
More informationMultivariate Random Variable
Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate
More informationOn the entropy flows to disorder
On the entropy flows to disorder School of Mathematics University of Manchester Manchester M 9PL, UK (e-mail: ctdodson@manchester.ac.uk) C.T.J. Dodson Abstract. Gamma distributions, which contain the exponential
More informationReduction of Random Variables in Structural Reliability Analysis
Reduction of Random Variables in Structural Reliability Analysis S. Adhikari and R. S. Langley Department of Engineering University of Cambridge Trumpington Street Cambridge CB2 1PZ (U.K.) February 21,
More informationGaussian Models (9/9/13)
STA561: Probabilistic machine learning Gaussian Models (9/9/13) Lecturer: Barbara Engelhardt Scribes: Xi He, Jiangwei Pan, Ali Razeen, Animesh Srivastava 1 Multivariate Normal Distribution The multivariate
More informationcomponent risk analysis
273: Urban Systems Modeling Lec. 3 component risk analysis instructor: Matteo Pozzi 273: Urban Systems Modeling Lec. 3 component reliability outline risk analysis for components uncertain demand and uncertain
More informationProbing the covariance matrix
Probing the covariance matrix Kenneth M. Hanson Los Alamos National Laboratory (ret.) BIE Users Group Meeting, September 24, 2013 This presentation available at http://kmh-lanl.hansonhub.com/ LA-UR-06-5241
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationStochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More informationCollocation based high dimensional model representation for stochastic partial differential equations
Collocation based high dimensional model representation for stochastic partial differential equations S Adhikari 1 1 Swansea University, UK ECCM 2010: IV European Conference on Computational Mechanics,
More informationDamping Modelling and Identification Using Generalized Proportional Damping
Damping Modelling and Identification Using Generalized Proportional Damping S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. Email: S.Adhikari@bristol.ac.uk Generalized
More informationVector Derivatives and the Gradient
ECE 275AB Lecture 10 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 1/1 Lecture 10 ECE 275A Vector Derivatives and the Gradient ECE 275AB Lecture 10 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego
More informationBayesian Decision Theory
Bayesian Decision Theory Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent University) 1 / 46 Bayesian
More informationPolynomial Chaos and Karhunen-Loeve Expansion
Polynomial Chaos and Karhunen-Loeve Expansion 1) Random Variables Consider a system that is modeled by R = M(x, t, X) where X is a random variable. We are interested in determining the probability of the
More informationMath 533 Extra Hour Material
Math 533 Extra Hour Material A Justification for Regression The Justification for Regression It is well-known that if we want to predict a random quantity Y using some quantity m according to a mean-squared
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationSolving the steady state diffusion equation with uncertainty Final Presentation
Solving the steady state diffusion equation with uncertainty Final Presentation Virginia Forstall vhfors@gmail.com Advisor: Howard Elman elman@cs.umd.edu Department of Computer Science May 6, 2012 Problem
More informationMultivariate Statistics
Multivariate Statistics Chapter 2: Multivariate distributions and inference Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2016/2017 Master in Mathematical
More informationSeparation of Variables in Linear PDE: One-Dimensional Problems
Separation of Variables in Linear PDE: One-Dimensional Problems Now we apply the theory of Hilbert spaces to linear differential equations with partial derivatives (PDE). We start with a particular example,
More informationInformation geometry for bivariate distribution control
Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic
More informationNotes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed
18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,
More informationElliptically Contoured Distributions
Elliptically Contoured Distributions Recall: if X N p µ, Σ), then { 1 f X x) = exp 1 } det πσ x µ) Σ 1 x µ) So f X x) depends on x only through x µ) Σ 1 x µ), and is therefore constant on the ellipsoidal
More informationarxiv: v2 [math-ph] 22 May 2009
On the entropy flows to disorder arxiv:8.438v [math-ph] May 9 C.T.J. Dodson School of Mathematics, University of Manchester, Manchester M3 9PL, UK ctdodson@manchester.ac.uk August 6, 8 Abstract Gamma distributions,
More informationUnsupervised Learning: Dimensionality Reduction
Unsupervised Learning: Dimensionality Reduction CMPSCI 689 Fall 2015 Sridhar Mahadevan Lecture 3 Outline In this lecture, we set about to solve the problem posed in the previous lecture Given a dataset,
More informationNotes on Random Vectors and Multivariate Normal
MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution
More informationAdaptive Collocation with Kernel Density Estimation
Examples of with Kernel Density Estimation Howard C. Elman Department of Computer Science University of Maryland at College Park Christopher W. Miller Applied Mathematics and Scientific Computing Program
More informationPolynomial chaos expansions for sensitivity analysis
c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty
More informationReview (Probability & Linear Algebra)
Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint
More informationMultivariate Gaussians. Sargur Srihari
Multivariate Gaussians Sargur srihari@cedar.buffalo.edu 1 Topics 1. Multivariate Gaussian: Basic Parameterization 2. Covariance and Information Form 3. Operations on Gaussians 4. Independencies in Gaussians
More informationStructural Reliability
Structural Reliability Thuong Van DANG May 28, 2018 1 / 41 2 / 41 Introduction to Structural Reliability Concept of Limit State and Reliability Review of Probability Theory First Order Second Moment Method
More information5 Operations on Multiple Random Variables
EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y
More informationSTT 843 Key to Homework 1 Spring 2018
STT 843 Key to Homework Spring 208 Due date: Feb 4, 208 42 (a Because σ = 2, σ 22 = and ρ 2 = 05, we have σ 2 = ρ 2 σ σ22 = 2/2 Then, the mean and covariance of the bivariate normal is µ = ( 0 2 and Σ
More informationMultivariate Analysis and Likelihood Inference
Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density
More informationCOMPUTATIONAL MULTI-POINT BME AND BME CONFIDENCE SETS
VII. COMPUTATIONAL MULTI-POINT BME AND BME CONFIDENCE SETS Until now we have considered the estimation of the value of a S/TRF at one point at the time, independently of the estimated values at other estimation
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More informationMaximum Likelihood Estimation. only training data is available to design a classifier
Introduction to Pattern Recognition [ Part 5 ] Mahdi Vasighi Introduction Bayesian Decision Theory shows that we could design an optimal classifier if we knew: P( i ) : priors p(x i ) : class-conditional
More informationLarge sample covariance matrices and the T 2 statistic
Large sample covariance matrices and the T 2 statistic EURANDOM, the Netherlands Joint work with W. Zhou Outline 1 2 Basic setting Let {X ij }, i, j =, be i.i.d. r.v. Write n s j = (X 1j,, X pj ) T and
More informationSubmitted to the Brazilian Journal of Probability and Statistics
Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a
More informationLecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:
tifica Lecture 1: Computation of Karhunen-Loeve Expansion: Alexander Litvinenko http://sri-uq.kaust.edu.sa/ Stochastic PDEs We consider div(κ(x, ω) u) = f (x, ω) in G, u = 0 on G, with stochastic coefficients
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More information5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors
EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we
More informationNovel Computational Methods for Solving High- Dimensional Random Eigenvalue Problems
University of Iowa Iowa Research Online Theses and Dissertations Summer 2013 Novel Computational Methods for Solving High- Dimensional Random Eigenvalue Problems Vaibhav Yadav University of Iowa Copyright
More informationIntroduction to gradient descent
6-1: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction to gradient descent Derivation and intuitions Hessian 6-2: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction Our
More informationDeterminantal Processes And The IID Gaussian Power Series
Determinantal Processes And The IID Gaussian Power Series Yuval Peres U.C. Berkeley Talk based on work joint with: J. Ben Hough Manjunath Krishnapur Bálint Virág 1 Samples of translation invariant point
More informationOn the entropy flows to disorder
On the entropy flows to disorder C.T.J. Dodson School of Mathematics, University of Manchester, Manchester M 9PL, UK ctdodson@manchester.ac.uk May, 9 Abstract Gamma distributions, which contain the exponential
More informationL2: Review of probability and statistics
Probability L2: Review of probability and statistics Definition of probability Axioms and properties Conditional probability Bayes theorem Random variables Definition of a random variable Cumulative distribution
More informationF denotes cumulative density. denotes probability density function; (.)
BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationMultiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions
Multiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions March 6, 2013 Contents 1 Wea second variation 2 1.1 Formulas for variation........................
More informationPhysics 403. Segev BenZvi. Numerical Methods, Maximum Likelihood, and Least Squares. Department of Physics and Astronomy University of Rochester
Physics 403 Numerical Methods, Maximum Likelihood, and Least Squares Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Quadratic Approximation
More informationNeural Network Training
Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification
More informationKernel Principal Component Analysis
Kernel Principal Component Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationECE 4400:693 - Information Theory
ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential
More informationFundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner
Fundamentals CS 281A: Statistical Learning Theory Yangqing Jia Based on tutorial slides by Lester Mackey and Ariel Kleiner August, 2011 Outline 1 Probability 2 Statistics 3 Linear Algebra 4 Optimization
More informationStatistical Inference and Methods
Department of Mathematics Imperial College London d.stephens@imperial.ac.uk http://stats.ma.ic.ac.uk/ das01/ 31st January 2006 Part VI Session 6: Filtering and Time to Event Data Session 6: Filtering and
More informationEstimation of Dynamic Regression Models
University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.
More informationA reduced-order stochastic finite element analysis for structures with uncertainties
A reduced-order stochastic finite element analysis for structures with uncertainties Ji Yang 1, Béatrice Faverjon 1,2, Herwig Peters 1, icole Kessissoglou 1 1 School of Mechanical and Manufacturing Engineering,
More information4. Distributions of Functions of Random Variables
4. Distributions of Functions of Random Variables Setup: Consider as given the joint distribution of X 1,..., X n (i.e. consider as given f X1,...,X n and F X1,...,X n ) Consider k functions g 1 : R n
More informationComputer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression
Group Prof. Daniel Cremers 4. Gaussian Processes - Regression Definition (Rep.) Definition: A Gaussian process is a collection of random variables, any finite number of which have a joint Gaussian distribution.
More informationσ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =
Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,
More informationStudent-t Process as Alternative to Gaussian Processes Discussion
Student-t Process as Alternative to Gaussian Processes Discussion A. Shah, A. G. Wilson and Z. Gharamani Discussion by: R. Henao Duke University June 20, 2014 Contributions The paper is concerned about
More informationSpectral methods for fuzzy structural dynamics: modal vs direct approach
Spectral methods for fuzzy structural dynamics: modal vs direct approach S Adhikari Zienkiewicz Centre for Computational Engineering, College of Engineering, Swansea University, Wales, UK IUTAM Symposium
More information3. Probability and Statistics
FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important
More informationTransformation of Probability Densities
Transformation of Probability Densities This Wikibook shows how to transform the probability density of a continuous random variable in both the one-dimensional and multidimensional case. In other words,
More informationModel Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao
Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley
More informationOn the Stability of the Best Reply Map for Noncooperative Differential Games
On the Stability of the Best Reply Map for Noncooperative Differential Games Alberto Bressan and Zipeng Wang Department of Mathematics, Penn State University, University Park, PA, 68, USA DPMMS, University
More informationIntroduction to Machine Learning
1, DATA11002 Introduction to Machine Learning Lecturer: Teemu Roos TAs: Ville Hyvönen and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer
More informationUniversity of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries
University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout :. The Multivariate Gaussian & Decision Boundaries..15.1.5 1 8 6 6 8 1 Mark Gales mjfg@eng.cam.ac.uk Lent
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationORTHOGONAL POLYNOMIAL EXPANSIONS FOR SOLVING RANDOM EIGENVALUE PROBLEMS
International Journal for Uncertainty Quantification, 1 (2: 163 187 (2011 ORTHOGONAL POLYNOMIAL EXPANSIONS FOR SOLVING RANDOM EIGENVALUE PROBLEMS Sharif Rahman & Vaibhav Yadav College of Engineering and
More informationEstimating functional uncertainty using polynomial chaos and adjoint equations
0. Estimating functional uncertainty using polynomial chaos and adjoint equations February 24, 2011 1 Florida State University, Tallahassee, Florida, Usa 2 Moscow Institute of Physics and Technology, Moscow,
More informationLecture 4: Numerical solution of ordinary differential equations
Lecture 4: Numerical solution of ordinary differential equations Department of Mathematics, ETH Zürich General explicit one-step method: Consistency; Stability; Convergence. High-order methods: Taylor
More information