Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics

Size: px
Start display at page:

Download "Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics"

Transcription

1 Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. URL: Random Matrix Eigenvalue Problems p.1/36

2 Outline of the Presentation Random eigenvalue problem Existing methods Exact methods Perturbation methods Asymptotic analysis of multidimensional integrals Joint moments and pdf of the natural frequencies Numerical examples & results Conclusions Random Matrix Eigenvalue Problems p.2/36

3 Random Eigenvalue Problem The random eigenvalue problem of undamped or proportionally damped linear systems: K(x)φ j = ω 2 jm(x)φ j (1) ω j natural frequencies; φ j eigenvectors; M(x) R N N mass matrix and K(x) R N N stiffness matrix. x R m is random parameter vector with pdf p x (x) = e L(x) L(x) is the log-likelihood function. Random Matrix Eigenvalue Problems p.3/36

4 The Objectives The aim is to obtain the joint probability density function of the natural frequencies and the eigenvectors in this work we look at the joint statistics of the eigenvalues while several papers are available on the distribution of individual eigenvalues, only first-order perturbation results are available for the joint pdf of the eigenvalues Random Matrix Eigenvalue Problems p.4/36

5 Exact Joint pdf Without any loss of generality the original eigenvalue problem can be expressed by where H(x)ψ j = ω 2 jψ j (2) and H(x) = M 1/2 (x)k(x)m 1/2 (x) R N N ψ j = M 1/2 φ j Random Matrix Eigenvalue Problems p.5/36

6 Exact Joint pdf The joint probability (following Muirhead, 1982) density function of the natural frequencies of an N-dimensional linear positive definite dynamic system is given by p Ω (ω 1, ω 2,, ω N ) = πn2 /2 ( ω 2 j ω 2 i ) Γ(N/2) i<j N ( p H ΨΩ 2 Ψ T) (dψ) (3) O(N) where H = M 1/2 KM 1/2 & p H (H) is the pdf of H. Random Matrix Eigenvalue Problems p.6/36

7 Limitations of the Exact Method the multidimensional integral over the orthogonal group O(N) is difficult to carry out in practice and exact closed-form results can be derived only for few special cases the derivation of an expression of the joint pdf of the system matrix p H (H) is non-trivial even if the joint pdf of the random system parameters x is known Random Matrix Eigenvalue Problems p.7/36

8 Limitations of the Exact Method even one can overcome the previous two problems, the joint pdf of the natural frequencies given by Eq. (3) is too much information to be useful for practical problems because it is not easy to visualize the joint pdf in the space of N natural frequencies, and the derivation of the marginal density functions of the natural frequencies from Eq. (3) is not straightforward, especially when N is large. Random Matrix Eigenvalue Problems p.8/36

9 Eigenvalues of GOE Matrices Suppose the system matrix H is from a Gaussian orthogonal ensemble (GOE). The pdf of H: p H (H) = exp ( θ 2 Trace ( H 2) + θ 1 Trace (H) + θ 0 ) The joint pdf of the natural frequencies: p Ω (ω 1, ω 2,, ω N ) = exp [ ( N j=1 θ 2 ω 4 j θ 1 ω 2 j θ 0 )] i<j ω 2 j ω 2 i Random Matrix Eigenvalue Problems p.9/36

10 Perturbation Method Taylor series expansion of ω j (x) about the mean x = µ ω j (x) ω j (µ) + d T ω j (µ) (x µ) (x µ)t D ωj (µ) (x µ) Here d ωj (µ) R m and D ωj (µ) R m m are respectively the gradient vector and the Hessian matrix of ω j (x) evaluated at x = µ. Random Matrix Eigenvalue Problems p.10/36

11 Joint Statistics Joint statistics of the natural frequencies can be obtained provided it is assumed that the x is Gaussian. Assuming x N(µ,Σ), first few cumulants can be obtained as κ (1,0) jk = E [ω j ] = ω j Trace( D ωj Σ ), κ (0,1) jk κ (1,1) jk = E [ω k ] = ω k Trace (D ω k Σ), = Cov (ω j, ω k ) = 1 2 Trace(( D ωj Σ ) (D ωk Σ) ) + d T ω j Σd ωk Random Matrix Eigenvalue Problems p.11/36

12 Multidimensional Integrals We want to evaluate an m-dimensional integral over the unbounded domain R m : J = R e f(x) dx m Assume f(x) is smooth and at least twice differentiable The maximum contribution to this integral comes from the neighborhood where f(x) reaches its global minimum, say θ R m Random Matrix Eigenvalue Problems p.12/36

13 Multidimensional Integrals Therefore, at x = θ f(x) x k = 0, k or d f (θ) = 0 Expand f(x) in a Taylor series about θ: { J = R e f(θ)+ 1 2(x θ) T D } f(θ)(x θ)+ε(x,θ) dx m = e f (θ) 2(x θ) T D f(θ)(x θ) ε(x,θ) dx R m e 1 Random Matrix Eigenvalue Problems p.13/36

14 Multidimensional Integrals The error ε(x, θ) depends on higher derivatives of f(x) at x = θ. If they are small compared to f (θ) their contribution will negligible to the value of the integral. So we assume that f(θ) is large so that 1 f (θ) D(j) (f (θ)) 0 for j > 2 where D (j) (f (θ)) is jth order derivative of f(x) evaluated at x = θ. Under such assumptions ε(x, θ) 0. Random Matrix Eigenvalue Problems p.14/36

15 Multidimensional Integrals Use the coordinate transformation: ξ = (x θ)d 1/2 f (θ) The Jacobian: J = D f (θ) 1/2 The integral becomes: J e f (θ) R D m f (θ) 1/2 e 1 2 ( ξ T or J (2π) m/2 e f (θ) Df (θ) 1/2 ξ ) dξ Random Matrix Eigenvalue Problems p.15/36

16 Moments of Single Eigenvalues An arbitrary rth order moment of the natural frequencies can be obtained from µ (r) j = E [ ωj(x) ] r = R ωr j(x)p x (x)dx m = R e (L(x) r lnωj(x)) dx, r = 1, 2, 3 m Previous result can be used by choosing f(x) = L(x) r ln ω j (x) Random Matrix Eigenvalue Problems p.16/36

17 Moments of Single Eigenvalues After some simplifications µ (r) j (2π) m/2 ωj(θ)e r L (θ) D L(θ) + 1 r d L(θ)d L (θ) T r ω j (θ) D ω j (θ) 1/2 r = 1, 2, 3, θ is obtained from: d ωj (θ)r = ω j (θ)d L (θ) Random Matrix Eigenvalue Problems p.17/36

18 Maximum Entropy pdf Constraints for u [0, ]: 0 0 p ωj (u)du = 1 u r p ωj (u)du = µ (r) j, r = 1, 2, 3,, n Maximizing Shannon s measure of entropy S = 0 p ωj (u) lnp ωj (u)du, the pdf of ω j is p ωj (u) = e {ρ 0 + n i=1 ρ iu i } = e ρ 0 e n i=1 ρ iu i, u 0 Random Matrix Eigenvalue Problems p.18/36

19 Maximum Entropy pdf Taking first two moments, the resulting pdf is a truncated Gaussian density function { } 1 p ωj (u) = 2πσj Φ ( ω j /σ j ) exp (u ω j) 2 2σj 2 where σ 2 j = µ(2) j ω 2 j Ensures that the probability of any natural frequencies becoming negative is zero Random Matrix Eigenvalue Problems p.19/36

20 Joint Moments of Two Eigenvalues Arbitrary r s-th order joint moment of two natural frequencies µ (rs) jl = E [ ωj(x)ω r l s (x) ] = R exp { (L(x) r ln ω j(x) s ln ω l (x))} dx, m r = 1, 2, 3 Choose f(x) = L(x) r ln ω j (x) s ln ω l (x) Random Matrix Eigenvalue Problems p.20/36

21 Joint Moments of Two Eigenvalues After some simplifications µ (rs) jl (2π) m/2 ω r j(θ)ω s l (θ) exp { L (θ)} D f (θ) 1/2 where θ is obtained from: d L (θ) = r ω j (θ) d ω j (θ) + s ω l (θ) d ω l (θ) and D f (θ) = D L (θ) + r ω 2 j(θ) d ω j (θ)d ωj (θ) T r ω j(θ) D ω j (θ) + s ω 2 l(θ) d ω l (θ)d ωl (θ) T s ω l(θ) D ω l (θ) Random Matrix Eigenvalue Problems p.21/36

22 Joint Moments of Multiple Eigenvalues We want to obtain µ (r 1r 2 r n ) { j 1 j 2 j n = ω r 1 R m j 1 (x)ω r 2 j 2 (x) ω r n j n (x) } p x (x)dx It can be shown that µ (r 1r 2 r n ) j 1 j 2 j n (2π) m/2 { ω r 1 j 1 (θ)ω r 2 j 2 (θ) ω r n j n (θ) } exp { L (θ)} D f (θ) 1/2 Random Matrix Eigenvalue Problems p.22/36

23 Joint Moments of Multiple Eigenvalues Here θ is obtained from d L (θ) = r 1 ω j1 (θ) d ω j1 (θ)+ r 2 ω j1 (θ) d ω j2 (θ)+ and the Hessian matrix is given by r n ω jn (θ) d ω j n (θ) D f (θ) = D L (θ)+ j n,r n j = j 1, j 2, r = r 1, r 2, r ω 2 j (θ)d ω j (θ)d ωj (θ) T r ω j (θ) D ω j (θ) Random Matrix Eigenvalue Problems p.23/36

24 Example System Undamped three degree-of-freedom random system: k 6 k 1 m 1 k 4 k 5 m 3 k 3 k 2 m 2 m i = 1.0 kg for i = 1, 2, 3; k i = 1.0 N/m for i = 1,, 5 and k 6 = 3.0 N/m Random Matrix Eigenvalue Problems p.24/36

25 Example System m i = m i (1 + ǫ m x i ), i = 1, 2, 3 k i = k i (1 + ǫ k x i+3 ), i = 1,, 6 Vector of random variables: x = {x 1,, x 9 } T R 9 x is standard Gaussian, µ = 0 and Σ = I Strength parameters ǫ m = 0.15 and ǫ k = 0.20 Random Matrix Eigenvalue Problems p.25/36

26 Computational Methods Following four methods are compared 1. First-order perturbation 2. Second-order perturbation 3. Asymptotic method 4. Monte Carlo Simulation (15K samples) - can be considered as benchmark. The percentage error: Error = ( ) ( ) MCS ( ) MCS 100 Random Matrix Eigenvalue Problems p.26/36

27 Scatter of the Eigenvalues ω 1 ω 2 ω Samples Eigenvalues, ω j (rad/s) Statistical scatter of the natural frequencies ω 1 = 1, ω 2 = 2, and ω 3 = 3 Random Matrix Eigenvalue Problems p.27/36

28 Error in the Mean Values st order perturbation 2nd order perturbation Asymptotic method 0.7 Percentage error wrt MCS Eigenvalue number j Error in the mean values Random Matrix Eigenvalue Problems p.28/36

29 Error in Covariance Matrix st order perturbation 2nd order perturbation Asymptotic method Percentage error wrt MCS (1,1) (1,2) (1,3) (2,2) (2,3) (3,3) Elements of the covariance matrix Error in the elements of the covariance matrix Random Matrix Eigenvalue Problems p.29/36

30 Mean and Covariance Using the asymptotic method, the mean and covariance matrix of the natural frequencies are obtained as µ Ω = {0.9962, , } T and Σ Ω = Individual pdf and joint pdf of the natural frequencies are computed using these values. Random Matrix Eigenvalue Problems p.30/36

31 Individual pdf 6 5 Asymptotic method Monte Carlo Simulation pdf of the natural frequencies Natural frequencies (rad/s) Individual pdf of the natural frequencies Random Matrix Eigenvalue Problems p.31/36

32 Analytical Joint pdf Joint pdf using asymptotic method Random Matrix Eigenvalue Problems p.32/36

33 Joint pdf from MCS Joint pdf from Monte Carlo Simulation Random Matrix Eigenvalue Problems p.33/36

34 Contours of the joint pdf 4 (ω 1,ω 3 ) 3.5 ω k (rad/s) (ω 2,ω 3 ) 2 (ω 1,ω 2 ) Asymptotic method Monte Carlo Simulation ω j (rad/s) Contours of the joint pdf Random Matrix Eigenvalue Problems p.34/36

35 Conclusions Statistics of the natural frequencies of linear stochastic dynamic systems has been considered usual assumption of small randomness is not employed in this study. a general expression of the joint pdf of the natural frequencies of linear stochastic systems has been given Random Matrix Eigenvalue Problems p.35/36

36 Conclusions a closed-form expression is obtained for the general order joint moments of the eigenvalues it was observed that the natural frequencies are not jointly Gaussian even they are so individually future studies will consider joint statistics of the eigenvalues and eigenvectors and dynamic response analysis using eigensolution distributions Random Matrix Eigenvalue Problems p.36/36

37 References Muirhead, R. J. (1982), Aspects of Multivariate Statistical Theory, John Wiely and Sons, New York, USA. 36-1

Random Eigenvalue Problems Revisited

Random Eigenvalue Problems Revisited Random Eigenvalue Problems Revisited S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. Email: S.Adhikari@bristol.ac.uk URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html

More information

Random Eigenvalue Problems in Structural Dynamics

Random Eigenvalue Problems in Structural Dynamics 45th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics & Materials Confer 19-22 April 24, Palm Springs, California AIAA 24-1747 Random Eigenvalue Problems in Structural Dynamics S. Adhiakri and M.

More information

An Efficient Computational Solution Scheme of the Random Eigenvalue Problems

An Efficient Computational Solution Scheme of the Random Eigenvalue Problems 50th AIAA SDM Conference, 4-7 May 2009 An Efficient Computational Solution Scheme of the Random Eigenvalue Problems Rajib Chowdhury & Sondipon Adhikari School of Engineering Swansea University Swansea,

More information

On the Nature of Random System Matrices in Structural Dynamics

On the Nature of Random System Matrices in Structural Dynamics On the Nature of Random System Matrices in Structural Dynamics S. ADHIKARI AND R. S. LANGLEY Cambridge University Engineering Department Cambridge, U.K. Nature of Random System Matrices p.1/20 Outline

More information

Dynamic System Identification using HDMR-Bayesian Technique

Dynamic System Identification using HDMR-Bayesian Technique Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in

More information

Dynamic Response of Structures With Frequency Dependent Damping

Dynamic Response of Structures With Frequency Dependent Damping Dynamic Response of Structures With Frequency Dependent Damping Blanca Pascual & S Adhikari School of Engineering, Swansea University, Swansea, UK Email: S.Adhikari@swansea.ac.uk URL: http://engweb.swan.ac.uk/

More information

Random Eigenvalue Problems in Structural Dynamics: An Experimental Investigation

Random Eigenvalue Problems in Structural Dynamics: An Experimental Investigation Random Eigenvalue Problems in Structural Dynamics: An Experimental Investigation S. Adhikari, A. Srikantha Phani and D. A. Pape School of Engineering, Swansea University, Swansea, UK Email: S.Adhikari@swansea.ac.uk

More information

Probabilistic Structural Dynamics: Parametric vs. Nonparametric Approach

Probabilistic Structural Dynamics: Parametric vs. Nonparametric Approach Probabilistic Structural Dynamics: Parametric vs. Nonparametric Approach S Adhikari School of Engineering, Swansea University, Swansea, UK Email: S.Adhikari@swansea.ac.uk URL: http://engweb.swan.ac.uk/

More information

The quantum mechanics approach to uncertainty modeling in structural dynamics

The quantum mechanics approach to uncertainty modeling in structural dynamics p. 1/3 The quantum mechanics approach to uncertainty modeling in structural dynamics Andreas Kyprianou Department of Mechanical and Manufacturing Engineering, University of Cyprus Outline Introduction

More information

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly. C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a

More information

Dynamic response of structures with uncertain properties

Dynamic response of structures with uncertain properties Dynamic response of structures with uncertain properties S. Adhikari 1 1 Chair of Aerospace Engineering, College of Engineering, Swansea University, Bay Campus, Fabian Way, Swansea, SA1 8EN, UK International

More information

Rayleigh s Classical Damping Revisited

Rayleigh s Classical Damping Revisited Rayleigh s Classical Damping Revisited S. Adhikari and A. Srikantha Phani Department of Aerospace Engineering, University of Bristol, Bristol, U.K. Email: S.Adhikari@bristol.ac.uk URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html

More information

Econ 508B: Lecture 5

Econ 508B: Lecture 5 Econ 508B: Lecture 5 Expectation, MGF and CGF Hongyi Liu Washington University in St. Louis July 31, 2017 Hongyi Liu (Washington University in St. Louis) Math Camp 2017 Stats July 31, 2017 1 / 23 Outline

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

COMPLEX MODES IN LINEAR STOCHASTIC SYSTEMS

COMPLEX MODES IN LINEAR STOCHASTIC SYSTEMS Proceedings of VETOMAC-I October 25-27, 2, Bangalore, INDIA COMPLEX MODES IN LINEAR STOCHASTIC SYSTEMS S. Adhikari Cambridge University Engineering Department Trumpington Street Cambridge CB2 1PZ (U.K.)

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

EEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as

EEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as L30-1 EEL 5544 Noise in Linear Systems Lecture 30 OTHER TRANSFORMS For a continuous, nonnegative RV X, the Laplace transform of X is X (s) = E [ e sx] = 0 f X (x)e sx dx. For a nonnegative RV, the Laplace

More information

DAMPING MODELLING AND IDENTIFICATION USING GENERALIZED PROPORTIONAL DAMPING

DAMPING MODELLING AND IDENTIFICATION USING GENERALIZED PROPORTIONAL DAMPING DAMPING MODELLING AND IDENTIFICATION USING GENERALIZED PROPORTIONAL DAMPING S. Adhikari Department of Aerospace Engineering, University of Bristol, Queens Building, University Walk, Bristol BS8 1TR (U.K.)

More information

An introduction to Birkhoff normal form

An introduction to Birkhoff normal form An introduction to Birkhoff normal form Dario Bambusi Dipartimento di Matematica, Universitá di Milano via Saldini 50, 0133 Milano (Italy) 19.11.14 1 Introduction The aim of this note is to present an

More information

Fast Numerical Methods for Stochastic Computations

Fast Numerical Methods for Stochastic Computations Fast AreviewbyDongbinXiu May 16 th,2013 Outline Motivation 1 Motivation 2 3 4 5 Example: Burgers Equation Let us consider the Burger s equation: u t + uu x = νu xx, x [ 1, 1] u( 1) =1 u(1) = 1 Example:

More information

Exam 2. Jeremy Morris. March 23, 2006

Exam 2. Jeremy Morris. March 23, 2006 Exam Jeremy Morris March 3, 006 4. Consider a bivariate normal population with µ 0, µ, σ, σ and ρ.5. a Write out the bivariate normal density. The multivariate normal density is defined by the following

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An

More information

On the entropy flows to disorder

On the entropy flows to disorder CHAOS 2009 Charnia, Crete 1-5 June 2009 On the entropy flows to disorder C.T.J. Dodson School of Mathematics University of Manchester, UK Abstract Gamma distributions, which contain the exponential as

More information

System Identification, Lecture 4

System Identification, Lecture 4 System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2012 F, FRI Uppsala University, Information Technology 30 Januari 2012 SI-2012 K. Pelckmans

More information

Formulas for probability theory and linear models SF2941

Formulas for probability theory and linear models SF2941 Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms

More information

Physics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester Physics 403 Propagation of Uncertainties Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals

More information

System Identification, Lecture 4

System Identification, Lecture 4 System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2016 F, FRI Uppsala University, Information Technology 13 April 2016 SI-2016 K. Pelckmans

More information

where r n = dn+1 x(t)

where r n = dn+1 x(t) Random Variables Overview Probability Random variables Transforms of pdfs Moments and cumulants Useful distributions Random vectors Linear transformations of random vectors The multivariate normal distribution

More information

Multivariate Random Variable

Multivariate Random Variable Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate

More information

On the entropy flows to disorder

On the entropy flows to disorder On the entropy flows to disorder School of Mathematics University of Manchester Manchester M 9PL, UK (e-mail: ctdodson@manchester.ac.uk) C.T.J. Dodson Abstract. Gamma distributions, which contain the exponential

More information

Reduction of Random Variables in Structural Reliability Analysis

Reduction of Random Variables in Structural Reliability Analysis Reduction of Random Variables in Structural Reliability Analysis S. Adhikari and R. S. Langley Department of Engineering University of Cambridge Trumpington Street Cambridge CB2 1PZ (U.K.) February 21,

More information

Gaussian Models (9/9/13)

Gaussian Models (9/9/13) STA561: Probabilistic machine learning Gaussian Models (9/9/13) Lecturer: Barbara Engelhardt Scribes: Xi He, Jiangwei Pan, Ali Razeen, Animesh Srivastava 1 Multivariate Normal Distribution The multivariate

More information

component risk analysis

component risk analysis 273: Urban Systems Modeling Lec. 3 component risk analysis instructor: Matteo Pozzi 273: Urban Systems Modeling Lec. 3 component reliability outline risk analysis for components uncertain demand and uncertain

More information

Probing the covariance matrix

Probing the covariance matrix Probing the covariance matrix Kenneth M. Hanson Los Alamos National Laboratory (ret.) BIE Users Group Meeting, September 24, 2013 This presentation available at http://kmh-lanl.hansonhub.com/ LA-UR-06-5241

More information

Continuous Random Variables

Continuous Random Variables 1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables

More information

Stochastic Spectral Approaches to Bayesian Inference

Stochastic Spectral Approaches to Bayesian Inference Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to

More information

Collocation based high dimensional model representation for stochastic partial differential equations

Collocation based high dimensional model representation for stochastic partial differential equations Collocation based high dimensional model representation for stochastic partial differential equations S Adhikari 1 1 Swansea University, UK ECCM 2010: IV European Conference on Computational Mechanics,

More information

Damping Modelling and Identification Using Generalized Proportional Damping

Damping Modelling and Identification Using Generalized Proportional Damping Damping Modelling and Identification Using Generalized Proportional Damping S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. Email: S.Adhikari@bristol.ac.uk Generalized

More information

Vector Derivatives and the Gradient

Vector Derivatives and the Gradient ECE 275AB Lecture 10 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego p. 1/1 Lecture 10 ECE 275A Vector Derivatives and the Gradient ECE 275AB Lecture 10 Fall 2008 V1.1 c K. Kreutz-Delgado, UC San Diego

More information

Bayesian Decision Theory

Bayesian Decision Theory Bayesian Decision Theory Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent University) 1 / 46 Bayesian

More information

Polynomial Chaos and Karhunen-Loeve Expansion

Polynomial Chaos and Karhunen-Loeve Expansion Polynomial Chaos and Karhunen-Loeve Expansion 1) Random Variables Consider a system that is modeled by R = M(x, t, X) where X is a random variable. We are interested in determining the probability of the

More information

Math 533 Extra Hour Material

Math 533 Extra Hour Material Math 533 Extra Hour Material A Justification for Regression The Justification for Regression It is well-known that if we want to predict a random quantity Y using some quantity m according to a mean-squared

More information

Lecture 2: Repetition of probability theory and statistics

Lecture 2: Repetition of probability theory and statistics Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:

More information

Solving the steady state diffusion equation with uncertainty Final Presentation

Solving the steady state diffusion equation with uncertainty Final Presentation Solving the steady state diffusion equation with uncertainty Final Presentation Virginia Forstall vhfors@gmail.com Advisor: Howard Elman elman@cs.umd.edu Department of Computer Science May 6, 2012 Problem

More information

Multivariate Statistics

Multivariate Statistics Multivariate Statistics Chapter 2: Multivariate distributions and inference Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2016/2017 Master in Mathematical

More information

Separation of Variables in Linear PDE: One-Dimensional Problems

Separation of Variables in Linear PDE: One-Dimensional Problems Separation of Variables in Linear PDE: One-Dimensional Problems Now we apply the theory of Hilbert spaces to linear differential equations with partial derivatives (PDE). We start with a particular example,

More information

Information geometry for bivariate distribution control

Information geometry for bivariate distribution control Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

Elliptically Contoured Distributions

Elliptically Contoured Distributions Elliptically Contoured Distributions Recall: if X N p µ, Σ), then { 1 f X x) = exp 1 } det πσ x µ) Σ 1 x µ) So f X x) depends on x only through x µ) Σ 1 x µ), and is therefore constant on the ellipsoidal

More information

arxiv: v2 [math-ph] 22 May 2009

arxiv: v2 [math-ph] 22 May 2009 On the entropy flows to disorder arxiv:8.438v [math-ph] May 9 C.T.J. Dodson School of Mathematics, University of Manchester, Manchester M3 9PL, UK ctdodson@manchester.ac.uk August 6, 8 Abstract Gamma distributions,

More information

Unsupervised Learning: Dimensionality Reduction

Unsupervised Learning: Dimensionality Reduction Unsupervised Learning: Dimensionality Reduction CMPSCI 689 Fall 2015 Sridhar Mahadevan Lecture 3 Outline In this lecture, we set about to solve the problem posed in the previous lecture Given a dataset,

More information

Notes on Random Vectors and Multivariate Normal

Notes on Random Vectors and Multivariate Normal MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution

More information

Adaptive Collocation with Kernel Density Estimation

Adaptive Collocation with Kernel Density Estimation Examples of with Kernel Density Estimation Howard C. Elman Department of Computer Science University of Maryland at College Park Christopher W. Miller Applied Mathematics and Scientific Computing Program

More information

Polynomial chaos expansions for sensitivity analysis

Polynomial chaos expansions for sensitivity analysis c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty

More information

Review (Probability & Linear Algebra)

Review (Probability & Linear Algebra) Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint

More information

Multivariate Gaussians. Sargur Srihari

Multivariate Gaussians. Sargur Srihari Multivariate Gaussians Sargur srihari@cedar.buffalo.edu 1 Topics 1. Multivariate Gaussian: Basic Parameterization 2. Covariance and Information Form 3. Operations on Gaussians 4. Independencies in Gaussians

More information

Structural Reliability

Structural Reliability Structural Reliability Thuong Van DANG May 28, 2018 1 / 41 2 / 41 Introduction to Structural Reliability Concept of Limit State and Reliability Review of Probability Theory First Order Second Moment Method

More information

5 Operations on Multiple Random Variables

5 Operations on Multiple Random Variables EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y

More information

STT 843 Key to Homework 1 Spring 2018

STT 843 Key to Homework 1 Spring 2018 STT 843 Key to Homework Spring 208 Due date: Feb 4, 208 42 (a Because σ = 2, σ 22 = and ρ 2 = 05, we have σ 2 = ρ 2 σ σ22 = 2/2 Then, the mean and covariance of the bivariate normal is µ = ( 0 2 and Σ

More information

Multivariate Analysis and Likelihood Inference

Multivariate Analysis and Likelihood Inference Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density

More information

COMPUTATIONAL MULTI-POINT BME AND BME CONFIDENCE SETS

COMPUTATIONAL MULTI-POINT BME AND BME CONFIDENCE SETS VII. COMPUTATIONAL MULTI-POINT BME AND BME CONFIDENCE SETS Until now we have considered the estimation of the value of a S/TRF at one point at the time, independently of the estimated values at other estimation

More information

A Probability Review

A Probability Review A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in

More information

Maximum Likelihood Estimation. only training data is available to design a classifier

Maximum Likelihood Estimation. only training data is available to design a classifier Introduction to Pattern Recognition [ Part 5 ] Mahdi Vasighi Introduction Bayesian Decision Theory shows that we could design an optimal classifier if we knew: P( i ) : priors p(x i ) : class-conditional

More information

Large sample covariance matrices and the T 2 statistic

Large sample covariance matrices and the T 2 statistic Large sample covariance matrices and the T 2 statistic EURANDOM, the Netherlands Joint work with W. Zhou Outline 1 2 Basic setting Let {X ij }, i, j =, be i.i.d. r.v. Write n s j = (X 1j,, X pj ) T and

More information

Submitted to the Brazilian Journal of Probability and Statistics

Submitted to the Brazilian Journal of Probability and Statistics Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a

More information

Lecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:

Lecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion: tifica Lecture 1: Computation of Karhunen-Loeve Expansion: Alexander Litvinenko http://sri-uq.kaust.edu.sa/ Stochastic PDEs We consider div(κ(x, ω) u) = f (x, ω) in G, u = 0 on G, with stochastic coefficients

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we

More information

Novel Computational Methods for Solving High- Dimensional Random Eigenvalue Problems

Novel Computational Methods for Solving High- Dimensional Random Eigenvalue Problems University of Iowa Iowa Research Online Theses and Dissertations Summer 2013 Novel Computational Methods for Solving High- Dimensional Random Eigenvalue Problems Vaibhav Yadav University of Iowa Copyright

More information

Introduction to gradient descent

Introduction to gradient descent 6-1: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction to gradient descent Derivation and intuitions Hessian 6-2: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction Our

More information

Determinantal Processes And The IID Gaussian Power Series

Determinantal Processes And The IID Gaussian Power Series Determinantal Processes And The IID Gaussian Power Series Yuval Peres U.C. Berkeley Talk based on work joint with: J. Ben Hough Manjunath Krishnapur Bálint Virág 1 Samples of translation invariant point

More information

On the entropy flows to disorder

On the entropy flows to disorder On the entropy flows to disorder C.T.J. Dodson School of Mathematics, University of Manchester, Manchester M 9PL, UK ctdodson@manchester.ac.uk May, 9 Abstract Gamma distributions, which contain the exponential

More information

L2: Review of probability and statistics

L2: Review of probability and statistics Probability L2: Review of probability and statistics Definition of probability Axioms and properties Conditional probability Bayes theorem Random variables Definition of a random variable Cumulative distribution

More information

F denotes cumulative density. denotes probability density function; (.)

F denotes cumulative density. denotes probability density function; (.) BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

Multiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions

Multiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions Multiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions March 6, 2013 Contents 1 Wea second variation 2 1.1 Formulas for variation........................

More information

Physics 403. Segev BenZvi. Numerical Methods, Maximum Likelihood, and Least Squares. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Numerical Methods, Maximum Likelihood, and Least Squares. Department of Physics and Astronomy University of Rochester Physics 403 Numerical Methods, Maximum Likelihood, and Least Squares Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Quadratic Approximation

More information

Neural Network Training

Neural Network Training Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification

More information

Kernel Principal Component Analysis

Kernel Principal Component Analysis Kernel Principal Component Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

ECE 4400:693 - Information Theory

ECE 4400:693 - Information Theory ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential

More information

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner Fundamentals CS 281A: Statistical Learning Theory Yangqing Jia Based on tutorial slides by Lester Mackey and Ariel Kleiner August, 2011 Outline 1 Probability 2 Statistics 3 Linear Algebra 4 Optimization

More information

Statistical Inference and Methods

Statistical Inference and Methods Department of Mathematics Imperial College London d.stephens@imperial.ac.uk http://stats.ma.ic.ac.uk/ das01/ 31st January 2006 Part VI Session 6: Filtering and Time to Event Data Session 6: Filtering and

More information

Estimation of Dynamic Regression Models

Estimation of Dynamic Regression Models University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.

More information

A reduced-order stochastic finite element analysis for structures with uncertainties

A reduced-order stochastic finite element analysis for structures with uncertainties A reduced-order stochastic finite element analysis for structures with uncertainties Ji Yang 1, Béatrice Faverjon 1,2, Herwig Peters 1, icole Kessissoglou 1 1 School of Mechanical and Manufacturing Engineering,

More information

4. Distributions of Functions of Random Variables

4. Distributions of Functions of Random Variables 4. Distributions of Functions of Random Variables Setup: Consider as given the joint distribution of X 1,..., X n (i.e. consider as given f X1,...,X n and F X1,...,X n ) Consider k functions g 1 : R n

More information

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression Group Prof. Daniel Cremers 4. Gaussian Processes - Regression Definition (Rep.) Definition: A Gaussian process is a collection of random variables, any finite number of which have a joint Gaussian distribution.

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

Student-t Process as Alternative to Gaussian Processes Discussion

Student-t Process as Alternative to Gaussian Processes Discussion Student-t Process as Alternative to Gaussian Processes Discussion A. Shah, A. G. Wilson and Z. Gharamani Discussion by: R. Henao Duke University June 20, 2014 Contributions The paper is concerned about

More information

Spectral methods for fuzzy structural dynamics: modal vs direct approach

Spectral methods for fuzzy structural dynamics: modal vs direct approach Spectral methods for fuzzy structural dynamics: modal vs direct approach S Adhikari Zienkiewicz Centre for Computational Engineering, College of Engineering, Swansea University, Wales, UK IUTAM Symposium

More information

3. Probability and Statistics

3. Probability and Statistics FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important

More information

Transformation of Probability Densities

Transformation of Probability Densities Transformation of Probability Densities This Wikibook shows how to transform the probability density of a continuous random variable in both the one-dimensional and multidimensional case. In other words,

More information

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley

More information

On the Stability of the Best Reply Map for Noncooperative Differential Games

On the Stability of the Best Reply Map for Noncooperative Differential Games On the Stability of the Best Reply Map for Noncooperative Differential Games Alberto Bressan and Zipeng Wang Department of Mathematics, Penn State University, University Park, PA, 68, USA DPMMS, University

More information

Introduction to Machine Learning

Introduction to Machine Learning 1, DATA11002 Introduction to Machine Learning Lecturer: Teemu Roos TAs: Ville Hyvönen and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer

More information

University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries

University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout :. The Multivariate Gaussian & Decision Boundaries..15.1.5 1 8 6 6 8 1 Mark Gales mjfg@eng.cam.ac.uk Lent

More information

Quick Tour of Basic Probability Theory and Linear Algebra

Quick Tour of Basic Probability Theory and Linear Algebra Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions

More information

ORTHOGONAL POLYNOMIAL EXPANSIONS FOR SOLVING RANDOM EIGENVALUE PROBLEMS

ORTHOGONAL POLYNOMIAL EXPANSIONS FOR SOLVING RANDOM EIGENVALUE PROBLEMS International Journal for Uncertainty Quantification, 1 (2: 163 187 (2011 ORTHOGONAL POLYNOMIAL EXPANSIONS FOR SOLVING RANDOM EIGENVALUE PROBLEMS Sharif Rahman & Vaibhav Yadav College of Engineering and

More information

Estimating functional uncertainty using polynomial chaos and adjoint equations

Estimating functional uncertainty using polynomial chaos and adjoint equations 0. Estimating functional uncertainty using polynomial chaos and adjoint equations February 24, 2011 1 Florida State University, Tallahassee, Florida, Usa 2 Moscow Institute of Physics and Technology, Moscow,

More information

Lecture 4: Numerical solution of ordinary differential equations

Lecture 4: Numerical solution of ordinary differential equations Lecture 4: Numerical solution of ordinary differential equations Department of Mathematics, ETH Zürich General explicit one-step method: Consistency; Stability; Convergence. High-order methods: Taylor

More information