On the Relationship Between Concentration and Inertia Hyperellipsoids

Size: px
Start display at page:

Download "On the Relationship Between Concentration and Inertia Hyperellipsoids"

Transcription

1 Applied Mathematical Sciences, Vol. 2, 2008, no. 10, On the Relationship Between Concentration and Inertia Hyperellipsoids Brian J. M c Cartin Applied Mathematics, Kettering University 1700 West Third Avenue, Flint, MI USA bmccarti@kettering.edu Abstract We extend the known intricate connections between the concentration ellipsoid of statistics [1, 2] and the inertia ellipsoid of mechanics [3, 4] from three to n dimensions. We then also extend the relationship involving the eccentricities of their principal cross-sections [5] to n dimensions. The shared degeneracies of these hyperellipsoids is noted. Mathematics Subject Classifications: 51N20, 62J05, 70F99 Keywords: concentration hyperellipsoid; correlation hyperellipsoid; hyperellipsoid of residuals; inertia hyperellipsoid; momental hyperellipsoid; hyperellipsoid of gyration; orthogonal regression; least squares approximation 1 Introduction Mathematical statistics and classical mechanics both employ certain hyperellipsoids in their analytical investigations. These hyperellipsoids are distinct in n>3 dimensions while the corresponding two dimensional ellipses essentially coincide. In what follows, we first develop the concentration hyperellipsoid of statistics [1, 2] and the inertia hperellipsoid of mechanics [3, 4]. The known connections between these two hyperellipsoids in three dimensions are then extended to higher dimensions. Next, a relationship concerning the eccentricities of their principal cross-sections [5] is also extended from three to higher dimensions. Finally, it is observed that the previously derived relationships imply that both hyperellipsoids degenerate to hyperspheroids under identical conditions.

2 490 B. J. McCartin 2 Concentration Hyperellipsoid Consider the set of m data points in R n with coordinates {(x (i) 1,...,x (i) n )} m with centroid ( x 1,..., x n ) defined by the mean values x j := 1 m m Furthermore, let us introduce the variances x (i) j ;(j =1,...,n). (1) σj 2 := 1 m m (x (i) j x j ) 2 ;(j =1,...,n), (2) and the covariances p jk := 1 m m (x (i) j x j ) (x (i) k x k); (j =1,...,n; k =1,...,n; j k). (3) Defining the symmetric covariance matrix as σ 2 1 p 12 p 1n p 12 σ 2 C := 2 p 2n....., (4). p 1n p 2n σn 2 observe that v := [a 1 a 2 a n ] T implies that v T C v = E{[a 1 (x 1 x 1 )+a 2 (x 2 x 2 )+ + a n (x n x n )] 2 } 0, (5) where E is the averaging operator E{u} = 1 m m u (i), with equality if and only if the data all lie on the hyperplane of dimension n 1 a 1 (x 1 x 1 )+a 2 (x 2 x 2 )+ + a n (x n x n )=0. (6) Thus, hyperplanarity of the data corresponds to a zero eigenvalue of the positive semidefinite matrix C and the corresponding eigenvector is normal to this hyperplane. Of course, the data may further degenerate to a hyperplane of even lower dimension. Thus, without loss of generality, we will assume that C is positive definite in the ensuing analysis. The concentration hyperellipsoid (also known as a correlation hyperellipsoid) is defined by ( r r) T C 1 ( r r) =n +2, (7) which has the same first and second moments about the centroid as does the data [2, p. 300]. In this sense, it is the hyperellipsoid which is most representative of the data points without any a priori statistical assumptions

3 Concentration and inertia hyperellipsoids 491 concerning their origin. The reciprocal of the concentration hyperellipsoid, obtained by replacement of C 1 by C in the quadratic form of Equation (7), will be referred to as the hyperellipsoid of residuals [1]. Denoting the eigenvalues of C 1 by ν 1 <ν 2 < < ν n, the principal semiaxes of the concentration hyperellipsoid pass through its centroid and lie in the directions of the the corresponding eigenvectors, v j (j =1,...,n), while their lengths are given by 1/ ν j [6, p. 309]. The hyperplanar cross-sections of dimension n p passing through the centroid and perpendicular to p of these semiaxes are the corresponding principal sections of dimension n p of the concentration ellipsoid. Note that these principal sections are hyperellipsoids of dimension n p with principal semiaxes specified by the remaining n p eigenpairs. In particular, for p = n 2, these n(n 1) principal sections are ellipses and will be referred 2 to simply as the principal sections of the concentration hyperellipsoid. For p =1,...,n, the hyperplane of dimension n p which minimizes the mean square distance of the points from the hyperplane is called the orthogonal regression hyperplane of dimension n p and may be thought of as the hyperplane of dimension n p which best fits the data. The following important facts were first established by Pearson [1] in three dimensions and have subsequently been extended to n dimensions to form the basis for so-called orthogonal regression. The sum of the p th smallest eigenvalues of C equals the minimum mean square deviation from the orthogonal regression hyperplane of dimension n p which passes through the centroid and is normal to the corresponding p eigenvectors. These orthogonal regression hyperplanes are nested in that, for p =1,...,n 1, the one of dimension n p contains that of dimension n p 1. 3 Inertia Hyperellipsoid We next focus our attention on orthogonal linear regression which selects (α 1,α 2,...,α n ), the direction angles of a line (i.e., a hyperplane of dimension one), L, passing through the centroid, so as to minimize m n 1 r 2 (α 1,α 2,...,α n ):= { [cos α j (x (i) k x k) cos α k (x (i) j x j )] 2 }, (8) j=1 k=j+1 which is the total square distance of the data from L [7, p. 20]. Alternatively, r 2 (α 1,α 2,...,α n ) may be viewed as the moment of inertia with respect to L of unit masses located at the data points [3, p. 66].

4 492 B. J. McCartin In any event, the mean square deviation is given by n 1 1 m r2 (α 1,α 2,...,α n )= j=1 k=j+1 cos 2 α j σ 2 k 2 cos α j cos α k p jk + cos 2 α k σ 2 j. (9) Now, along L, mark off the two points at a distance l(α 1,α 2,...,α n ):= 1 r from the centroid. As α 2 (α 1,α 2,...,α n) 1,α 2,...,α n vary, these points sweep out a hyperellipsoid centered at the centroid. We see this as follows. Since cos α j =(x j x j )/l, 1 l 2 = r2 (α 1,α 2,...,α n )= j=1 jj (x j x j ) 2 n 1 I +2 l 2 j=1 k=j+1 I jk (x j x j )(x k x k ) l 2, (10) where I jj := m k j σ 2 k and I jk := m p jk are, respectively, the moments of inertia and the products of inertia about the coordinate axes [3, p. 67]. Equation (10) may be rewritten in matrix form as ( r r) T I( r r) =1, (11) where r =(x 1,x 2,...,x n ) T, r =( x 1, x 2,..., x n ) T, and the inertia matrix I is defined as I 11 I 12 I 1n I 12 I 22 I 2n I := (12). I 1n I 2n I nn That this n-dimensional quadric surface is a hyperellipsoid follows from the relation I = m [tr(c)i C], (13) so that the positive definiteness of I follows directly from that of C. This hyperellipsoid with l 2 (α 1,α 2,...,α n ) = 1/(moment of inertia) is called the inertia hyperellipsoid (also known as the momental hyperellipsoid) because of its role in rotational mechanics [3, p. 66]. Replacement of I by I 1 in the quadratic form of Equation (11) yields the reciprocal hyperellipsoid of gyration [4, p. 233]. The minimum value of r 2 (α 1,α 2,...,α n ) is achieved precisely when l 2 (α 1,α 2,...,α n ) assumes its maximum value which will, of course, correspond to the semimajor axis of the inertia hyperellipsoid (i.e., the largest of the principal semiaxes). Therefore, the minimum total square deviation is given by s 2 := r 2 (α 1,α 2,...,α n)=1/(semimajor axis) 2 = λ min, (14)

5 Concentration and inertia hyperellipsoids 493 where (α 1,α 2,...,α n ) are the direction angles of the semimajor axis and λ min is the smallest eigenvalue of I whose corresponding eigenvector points along the semimajor axis of the inertia hyperellipsoid. Alternatively, the orthogonal regression line is parallel to the eigenvector corresponding to the largest eigenvalue of the hyperellipsoid of gyration. In stark contrast to the two-dimensional scenario [5], the inertia hyperellipsoid is not homothetic (i.e., similar ) to the concentration hyperellipsoid for n>2. However, as we now will see, they are indeed closely related. 4 Hyperellipsoidal Relationships Equation (13) indicates that there is an intimate relationship between the inertia and concentration hyperellipsoids, a state of affairs which we now explore. In this investigation, we will denote the eigenvalues of I as λ 1 <λ 2 < <λ n, the eigenvalues of C as μ 1 <μ 2 < <μ n, and the eigenvalues of C 1 as ν 1 <ν 2 < <ν n. The degenerate case of multiple eigenvalues will be subsequently studied. In consequence of Equation (13) we have for i =1,...,n the relations λ i = m (μ 1 + μ μ n μ n i+1 ) with same eigenvector as μ n i+1. (15) Inverting these expressions we arrive at the complementary relations [4, p. 230] μ i = λ 1 + λ λ n (n 1)λ n i+1. (16) m(n 1) Observe that the positive definiteness of C implies the important inertia matrix inequality λ n < (λ 1 + λ λ n 1 )/(n 2) [4, p. 231]. Furthermore, Equations (15) and (16) together imply the desired relationship between the spectra of the inertia and concentration hyperellipsoids: ν i = 1 m(n 1) = with same eigenvector as λ i. μ n i+1 λ 1 + λ λ n (n 1)λ i (17) Thus, the corresponding principal semiaxes of the inertia and concentration hyperellipsoids are parallel. In addition, Equations (15) through (17) together imply the following collection of inequalities: Theorem 1 (Eccentricity Inequalities) For n>2, ν j ν k < λ j λ k (j =1,...,n 1; k = j +1,...,n). (18)

6 494 B. J. McCartin Proof: Since μ i (i =1,...,n) are all positive and n>2, we have Now, since we have μ n j+1 + μ n k+1 < μ i. j<k μ n k+1 <μ n j+1, (μ n j+1 + μ n k+1 ) (μ n j+1 μ n k+1 ) < μ i (μ n j+1 μ n k+1 ). This may be rearranged to produce μ n k+1 ( μ i μ n j+1 μ n k+1 ) <μ n j+1 ( μ i μ n k+1 μ n j+1 ). Adding μ n j+1 μ n k+1 to both sides yields μ n k+1 ( μ i μ n k+1 ) <μ n j+1 ( μ i μ n j+1 ), which implies that μ n n k+1 μ i μ n j+1 < μ n. n j+1 μ i μ n k+1 Finally, this may be recast as ν j < λ j. ν k λ k Note that if n = 2 then all of the above inequalities become equalities. Thus, not only are the principal axes of the inertia and concentration hyperellipsoids parallel, but each principal section of the inertia hyperellipsoid is less eccentric than the corresponding principal section of the concentration hyperellipsoid. This hyperellipsoidal relationship is a natural generalization of the corresponding three dimensional result first presented in [5]. 5 Conclusion We conclude with the straightforward observation that the inertia and concentration hyperellipsoids share degeneracies to hypershperoids, i.e. multiple eigenvalues, μ j 1 <μ j = = μ k <μ k+1, (19) in which case the orthogonal regression planes of dimension n j through n k + 1 will be indeterminate. An auxiliary consequence of the preceding

7 Concentration and inertia hyperellipsoids 495 considerations is that the concentration hyperellipsoid also shares degeneracies with the hyperellipsoid of residuals except that now, due to their reciprocal relationship, the inequalities of Equation (19) become ν n k <ν n k+1 = = ν n j+1 <ν n j+2. (20) The characteristic polynomial of C is called the discriminating polynomial of degree n and multiplicity of its roots is equivalent to degeneracy of these hyperellipsoids to hyperspheroids. Unfortunately, necessary and sufficient conditions for multiple roots of the discriminating polynomial are not readily available for n>3. Acknowledgement. This paper is dedicated to my loving wife, Barbara A. McCartin, for her support of all of my mathematical endeavors. References [1] K. Pearson, On Lines and Planes of Closest Fit to Systems of Points in Space, Phil. Mag. 2 (1901), [2] H. Cramér, Mathematical Methods of Statistics, Princeton University Press, Princeton, NJ, [3] H. Lamb, Higher Mechanics, Second Edition, Cambridge University Press, Cambridge, [4] A. G. Webster, Dynamics of Particles and of Rigid, Elastic, and Fluid Bodies, Third Edition, Stechert, New York, [5] B. J. McCartin, On Concentration and Inertia Ellipsoids, Applied Mathematical Sciences 1(1) (2007), [6] F. R. Gantmacher, Matrix Theory: Volume I, Chelsea, New York, [7] M. G. Kendall, A Course in the Geometry of n Dimensions, Hafner, New York, NY, Received: August 26, 2007

Geometric Constructions Relating Various Lines of Regression

Geometric Constructions Relating Various Lines of Regression Applied Mathematical Sciences, Vol. 0, 06, no. 3, - 30 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/0.988/ams.06.5735 Geometric Constructions Relating Various Lines of Regression Brian J. M c Cartin

More information

A Matrix-Vector Analytic Demonstration of Pappus Construction of an Ellipse from a Pair of Conjugate Diameters

A Matrix-Vector Analytic Demonstration of Pappus Construction of an Ellipse from a Pair of Conjugate Diameters Applied Mathematical Sciences, Vol. 9, 015, no. 14, 679-687 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.1988/ams.015.411035 A Matrix-Vector Analytic Demonstration of Pappus Construction of an Ellipse

More information

A A x i x j i j (i, j) (j, i) Let. Compute the value of for and

A A x i x j i j (i, j) (j, i) Let. Compute the value of for and 7.2 - Quadratic Forms quadratic form on is a function defined on whose value at a vector in can be computed by an expression of the form, where is an symmetric matrix. The matrix R n Q R n x R n Q(x) =

More information

Regression. Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables X and Y.

Regression. Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables X and Y. Regression Bivariate i linear regression: Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables and. Generally describe as a

More information

ECE 275A Homework 6 Solutions

ECE 275A Homework 6 Solutions ECE 275A Homework 6 Solutions. The notation used in the solutions for the concentration (hyper) ellipsoid problems is defined in the lecture supplement on concentration ellipsoids. Note that θ T Σ θ =

More information

Linear Algebra And Its Applications Chapter 6. Positive Definite Matrix

Linear Algebra And Its Applications Chapter 6. Positive Definite Matrix Linear Algebra And Its Applications Chapter 6. Positive Definite Matrix KAIST wit Lab 2012. 07. 10 남성호 Introduction The signs of the eigenvalues can be important. The signs can also be related to the minima,

More information

Geometric Construction of an Ellipse from Its Moments

Geometric Construction of an Ellipse from Its Moments Applied Mathematical Sciences, Vol. 10, 2016, no. 3, 127-135 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2016.511704 Geometric Construction of an llipse from Its Moments Brian J. M c Cartin

More information

UNIT 6: The singular value decomposition.

UNIT 6: The singular value decomposition. UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T

More information

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision) CS4495/6495 Introduction to Computer Vision 8B-L2 Principle Component Analysis (and its use in Computer Vision) Wavelength 2 Wavelength 2 Principal Components Principal components are all about the directions

More information

Properties of surfaces II: Second moment of area

Properties of surfaces II: Second moment of area Properties of surfaces II: Second moment of area Just as we have discussing first moment of an area and its relation with problems in mechanics, we will now describe second moment and product of area of

More information

Principal Components Theory Notes

Principal Components Theory Notes Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory

More information

Minimum Error Rate Classification

Minimum Error Rate Classification Minimum Error Rate Classification Dr. K.Vijayarekha Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur-613 401 Table of Contents 1.Minimum Error Rate Classification...

More information

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson Name: TA Name and section: NO CALCULATORS, SHOW ALL WORK, NO OTHER PAPERS ON DESK. There is very little actual work to be done on this exam if

More information

Bayesian Decision Theory

Bayesian Decision Theory Bayesian Decision Theory Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent University) 1 / 46 Bayesian

More information

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x = Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.

More information

Concentration Ellipsoids

Concentration Ellipsoids Concentration Ellipsoids ECE275A Lecture Supplement Fall 2008 Kenneth Kreutz Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California, San Diego VERSION LSECE275CE

More information

EECS490: Digital Image Processing. Lecture #26

EECS490: Digital Image Processing. Lecture #26 Lecture #26 Moments; invariant moments Eigenvector, principal component analysis Boundary coding Image primitives Image representation: trees, graphs Object recognition and classes Minimum distance classifiers

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Least Squares Optimization

Least Squares Optimization Least Squares Optimization The following is a brief review of least squares optimization and constrained optimization techniques. I assume the reader is familiar with basic linear algebra, including the

More information

Functions of Several Variables

Functions of Several Variables Functions of Several Variables The Unconstrained Minimization Problem where In n dimensions the unconstrained problem is stated as f() x variables. minimize f()x x, is a scalar objective function of vector

More information

Algebra II. Paulius Drungilas and Jonas Jankauskas

Algebra II. Paulius Drungilas and Jonas Jankauskas Algebra II Paulius Drungilas and Jonas Jankauskas Contents 1. Quadratic forms 3 What is quadratic form? 3 Change of variables. 3 Equivalence of quadratic forms. 4 Canonical form. 4 Normal form. 7 Positive

More information

SOLUTION KEY TO THE LINEAR ALGEBRA FINAL EXAM 1 2 ( 2) ( 1) c a = 1 0

SOLUTION KEY TO THE LINEAR ALGEBRA FINAL EXAM 1 2 ( 2) ( 1) c a = 1 0 SOLUTION KEY TO THE LINEAR ALGEBRA FINAL EXAM () We find a least squares solution to ( ) ( ) A x = y or 0 0 a b = c 4 0 0. 0 The normal equation is A T A x = A T y = y or 5 0 0 0 0 0 a b = 5 9. 0 0 4 7

More information

Quadratic forms. Defn

Quadratic forms. Defn Quadratic forms Aim lecture: We use the spectral thm for self-adjoint operators to study solns to some multi-variable quadratic eqns. In this lecture, we work over the real field F = R. We use the notn

More information

Principal component analysis

Principal component analysis Principal component analysis Motivation i for PCA came from major-axis regression. Strong assumption: single homogeneous sample. Free of assumptions when used for exploration. Classical tests of significance

More information

Symmetric matrices and dot products

Symmetric matrices and dot products Symmetric matrices and dot products Proposition An n n matrix A is symmetric iff, for all x, y in R n, (Ax) y = x (Ay). Proof. If A is symmetric, then (Ax) y = x T A T y = x T Ay = x (Ay). If equality

More information

5. Discriminant analysis

5. Discriminant analysis 5. Discriminant analysis We continue from Bayes s rule presented in Section 3 on p. 85 (5.1) where c i is a class, x isap-dimensional vector (data case) and we use class conditional probability (density

More information

Massachusetts Institute of Technology Instrumentation Laboratory Cambridge, Massachusetts

Massachusetts Institute of Technology Instrumentation Laboratory Cambridge, Massachusetts Massachusetts Institute of Technology Instrumentation Laboratory Cambridge, Massachusetts Space Guidance Analysis Memo #9 To: From: SGA Distribution James E. Potter Date: November 0, 96 Subject: Error

More information

Wiley. Methods and Applications of Linear Models. Regression and the Analysis. of Variance. Third Edition. Ishpeming, Michigan RONALD R.

Wiley. Methods and Applications of Linear Models. Regression and the Analysis. of Variance. Third Edition. Ishpeming, Michigan RONALD R. Methods and Applications of Linear Models Regression and the Analysis of Variance Third Edition RONALD R. HOCKING PenHock Statistical Consultants Ishpeming, Michigan Wiley Contents Preface to the Third

More information

On Polygonal Domains with Trigonometric Eigenfunctions of the Laplacian under Dirichlet or Neumann Boundary Conditions

On Polygonal Domains with Trigonometric Eigenfunctions of the Laplacian under Dirichlet or Neumann Boundary Conditions Applied Mathematical Sciences, Vol. 2, 2008, no. 58, 2891-2901 On Polygonal Domains with Trigonometric Eigenfunctions of the Laplacian under Dirichlet or Neumann Boundary Conditions Brian J. M c Cartin

More information

Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as

Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as MAHALANOBIS DISTANCE Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as d E (x, y) = (x 1 y 1 ) 2 + +(x p y p ) 2

More information

Math Matrix Algebra

Math Matrix Algebra Math 44 - Matrix Algebra Review notes - (Alberto Bressan, Spring 7) sec: Orthogonal diagonalization of symmetric matrices When we seek to diagonalize a general n n matrix A, two difficulties may arise:

More information

8. Diagonalization.

8. Diagonalization. 8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard

More information

Lagrange Multipliers

Lagrange Multipliers Optimization with Constraints As long as algebra and geometry have been separated, their progress have been slow and their uses limited; but when these two sciences have been united, they have lent each

More information

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms De La Fuente notes that, if an n n matrix has n distinct eigenvalues, it can be diagonalized. In this supplement, we will provide

More information

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice 3. Eigenvalues and Eigenvectors, Spectral Representation 3.. Eigenvalues and Eigenvectors A vector ' is eigenvector of a matrix K, if K' is parallel to ' and ' 6, i.e., K' k' k is the eigenvalue. If is

More information

Large Sample Properties of Estimators in the Classical Linear Regression Model

Large Sample Properties of Estimators in the Classical Linear Regression Model Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in

More information

Diagonalization of Matrices

Diagonalization of Matrices LECTURE 4 Diagonalization of Matrices Recall that a diagonal matrix is a square n n matrix with non-zero entries only along the diagonal from the upper left to the lower right (the main diagonal) Diagonal

More information

ELLIPSES. Problem: Find the points on the locus. Q(x, y) = 865x 2 294xy + 585y 2 = closest to, and farthest from, the origin. Answer.

ELLIPSES. Problem: Find the points on the locus. Q(x, y) = 865x 2 294xy + 585y 2 = closest to, and farthest from, the origin. Answer. ELLIPSES Problem: Find the points on the locus closest to, and farthest from, the origin. Answer. Q(x, y) 865x 2 294xy + 585y 2 1450 This is a Lagrange multiplier problem: we want to extremize f(x, y)

More information

Least Squares Fitting of Data

Least Squares Fitting of Data Least Squares Fitting of Data David Eberly, Geoetric Tools, Redond WA 98052 https://www.geoetrictools.co/ This work is licensed under the Creative Coons Attribution 4.0 International License. To view a

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

Review problems for Math 511

Review problems for Math 511 Review problems for Math 511 A Eremenko Spring 018 1 Evaluate the determinants: 13547 13647 843 853, 46 47 37 1014 543 443 34 71 61 No electronic devices and no partial credit! Hint: you should not multiply

More information

Basic Concepts in Matrix Algebra

Basic Concepts in Matrix Algebra Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

Linear Algebra. Chapter 8: Eigenvalues: Further Applications and Computations Section 8.2. Applications to Geometry Proofs of Theorems.

Linear Algebra. Chapter 8: Eigenvalues: Further Applications and Computations Section 8.2. Applications to Geometry Proofs of Theorems. Linear Algebra Chapter 8: Eigenvalues: Further Applications and Computations Section 8.2. Applications to Geometry Proofs of Theorems May 1, 2018 () Linear Algebra May 1, 2018 1 / 8 Table of contents 1

More information

Chapter 6. Eigenvalues. Josef Leydold Mathematical Methods WS 2018/19 6 Eigenvalues 1 / 45

Chapter 6. Eigenvalues. Josef Leydold Mathematical Methods WS 2018/19 6 Eigenvalues 1 / 45 Chapter 6 Eigenvalues Josef Leydold Mathematical Methods WS 2018/19 6 Eigenvalues 1 / 45 Closed Leontief Model In a closed Leontief input-output-model consumption and production coincide, i.e. V x = x

More information

Non-linear least squares

Non-linear least squares Non-linear least squares Concept of non-linear least squares We have extensively studied linear least squares or linear regression. We see that there is a unique regression line that can be determined

More information

1. Addition: To every pair of vectors x, y X corresponds an element x + y X such that the commutative and associative properties hold

1. Addition: To every pair of vectors x, y X corresponds an element x + y X such that the commutative and associative properties hold Appendix B Y Mathematical Refresher This appendix presents mathematical concepts we use in developing our main arguments in the text of this book. This appendix can be read in the order in which it appears,

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information

Convex Optimization M2

Convex Optimization M2 Convex Optimization M2 Lecture 8 A. d Aspremont. Convex Optimization M2. 1/57 Applications A. d Aspremont. Convex Optimization M2. 2/57 Outline Geometrical problems Approximation problems Combinatorial

More information

x i e i ) + Q(x n e n ) + ( i<n c ij x i x j

x i e i ) + Q(x n e n ) + ( i<n c ij x i x j Math 210A. Quadratic spaces over R 1. Algebraic preliminaries Let V be a finite free module over a nonzero commutative ring F. Recall that a quadratic form on V is a map Q : V F such that Q(cv) = c 2 Q(v)

More information

Linear Algebra: Characteristic Value Problem

Linear Algebra: Characteristic Value Problem Linear Algebra: Characteristic Value Problem . The Characteristic Value Problem Let < be the set of real numbers and { be the set of complex numbers. Given an n n real matrix A; does there exist a number

More information

Exercise Sheet 1.

Exercise Sheet 1. Exercise Sheet 1 You can download my lecture and exercise sheets at the address http://sami.hust.edu.vn/giang-vien/?name=huynt 1) Let A, B be sets. What does the statement "A is not a subset of B " mean?

More information

Kernel Methods. Machine Learning A W VO

Kernel Methods. Machine Learning A W VO Kernel Methods Machine Learning A 708.063 07W VO Outline 1. Dual representation 2. The kernel concept 3. Properties of kernels 4. Examples of kernel machines Kernel PCA Support vector regression (Relevance

More information

Digital Image Processing Chapter 11 Representation and Description

Digital Image Processing Chapter 11 Representation and Description Digital Image Processing Chapter 11 Representation and Description Last Updated: July 20, 2003 Preview 11.1 Representation 11.1.1 Chain Codes Chain codes are used to represent a boundary by a connected

More information

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C. Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal

More information

Roundoff Error. Monday, August 29, 11

Roundoff Error. Monday, August 29, 11 Roundoff Error A round-off error (rounding error), is the difference between the calculated approximation of a number and its exact mathematical value. Numerical analysis specifically tries to estimate

More information

Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010

Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010 Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010 Diagonalization of Symmetric Real Matrices (from Handout Definition 1 Let δ ij = { 1 if i = j 0 if i j A basis V = {v 1,..., v n } of R n

More information

TBP MATH33A Review Sheet. November 24, 2018

TBP MATH33A Review Sheet. November 24, 2018 TBP MATH33A Review Sheet November 24, 2018 General Transformation Matrices: Function Scaling by k Orthogonal projection onto line L Implementation If we want to scale I 2 by k, we use the following: [

More information

Introduction to Machine Learning

Introduction to Machine Learning 1, DATA11002 Introduction to Machine Learning Lecturer: Teemu Roos TAs: Ville Hyvönen and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer

More information

Numerical Optimization

Numerical Optimization Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function

More information

Flexural-Torsional Buckling of General Cold-Formed Steel Columns with Unequal Unbraced Lengths

Flexural-Torsional Buckling of General Cold-Formed Steel Columns with Unequal Unbraced Lengths Proceedings of the Annual Stability Conference Structural Stability Research Council San Antonio, Texas, March 21-24, 2017 Flexural-Torsional Buckling of General Cold-Formed Steel Columns with Unequal

More information

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.) Prof. Daniel Cremers 2. Regression (cont.) Regression with MLE (Rep.) Assume that y is affected by Gaussian noise : t = f(x, w)+ where Thus, we have p(t x, w, )=N (t; f(x, w), 2 ) 2 Maximum A-Posteriori

More information

Review (Probability & Linear Algebra)

Review (Probability & Linear Algebra) Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint

More information

Principal Component Analysis (PCA) Theory, Practice, and Examples

Principal Component Analysis (PCA) Theory, Practice, and Examples Principal Component Analysis (PCA) Theory, Practice, and Examples Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite) variables. p k n A

More information

Motivating the Covariance Matrix

Motivating the Covariance Matrix Motivating the Covariance Matrix Raúl Rojas Computer Science Department Freie Universität Berlin January 2009 Abstract This note reviews some interesting properties of the covariance matrix and its role

More information

Math 302 Outcome Statements Winter 2013

Math 302 Outcome Statements Winter 2013 Math 302 Outcome Statements Winter 2013 1 Rectangular Space Coordinates; Vectors in the Three-Dimensional Space (a) Cartesian coordinates of a point (b) sphere (c) symmetry about a point, a line, and a

More information

Least Squares Fitting of Data by Linear or Quadratic Structures

Least Squares Fitting of Data by Linear or Quadratic Structures Least Squares Fitting of Data by Linear or Quadratic Structures David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution

More information

Assignment 11 (C + C ) = (C + C ) = (C + C) i(c C ) ] = i(c C) (AB) = (AB) = B A = BA 0 = [A, B] = [A, B] = (AB BA) = (AB) AB

Assignment 11 (C + C ) = (C + C ) = (C + C) i(c C ) ] = i(c C) (AB) = (AB) = B A = BA 0 = [A, B] = [A, B] = (AB BA) = (AB) AB Arfken 3.4.6 Matrix C is not Hermition. But which is Hermitian. Likewise, Assignment 11 (C + C ) = (C + C ) = (C + C) [ i(c C ) ] = i(c C ) = i(c C) = i ( C C ) Arfken 3.4.9 The matrices A and B are both

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Rotations and vibrations of polyatomic molecules

Rotations and vibrations of polyatomic molecules Rotations and vibrations of polyatomic molecules When the potential energy surface V( R 1, R 2,..., R N ) is known we can compute the energy levels of the molecule. These levels can be an effect of: Rotation

More information

We simply compute: for v = x i e i, bilinearity of B implies that Q B (v) = B(v, v) is given by xi x j B(e i, e j ) =

We simply compute: for v = x i e i, bilinearity of B implies that Q B (v) = B(v, v) is given by xi x j B(e i, e j ) = Math 395. Quadratic spaces over R 1. Algebraic preliminaries Let V be a vector space over a field F. Recall that a quadratic form on V is a map Q : V F such that Q(cv) = c 2 Q(v) for all v V and c F, and

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

p(x ω i 0.4 ω 2 ω

p(x ω i 0.4 ω 2 ω p( ω i ). ω.3.. 9 3 FIGURE.. Hypothetical class-conditional probability density functions show the probability density of measuring a particular feature value given the pattern is in category ω i.if represents

More information

Determine the moment of inertia of the area about the x axis Determine the moment of inertia of the area about the y axis.

Determine the moment of inertia of the area about the x axis Determine the moment of inertia of the area about the y axis. 10 Solutions 44918 1/28/09 4:21 PM Page 941 2010 Pearson Education, Inc., Upper Saddle River, NJ. All rights reserved. This material is protected under all copright laws as the currentl eist. No portion

More information

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45 address 12 adjoint matrix 118 alternating 112 alternating 203 angle 159 angle 33 angle 60 area 120 associative 180 augmented matrix 11 axes 5 Axiom of Choice 153 basis 178 basis 210 basis 74 basis test

More information

41514 Dynamics of Machinery

41514 Dynamics of Machinery 41514 Dynamics of Machinery Theory, Experiment, Phenomenology and Industrial Applications Ilmar Ferreira Santos 1. Recapitulation Mathematical Modeling & Steps 2. Example System of Particle 3. Example

More information

Rigid bodies - general theory

Rigid bodies - general theory Rigid bodies - general theory Kinetic Energy: based on FW-26 Consider a system on N particles with all their relative separations fixed: it has 3 translational and 3 rotational degrees of freedom. Motion

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Data Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396

Data Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396 Data Mining Linear & nonlinear classifiers Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1396 1 / 31 Table of contents 1 Introduction

More information

What is Principal Component Analysis?

What is Principal Component Analysis? What is Principal Component Analysis? Principal component analysis (PCA) Reduce the dimensionality of a data set by finding a new set of variables, smaller than the original set of variables Retains most

More information

EXPLICIT MULTIVARIATE BOUNDS OF CHEBYSHEV TYPE

EXPLICIT MULTIVARIATE BOUNDS OF CHEBYSHEV TYPE Annales Univ. Sci. Budapest., Sect. Comp. 42 2014) 109 125 EXPLICIT MULTIVARIATE BOUNDS OF CHEBYSHEV TYPE Villő Csiszár Budapest, Hungary) Tamás Fegyverneki Budapest, Hungary) Tamás F. Móri Budapest, Hungary)

More information

Seminar on Linear Algebra

Seminar on Linear Algebra Supplement Seminar on Linear Algebra Projection, Singular Value Decomposition, Pseudoinverse Kenichi Kanatani Kyoritsu Shuppan Co., Ltd. Contents 1 Linear Space and Projection 1 1.1 Expression of Linear

More information

Principal component analysis

Principal component analysis Principal component analysis Angela Montanari 1 Introduction Principal component analysis (PCA) is one of the most popular multivariate statistical methods. It was first introduced by Pearson (1901) and

More information

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)-matrix Consider the following row operations on A (1) Swap the positions any

More information

We wish the reader success in future encounters with the concepts of linear algebra.

We wish the reader success in future encounters with the concepts of linear algebra. Afterword Our path through linear algebra has emphasized spaces of vectors in dimension 2, 3, and 4 as a means of introducing concepts which go forward to IRn for arbitrary n. But linear algebra does not

More information

Table of Contents. Multivariate methods. Introduction II. Introduction I

Table of Contents. Multivariate methods. Introduction II. Introduction I Table of Contents Introduction Antti Penttilä Department of Physics University of Helsinki Exactum summer school, 04 Construction of multinormal distribution Test of multinormality with 3 Interpretation

More information

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is, 65 Diagonalizable Matrices It is useful to introduce few more concepts, that are common in the literature Definition 65 The characteristic polynomial of an n n matrix A is the function p(λ) det(a λi) Example

More information

AS and A level Further mathematics contents lists

AS and A level Further mathematics contents lists AS and A level Further mathematics contents lists Contents Core Pure Mathematics Book 1/AS... 2 Core Pure Mathematics Book 2... 4 Further Pure Mathematics 1... 6 Further Pure Mathematics 2... 8 Further

More information

Algebra Workshops 10 and 11

Algebra Workshops 10 and 11 Algebra Workshops 1 and 11 Suggestion: For Workshop 1 please do questions 2,3 and 14. For the other questions, it s best to wait till the material is covered in lectures. Bilinear and Quadratic Forms on

More information

Discriminant analysis and supervised classification

Discriminant analysis and supervised classification Discriminant analysis and supervised classification Angela Montanari 1 Linear discriminant analysis Linear discriminant analysis (LDA) also known as Fisher s linear discriminant analysis or as Canonical

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization April 6, 2018 1 / 34 This material is covered in the textbook, Chapters 9 and 10. Some of the materials are taken from it. Some of

More information

Unsupervised Learning: Dimensionality Reduction

Unsupervised Learning: Dimensionality Reduction Unsupervised Learning: Dimensionality Reduction CMPSCI 689 Fall 2015 Sridhar Mahadevan Lecture 3 Outline In this lecture, we set about to solve the problem posed in the previous lecture Given a dataset,

More information

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015 EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,

More information

DOA Estimation using MUSIC and Root MUSIC Methods

DOA Estimation using MUSIC and Root MUSIC Methods DOA Estimation using MUSIC and Root MUSIC Methods EE602 Statistical signal Processing 4/13/2009 Presented By: Chhavipreet Singh(Y515) Siddharth Sahoo(Y5827447) 2 Table of Contents 1 Introduction... 3 2

More information

The Multivariate Gaussian Distribution

The Multivariate Gaussian Distribution The Multivariate Gaussian Distribution Chuong B. Do October, 8 A vector-valued random variable X = T X X n is said to have a multivariate normal or Gaussian) distribution with mean µ R n and covariance

More information

Mathematical Properties of Stiffness Matrices

Mathematical Properties of Stiffness Matrices Mathematical Properties of Stiffness Matrices CEE 4L. Matrix Structural Analysis Department of Civil and Environmental Engineering Duke University Henri P. Gavin Fall, 0 These notes describe some of the

More information

Review of Some Concepts from Linear Algebra: Part 2

Review of Some Concepts from Linear Algebra: Part 2 Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set

More information

CSE 554 Lecture 7: Alignment

CSE 554 Lecture 7: Alignment CSE 554 Lecture 7: Alignment Fall 2012 CSE554 Alignment Slide 1 Review Fairing (smoothing) Relocating vertices to achieve a smoother appearance Method: centroid averaging Simplification Reducing vertex

More information