On the Relationship Between Concentration and Inertia Hyperellipsoids
|
|
- Beverley Sheila Stevens
- 5 years ago
- Views:
Transcription
1 Applied Mathematical Sciences, Vol. 2, 2008, no. 10, On the Relationship Between Concentration and Inertia Hyperellipsoids Brian J. M c Cartin Applied Mathematics, Kettering University 1700 West Third Avenue, Flint, MI USA bmccarti@kettering.edu Abstract We extend the known intricate connections between the concentration ellipsoid of statistics [1, 2] and the inertia ellipsoid of mechanics [3, 4] from three to n dimensions. We then also extend the relationship involving the eccentricities of their principal cross-sections [5] to n dimensions. The shared degeneracies of these hyperellipsoids is noted. Mathematics Subject Classifications: 51N20, 62J05, 70F99 Keywords: concentration hyperellipsoid; correlation hyperellipsoid; hyperellipsoid of residuals; inertia hyperellipsoid; momental hyperellipsoid; hyperellipsoid of gyration; orthogonal regression; least squares approximation 1 Introduction Mathematical statistics and classical mechanics both employ certain hyperellipsoids in their analytical investigations. These hyperellipsoids are distinct in n>3 dimensions while the corresponding two dimensional ellipses essentially coincide. In what follows, we first develop the concentration hyperellipsoid of statistics [1, 2] and the inertia hperellipsoid of mechanics [3, 4]. The known connections between these two hyperellipsoids in three dimensions are then extended to higher dimensions. Next, a relationship concerning the eccentricities of their principal cross-sections [5] is also extended from three to higher dimensions. Finally, it is observed that the previously derived relationships imply that both hyperellipsoids degenerate to hyperspheroids under identical conditions.
2 490 B. J. McCartin 2 Concentration Hyperellipsoid Consider the set of m data points in R n with coordinates {(x (i) 1,...,x (i) n )} m with centroid ( x 1,..., x n ) defined by the mean values x j := 1 m m Furthermore, let us introduce the variances x (i) j ;(j =1,...,n). (1) σj 2 := 1 m m (x (i) j x j ) 2 ;(j =1,...,n), (2) and the covariances p jk := 1 m m (x (i) j x j ) (x (i) k x k); (j =1,...,n; k =1,...,n; j k). (3) Defining the symmetric covariance matrix as σ 2 1 p 12 p 1n p 12 σ 2 C := 2 p 2n....., (4). p 1n p 2n σn 2 observe that v := [a 1 a 2 a n ] T implies that v T C v = E{[a 1 (x 1 x 1 )+a 2 (x 2 x 2 )+ + a n (x n x n )] 2 } 0, (5) where E is the averaging operator E{u} = 1 m m u (i), with equality if and only if the data all lie on the hyperplane of dimension n 1 a 1 (x 1 x 1 )+a 2 (x 2 x 2 )+ + a n (x n x n )=0. (6) Thus, hyperplanarity of the data corresponds to a zero eigenvalue of the positive semidefinite matrix C and the corresponding eigenvector is normal to this hyperplane. Of course, the data may further degenerate to a hyperplane of even lower dimension. Thus, without loss of generality, we will assume that C is positive definite in the ensuing analysis. The concentration hyperellipsoid (also known as a correlation hyperellipsoid) is defined by ( r r) T C 1 ( r r) =n +2, (7) which has the same first and second moments about the centroid as does the data [2, p. 300]. In this sense, it is the hyperellipsoid which is most representative of the data points without any a priori statistical assumptions
3 Concentration and inertia hyperellipsoids 491 concerning their origin. The reciprocal of the concentration hyperellipsoid, obtained by replacement of C 1 by C in the quadratic form of Equation (7), will be referred to as the hyperellipsoid of residuals [1]. Denoting the eigenvalues of C 1 by ν 1 <ν 2 < < ν n, the principal semiaxes of the concentration hyperellipsoid pass through its centroid and lie in the directions of the the corresponding eigenvectors, v j (j =1,...,n), while their lengths are given by 1/ ν j [6, p. 309]. The hyperplanar cross-sections of dimension n p passing through the centroid and perpendicular to p of these semiaxes are the corresponding principal sections of dimension n p of the concentration ellipsoid. Note that these principal sections are hyperellipsoids of dimension n p with principal semiaxes specified by the remaining n p eigenpairs. In particular, for p = n 2, these n(n 1) principal sections are ellipses and will be referred 2 to simply as the principal sections of the concentration hyperellipsoid. For p =1,...,n, the hyperplane of dimension n p which minimizes the mean square distance of the points from the hyperplane is called the orthogonal regression hyperplane of dimension n p and may be thought of as the hyperplane of dimension n p which best fits the data. The following important facts were first established by Pearson [1] in three dimensions and have subsequently been extended to n dimensions to form the basis for so-called orthogonal regression. The sum of the p th smallest eigenvalues of C equals the minimum mean square deviation from the orthogonal regression hyperplane of dimension n p which passes through the centroid and is normal to the corresponding p eigenvectors. These orthogonal regression hyperplanes are nested in that, for p =1,...,n 1, the one of dimension n p contains that of dimension n p 1. 3 Inertia Hyperellipsoid We next focus our attention on orthogonal linear regression which selects (α 1,α 2,...,α n ), the direction angles of a line (i.e., a hyperplane of dimension one), L, passing through the centroid, so as to minimize m n 1 r 2 (α 1,α 2,...,α n ):= { [cos α j (x (i) k x k) cos α k (x (i) j x j )] 2 }, (8) j=1 k=j+1 which is the total square distance of the data from L [7, p. 20]. Alternatively, r 2 (α 1,α 2,...,α n ) may be viewed as the moment of inertia with respect to L of unit masses located at the data points [3, p. 66].
4 492 B. J. McCartin In any event, the mean square deviation is given by n 1 1 m r2 (α 1,α 2,...,α n )= j=1 k=j+1 cos 2 α j σ 2 k 2 cos α j cos α k p jk + cos 2 α k σ 2 j. (9) Now, along L, mark off the two points at a distance l(α 1,α 2,...,α n ):= 1 r from the centroid. As α 2 (α 1,α 2,...,α n) 1,α 2,...,α n vary, these points sweep out a hyperellipsoid centered at the centroid. We see this as follows. Since cos α j =(x j x j )/l, 1 l 2 = r2 (α 1,α 2,...,α n )= j=1 jj (x j x j ) 2 n 1 I +2 l 2 j=1 k=j+1 I jk (x j x j )(x k x k ) l 2, (10) where I jj := m k j σ 2 k and I jk := m p jk are, respectively, the moments of inertia and the products of inertia about the coordinate axes [3, p. 67]. Equation (10) may be rewritten in matrix form as ( r r) T I( r r) =1, (11) where r =(x 1,x 2,...,x n ) T, r =( x 1, x 2,..., x n ) T, and the inertia matrix I is defined as I 11 I 12 I 1n I 12 I 22 I 2n I := (12). I 1n I 2n I nn That this n-dimensional quadric surface is a hyperellipsoid follows from the relation I = m [tr(c)i C], (13) so that the positive definiteness of I follows directly from that of C. This hyperellipsoid with l 2 (α 1,α 2,...,α n ) = 1/(moment of inertia) is called the inertia hyperellipsoid (also known as the momental hyperellipsoid) because of its role in rotational mechanics [3, p. 66]. Replacement of I by I 1 in the quadratic form of Equation (11) yields the reciprocal hyperellipsoid of gyration [4, p. 233]. The minimum value of r 2 (α 1,α 2,...,α n ) is achieved precisely when l 2 (α 1,α 2,...,α n ) assumes its maximum value which will, of course, correspond to the semimajor axis of the inertia hyperellipsoid (i.e., the largest of the principal semiaxes). Therefore, the minimum total square deviation is given by s 2 := r 2 (α 1,α 2,...,α n)=1/(semimajor axis) 2 = λ min, (14)
5 Concentration and inertia hyperellipsoids 493 where (α 1,α 2,...,α n ) are the direction angles of the semimajor axis and λ min is the smallest eigenvalue of I whose corresponding eigenvector points along the semimajor axis of the inertia hyperellipsoid. Alternatively, the orthogonal regression line is parallel to the eigenvector corresponding to the largest eigenvalue of the hyperellipsoid of gyration. In stark contrast to the two-dimensional scenario [5], the inertia hyperellipsoid is not homothetic (i.e., similar ) to the concentration hyperellipsoid for n>2. However, as we now will see, they are indeed closely related. 4 Hyperellipsoidal Relationships Equation (13) indicates that there is an intimate relationship between the inertia and concentration hyperellipsoids, a state of affairs which we now explore. In this investigation, we will denote the eigenvalues of I as λ 1 <λ 2 < <λ n, the eigenvalues of C as μ 1 <μ 2 < <μ n, and the eigenvalues of C 1 as ν 1 <ν 2 < <ν n. The degenerate case of multiple eigenvalues will be subsequently studied. In consequence of Equation (13) we have for i =1,...,n the relations λ i = m (μ 1 + μ μ n μ n i+1 ) with same eigenvector as μ n i+1. (15) Inverting these expressions we arrive at the complementary relations [4, p. 230] μ i = λ 1 + λ λ n (n 1)λ n i+1. (16) m(n 1) Observe that the positive definiteness of C implies the important inertia matrix inequality λ n < (λ 1 + λ λ n 1 )/(n 2) [4, p. 231]. Furthermore, Equations (15) and (16) together imply the desired relationship between the spectra of the inertia and concentration hyperellipsoids: ν i = 1 m(n 1) = with same eigenvector as λ i. μ n i+1 λ 1 + λ λ n (n 1)λ i (17) Thus, the corresponding principal semiaxes of the inertia and concentration hyperellipsoids are parallel. In addition, Equations (15) through (17) together imply the following collection of inequalities: Theorem 1 (Eccentricity Inequalities) For n>2, ν j ν k < λ j λ k (j =1,...,n 1; k = j +1,...,n). (18)
6 494 B. J. McCartin Proof: Since μ i (i =1,...,n) are all positive and n>2, we have Now, since we have μ n j+1 + μ n k+1 < μ i. j<k μ n k+1 <μ n j+1, (μ n j+1 + μ n k+1 ) (μ n j+1 μ n k+1 ) < μ i (μ n j+1 μ n k+1 ). This may be rearranged to produce μ n k+1 ( μ i μ n j+1 μ n k+1 ) <μ n j+1 ( μ i μ n k+1 μ n j+1 ). Adding μ n j+1 μ n k+1 to both sides yields μ n k+1 ( μ i μ n k+1 ) <μ n j+1 ( μ i μ n j+1 ), which implies that μ n n k+1 μ i μ n j+1 < μ n. n j+1 μ i μ n k+1 Finally, this may be recast as ν j < λ j. ν k λ k Note that if n = 2 then all of the above inequalities become equalities. Thus, not only are the principal axes of the inertia and concentration hyperellipsoids parallel, but each principal section of the inertia hyperellipsoid is less eccentric than the corresponding principal section of the concentration hyperellipsoid. This hyperellipsoidal relationship is a natural generalization of the corresponding three dimensional result first presented in [5]. 5 Conclusion We conclude with the straightforward observation that the inertia and concentration hyperellipsoids share degeneracies to hypershperoids, i.e. multiple eigenvalues, μ j 1 <μ j = = μ k <μ k+1, (19) in which case the orthogonal regression planes of dimension n j through n k + 1 will be indeterminate. An auxiliary consequence of the preceding
7 Concentration and inertia hyperellipsoids 495 considerations is that the concentration hyperellipsoid also shares degeneracies with the hyperellipsoid of residuals except that now, due to their reciprocal relationship, the inequalities of Equation (19) become ν n k <ν n k+1 = = ν n j+1 <ν n j+2. (20) The characteristic polynomial of C is called the discriminating polynomial of degree n and multiplicity of its roots is equivalent to degeneracy of these hyperellipsoids to hyperspheroids. Unfortunately, necessary and sufficient conditions for multiple roots of the discriminating polynomial are not readily available for n>3. Acknowledgement. This paper is dedicated to my loving wife, Barbara A. McCartin, for her support of all of my mathematical endeavors. References [1] K. Pearson, On Lines and Planes of Closest Fit to Systems of Points in Space, Phil. Mag. 2 (1901), [2] H. Cramér, Mathematical Methods of Statistics, Princeton University Press, Princeton, NJ, [3] H. Lamb, Higher Mechanics, Second Edition, Cambridge University Press, Cambridge, [4] A. G. Webster, Dynamics of Particles and of Rigid, Elastic, and Fluid Bodies, Third Edition, Stechert, New York, [5] B. J. McCartin, On Concentration and Inertia Ellipsoids, Applied Mathematical Sciences 1(1) (2007), [6] F. R. Gantmacher, Matrix Theory: Volume I, Chelsea, New York, [7] M. G. Kendall, A Course in the Geometry of n Dimensions, Hafner, New York, NY, Received: August 26, 2007
Geometric Constructions Relating Various Lines of Regression
Applied Mathematical Sciences, Vol. 0, 06, no. 3, - 30 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/0.988/ams.06.5735 Geometric Constructions Relating Various Lines of Regression Brian J. M c Cartin
More informationA Matrix-Vector Analytic Demonstration of Pappus Construction of an Ellipse from a Pair of Conjugate Diameters
Applied Mathematical Sciences, Vol. 9, 015, no. 14, 679-687 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.1988/ams.015.411035 A Matrix-Vector Analytic Demonstration of Pappus Construction of an Ellipse
More informationA A x i x j i j (i, j) (j, i) Let. Compute the value of for and
7.2 - Quadratic Forms quadratic form on is a function defined on whose value at a vector in can be computed by an expression of the form, where is an symmetric matrix. The matrix R n Q R n x R n Q(x) =
More informationRegression. Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables X and Y.
Regression Bivariate i linear regression: Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables and. Generally describe as a
More informationECE 275A Homework 6 Solutions
ECE 275A Homework 6 Solutions. The notation used in the solutions for the concentration (hyper) ellipsoid problems is defined in the lecture supplement on concentration ellipsoids. Note that θ T Σ θ =
More informationLinear Algebra And Its Applications Chapter 6. Positive Definite Matrix
Linear Algebra And Its Applications Chapter 6. Positive Definite Matrix KAIST wit Lab 2012. 07. 10 남성호 Introduction The signs of the eigenvalues can be important. The signs can also be related to the minima,
More informationGeometric Construction of an Ellipse from Its Moments
Applied Mathematical Sciences, Vol. 10, 2016, no. 3, 127-135 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2016.511704 Geometric Construction of an llipse from Its Moments Brian J. M c Cartin
More informationUNIT 6: The singular value decomposition.
UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T
More informationCS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)
CS4495/6495 Introduction to Computer Vision 8B-L2 Principle Component Analysis (and its use in Computer Vision) Wavelength 2 Wavelength 2 Principal Components Principal components are all about the directions
More informationProperties of surfaces II: Second moment of area
Properties of surfaces II: Second moment of area Just as we have discussing first moment of an area and its relation with problems in mechanics, we will now describe second moment and product of area of
More informationPrincipal Components Theory Notes
Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory
More informationMinimum Error Rate Classification
Minimum Error Rate Classification Dr. K.Vijayarekha Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur-613 401 Table of Contents 1.Minimum Error Rate Classification...
More informationFinal Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson
Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson Name: TA Name and section: NO CALCULATORS, SHOW ALL WORK, NO OTHER PAPERS ON DESK. There is very little actual work to be done on this exam if
More informationBayesian Decision Theory
Bayesian Decision Theory Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent University) 1 / 46 Bayesian
More informationVectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =
Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.
More informationConcentration Ellipsoids
Concentration Ellipsoids ECE275A Lecture Supplement Fall 2008 Kenneth Kreutz Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California, San Diego VERSION LSECE275CE
More informationEECS490: Digital Image Processing. Lecture #26
Lecture #26 Moments; invariant moments Eigenvector, principal component analysis Boundary coding Image primitives Image representation: trees, graphs Object recognition and classes Minimum distance classifiers
More informationLinear Algebra: Matrix Eigenvalue Problems
CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given
More informationLeast Squares Optimization
Least Squares Optimization The following is a brief review of least squares optimization and constrained optimization techniques. I assume the reader is familiar with basic linear algebra, including the
More informationFunctions of Several Variables
Functions of Several Variables The Unconstrained Minimization Problem where In n dimensions the unconstrained problem is stated as f() x variables. minimize f()x x, is a scalar objective function of vector
More informationAlgebra II. Paulius Drungilas and Jonas Jankauskas
Algebra II Paulius Drungilas and Jonas Jankauskas Contents 1. Quadratic forms 3 What is quadratic form? 3 Change of variables. 3 Equivalence of quadratic forms. 4 Canonical form. 4 Normal form. 7 Positive
More informationSOLUTION KEY TO THE LINEAR ALGEBRA FINAL EXAM 1 2 ( 2) ( 1) c a = 1 0
SOLUTION KEY TO THE LINEAR ALGEBRA FINAL EXAM () We find a least squares solution to ( ) ( ) A x = y or 0 0 a b = c 4 0 0. 0 The normal equation is A T A x = A T y = y or 5 0 0 0 0 0 a b = 5 9. 0 0 4 7
More informationQuadratic forms. Defn
Quadratic forms Aim lecture: We use the spectral thm for self-adjoint operators to study solns to some multi-variable quadratic eqns. In this lecture, we work over the real field F = R. We use the notn
More informationPrincipal component analysis
Principal component analysis Motivation i for PCA came from major-axis regression. Strong assumption: single homogeneous sample. Free of assumptions when used for exploration. Classical tests of significance
More informationSymmetric matrices and dot products
Symmetric matrices and dot products Proposition An n n matrix A is symmetric iff, for all x, y in R n, (Ax) y = x (Ay). Proof. If A is symmetric, then (Ax) y = x T A T y = x T Ay = x (Ay). If equality
More information5. Discriminant analysis
5. Discriminant analysis We continue from Bayes s rule presented in Section 3 on p. 85 (5.1) where c i is a class, x isap-dimensional vector (data case) and we use class conditional probability (density
More informationMassachusetts Institute of Technology Instrumentation Laboratory Cambridge, Massachusetts
Massachusetts Institute of Technology Instrumentation Laboratory Cambridge, Massachusetts Space Guidance Analysis Memo #9 To: From: SGA Distribution James E. Potter Date: November 0, 96 Subject: Error
More informationWiley. Methods and Applications of Linear Models. Regression and the Analysis. of Variance. Third Edition. Ishpeming, Michigan RONALD R.
Methods and Applications of Linear Models Regression and the Analysis of Variance Third Edition RONALD R. HOCKING PenHock Statistical Consultants Ishpeming, Michigan Wiley Contents Preface to the Third
More informationOn Polygonal Domains with Trigonometric Eigenfunctions of the Laplacian under Dirichlet or Neumann Boundary Conditions
Applied Mathematical Sciences, Vol. 2, 2008, no. 58, 2891-2901 On Polygonal Domains with Trigonometric Eigenfunctions of the Laplacian under Dirichlet or Neumann Boundary Conditions Brian J. M c Cartin
More informationDef. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as
MAHALANOBIS DISTANCE Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as d E (x, y) = (x 1 y 1 ) 2 + +(x p y p ) 2
More informationMath Matrix Algebra
Math 44 - Matrix Algebra Review notes - (Alberto Bressan, Spring 7) sec: Orthogonal diagonalization of symmetric matrices When we seek to diagonalize a general n n matrix A, two difficulties may arise:
More information8. Diagonalization.
8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard
More informationLagrange Multipliers
Optimization with Constraints As long as algebra and geometry have been separated, their progress have been slow and their uses limited; but when these two sciences have been united, they have lent each
More informationEcon 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis
Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms De La Fuente notes that, if an n n matrix has n distinct eigenvalues, it can be diagonalized. In this supplement, we will provide
More informationlinearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice
3. Eigenvalues and Eigenvectors, Spectral Representation 3.. Eigenvalues and Eigenvectors A vector ' is eigenvector of a matrix K, if K' is parallel to ' and ' 6, i.e., K' k' k is the eigenvalue. If is
More informationLarge Sample Properties of Estimators in the Classical Linear Regression Model
Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in
More informationDiagonalization of Matrices
LECTURE 4 Diagonalization of Matrices Recall that a diagonal matrix is a square n n matrix with non-zero entries only along the diagonal from the upper left to the lower right (the main diagonal) Diagonal
More informationELLIPSES. Problem: Find the points on the locus. Q(x, y) = 865x 2 294xy + 585y 2 = closest to, and farthest from, the origin. Answer.
ELLIPSES Problem: Find the points on the locus closest to, and farthest from, the origin. Answer. Q(x, y) 865x 2 294xy + 585y 2 1450 This is a Lagrange multiplier problem: we want to extremize f(x, y)
More informationLeast Squares Fitting of Data
Least Squares Fitting of Data David Eberly, Geoetric Tools, Redond WA 98052 https://www.geoetrictools.co/ This work is licensed under the Creative Coons Attribution 4.0 International License. To view a
More informationPrinciple Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA
Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In
More informationReview problems for Math 511
Review problems for Math 511 A Eremenko Spring 018 1 Evaluate the determinants: 13547 13647 843 853, 46 47 37 1014 543 443 34 71 61 No electronic devices and no partial credit! Hint: you should not multiply
More informationBasic Concepts in Matrix Algebra
Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1
More informationx. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).
.8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics
More informationLinear Algebra. Chapter 8: Eigenvalues: Further Applications and Computations Section 8.2. Applications to Geometry Proofs of Theorems.
Linear Algebra Chapter 8: Eigenvalues: Further Applications and Computations Section 8.2. Applications to Geometry Proofs of Theorems May 1, 2018 () Linear Algebra May 1, 2018 1 / 8 Table of contents 1
More informationChapter 6. Eigenvalues. Josef Leydold Mathematical Methods WS 2018/19 6 Eigenvalues 1 / 45
Chapter 6 Eigenvalues Josef Leydold Mathematical Methods WS 2018/19 6 Eigenvalues 1 / 45 Closed Leontief Model In a closed Leontief input-output-model consumption and production coincide, i.e. V x = x
More informationNon-linear least squares
Non-linear least squares Concept of non-linear least squares We have extensively studied linear least squares or linear regression. We see that there is a unique regression line that can be determined
More information1. Addition: To every pair of vectors x, y X corresponds an element x + y X such that the commutative and associative properties hold
Appendix B Y Mathematical Refresher This appendix presents mathematical concepts we use in developing our main arguments in the text of this book. This appendix can be read in the order in which it appears,
More informationRecall the convention that, for us, all vectors are column vectors.
Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists
More informationConvex Optimization M2
Convex Optimization M2 Lecture 8 A. d Aspremont. Convex Optimization M2. 1/57 Applications A. d Aspremont. Convex Optimization M2. 2/57 Outline Geometrical problems Approximation problems Combinatorial
More informationx i e i ) + Q(x n e n ) + ( i<n c ij x i x j
Math 210A. Quadratic spaces over R 1. Algebraic preliminaries Let V be a finite free module over a nonzero commutative ring F. Recall that a quadratic form on V is a map Q : V F such that Q(cv) = c 2 Q(v)
More informationLinear Algebra: Characteristic Value Problem
Linear Algebra: Characteristic Value Problem . The Characteristic Value Problem Let < be the set of real numbers and { be the set of complex numbers. Given an n n real matrix A; does there exist a number
More informationExercise Sheet 1.
Exercise Sheet 1 You can download my lecture and exercise sheets at the address http://sami.hust.edu.vn/giang-vien/?name=huynt 1) Let A, B be sets. What does the statement "A is not a subset of B " mean?
More informationKernel Methods. Machine Learning A W VO
Kernel Methods Machine Learning A 708.063 07W VO Outline 1. Dual representation 2. The kernel concept 3. Properties of kernels 4. Examples of kernel machines Kernel PCA Support vector regression (Relevance
More informationDigital Image Processing Chapter 11 Representation and Description
Digital Image Processing Chapter 11 Representation and Description Last Updated: July 20, 2003 Preview 11.1 Representation 11.1.1 Chain Codes Chain codes are used to represent a boundary by a connected
More informationThroughout these notes we assume V, W are finite dimensional inner product spaces over C.
Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal
More informationRoundoff Error. Monday, August 29, 11
Roundoff Error A round-off error (rounding error), is the difference between the calculated approximation of a number and its exact mathematical value. Numerical analysis specifically tries to estimate
More informationEconomics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010
Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010 Diagonalization of Symmetric Real Matrices (from Handout Definition 1 Let δ ij = { 1 if i = j 0 if i j A basis V = {v 1,..., v n } of R n
More informationTBP MATH33A Review Sheet. November 24, 2018
TBP MATH33A Review Sheet November 24, 2018 General Transformation Matrices: Function Scaling by k Orthogonal projection onto line L Implementation If we want to scale I 2 by k, we use the following: [
More informationIntroduction to Machine Learning
1, DATA11002 Introduction to Machine Learning Lecturer: Teemu Roos TAs: Ville Hyvönen and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer
More informationNumerical Optimization
Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function
More informationFlexural-Torsional Buckling of General Cold-Formed Steel Columns with Unequal Unbraced Lengths
Proceedings of the Annual Stability Conference Structural Stability Research Council San Antonio, Texas, March 21-24, 2017 Flexural-Torsional Buckling of General Cold-Formed Steel Columns with Unequal
More informationComputer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)
Prof. Daniel Cremers 2. Regression (cont.) Regression with MLE (Rep.) Assume that y is affected by Gaussian noise : t = f(x, w)+ where Thus, we have p(t x, w, )=N (t; f(x, w), 2 ) 2 Maximum A-Posteriori
More informationReview (Probability & Linear Algebra)
Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint
More informationPrincipal Component Analysis (PCA) Theory, Practice, and Examples
Principal Component Analysis (PCA) Theory, Practice, and Examples Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite) variables. p k n A
More informationMotivating the Covariance Matrix
Motivating the Covariance Matrix Raúl Rojas Computer Science Department Freie Universität Berlin January 2009 Abstract This note reviews some interesting properties of the covariance matrix and its role
More informationMath 302 Outcome Statements Winter 2013
Math 302 Outcome Statements Winter 2013 1 Rectangular Space Coordinates; Vectors in the Three-Dimensional Space (a) Cartesian coordinates of a point (b) sphere (c) symmetry about a point, a line, and a
More informationLeast Squares Fitting of Data by Linear or Quadratic Structures
Least Squares Fitting of Data by Linear or Quadratic Structures David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution
More informationAssignment 11 (C + C ) = (C + C ) = (C + C) i(c C ) ] = i(c C) (AB) = (AB) = B A = BA 0 = [A, B] = [A, B] = (AB BA) = (AB) AB
Arfken 3.4.6 Matrix C is not Hermition. But which is Hermitian. Likewise, Assignment 11 (C + C ) = (C + C ) = (C + C) [ i(c C ) ] = i(c C ) = i(c C) = i ( C C ) Arfken 3.4.9 The matrices A and B are both
More informationChapter 3 Transformations
Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases
More informationRotations and vibrations of polyatomic molecules
Rotations and vibrations of polyatomic molecules When the potential energy surface V( R 1, R 2,..., R N ) is known we can compute the energy levels of the molecule. These levels can be an effect of: Rotation
More informationWe simply compute: for v = x i e i, bilinearity of B implies that Q B (v) = B(v, v) is given by xi x j B(e i, e j ) =
Math 395. Quadratic spaces over R 1. Algebraic preliminaries Let V be a vector space over a field F. Recall that a quadratic form on V is a map Q : V F such that Q(cv) = c 2 Q(v) for all v V and c F, and
More informationLecture 7: Positive Semidefinite Matrices
Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.
More informationp(x ω i 0.4 ω 2 ω
p( ω i ). ω.3.. 9 3 FIGURE.. Hypothetical class-conditional probability density functions show the probability density of measuring a particular feature value given the pattern is in category ω i.if represents
More informationDetermine the moment of inertia of the area about the x axis Determine the moment of inertia of the area about the y axis.
10 Solutions 44918 1/28/09 4:21 PM Page 941 2010 Pearson Education, Inc., Upper Saddle River, NJ. All rights reserved. This material is protected under all copright laws as the currentl eist. No portion
More informationhomogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45
address 12 adjoint matrix 118 alternating 112 alternating 203 angle 159 angle 33 angle 60 area 120 associative 180 augmented matrix 11 axes 5 Axiom of Choice 153 basis 178 basis 210 basis 74 basis test
More information41514 Dynamics of Machinery
41514 Dynamics of Machinery Theory, Experiment, Phenomenology and Industrial Applications Ilmar Ferreira Santos 1. Recapitulation Mathematical Modeling & Steps 2. Example System of Particle 3. Example
More informationRigid bodies - general theory
Rigid bodies - general theory Kinetic Energy: based on FW-26 Consider a system on N particles with all their relative separations fixed: it has 3 translational and 3 rotational degrees of freedom. Motion
More informationEigenvalues, Eigenvectors, and an Intro to PCA
Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.
More informationEigenvalues, Eigenvectors, and an Intro to PCA
Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.
More informationData Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396
Data Mining Linear & nonlinear classifiers Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1396 1 / 31 Table of contents 1 Introduction
More informationWhat is Principal Component Analysis?
What is Principal Component Analysis? Principal component analysis (PCA) Reduce the dimensionality of a data set by finding a new set of variables, smaller than the original set of variables Retains most
More informationEXPLICIT MULTIVARIATE BOUNDS OF CHEBYSHEV TYPE
Annales Univ. Sci. Budapest., Sect. Comp. 42 2014) 109 125 EXPLICIT MULTIVARIATE BOUNDS OF CHEBYSHEV TYPE Villő Csiszár Budapest, Hungary) Tamás Fegyverneki Budapest, Hungary) Tamás F. Móri Budapest, Hungary)
More informationSeminar on Linear Algebra
Supplement Seminar on Linear Algebra Projection, Singular Value Decomposition, Pseudoinverse Kenichi Kanatani Kyoritsu Shuppan Co., Ltd. Contents 1 Linear Space and Projection 1 1.1 Expression of Linear
More informationPrincipal component analysis
Principal component analysis Angela Montanari 1 Introduction Principal component analysis (PCA) is one of the most popular multivariate statistical methods. It was first introduced by Pearson (1901) and
More informationANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3
ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)-matrix Consider the following row operations on A (1) Swap the positions any
More informationWe wish the reader success in future encounters with the concepts of linear algebra.
Afterword Our path through linear algebra has emphasized spaces of vectors in dimension 2, 3, and 4 as a means of introducing concepts which go forward to IRn for arbitrary n. But linear algebra does not
More informationTable of Contents. Multivariate methods. Introduction II. Introduction I
Table of Contents Introduction Antti Penttilä Department of Physics University of Helsinki Exactum summer school, 04 Construction of multinormal distribution Test of multinormality with 3 Interpretation
More informationA = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,
65 Diagonalizable Matrices It is useful to introduce few more concepts, that are common in the literature Definition 65 The characteristic polynomial of an n n matrix A is the function p(λ) det(a λi) Example
More informationAS and A level Further mathematics contents lists
AS and A level Further mathematics contents lists Contents Core Pure Mathematics Book 1/AS... 2 Core Pure Mathematics Book 2... 4 Further Pure Mathematics 1... 6 Further Pure Mathematics 2... 8 Further
More informationAlgebra Workshops 10 and 11
Algebra Workshops 1 and 11 Suggestion: For Workshop 1 please do questions 2,3 and 14. For the other questions, it s best to wait till the material is covered in lectures. Bilinear and Quadratic Forms on
More informationDiscriminant analysis and supervised classification
Discriminant analysis and supervised classification Angela Montanari 1 Linear discriminant analysis Linear discriminant analysis (LDA) also known as Fisher s linear discriminant analysis or as Canonical
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationCSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization
CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization April 6, 2018 1 / 34 This material is covered in the textbook, Chapters 9 and 10. Some of the materials are taken from it. Some of
More informationUnsupervised Learning: Dimensionality Reduction
Unsupervised Learning: Dimensionality Reduction CMPSCI 689 Fall 2015 Sridhar Mahadevan Lecture 3 Outline In this lecture, we set about to solve the problem posed in the previous lecture Given a dataset,
More informationEE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015
EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,
More informationDOA Estimation using MUSIC and Root MUSIC Methods
DOA Estimation using MUSIC and Root MUSIC Methods EE602 Statistical signal Processing 4/13/2009 Presented By: Chhavipreet Singh(Y515) Siddharth Sahoo(Y5827447) 2 Table of Contents 1 Introduction... 3 2
More informationThe Multivariate Gaussian Distribution
The Multivariate Gaussian Distribution Chuong B. Do October, 8 A vector-valued random variable X = T X X n is said to have a multivariate normal or Gaussian) distribution with mean µ R n and covariance
More informationMathematical Properties of Stiffness Matrices
Mathematical Properties of Stiffness Matrices CEE 4L. Matrix Structural Analysis Department of Civil and Environmental Engineering Duke University Henri P. Gavin Fall, 0 These notes describe some of the
More informationReview of Some Concepts from Linear Algebra: Part 2
Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set
More informationCSE 554 Lecture 7: Alignment
CSE 554 Lecture 7: Alignment Fall 2012 CSE554 Alignment Slide 1 Review Fairing (smoothing) Relocating vertices to achieve a smoother appearance Method: centroid averaging Simplification Reducing vertex
More information