A Derivation of the 3-entity-type model
|
|
- Malcolm Stevens
- 5 years ago
- Views:
Transcription
1 A Derivation of the 3-entity-type model The 3-entity-type model consists of two relationship matrices, a m n matrix X and a n r matrix Y In the spirit of generalized linear models we define the low rank representations of the relationship matrices to be X f 1 UV T ) and Y f V Z T ) where f 1 : R m n R m n and f : R n r R n r are the prediction links, and U R m k, V R n k, and Z R r k are the parameters of the model for k Z We further define parameter transformations G : R m k R m k, H : R n k R n k, I : R r k R r k, to model prior knowledge about our parameters, eg, regularizers We additionally required the convex conjugate, G U) = sup [U A GA)], U doma) where U A = tru T A) = ij U ija ij is the matrix dot product The overall loss function for our model is 1) LU, V, Z W, W) = αl 1 U, V W) + 1 α)l V, Z W) where we introduce fixed weight matrices for the observations, W R m n and W R n r The individual objectives on the reconstruction of X and Y are, respectively, ) 3) L 1 U, V W) = W F 1 UV T ) X UV T) + G U) + H V ), L V, Z W) = W F V Z T ) Y V Z T) + H V ) + I Z) The objective LU, V, Z W, W) is convex in any one of its arguments, but is in general non-convex in all its arguments As such, we use an alternating minimization scheme that optimizes one factor U, V, or Z at a time This appendix describes the derivation of both gradient and Newton update rules for U, V, and Z For completeness, Section A1 reviews useful definitions from matrix calculus The gradient of the objective, with respect to each argument is derived in Section A Finally, by assuming the loss is decomposable, we derive the Newton update in Section A3, whose additional cost over its gradient analogue is essentially a factor of k times more expensive A1 Matrix Calculus For the sake of completeness we define matrix derivatives, which generalizes both scalar and vector derivatives Using this definition of matrix derivatives, we also generalize the scalar chain and product rules to matrices The discussion herein is based on Magnus et al [1, ] Let M be an n q matrix of variables, where m j denotes the j-th column of M The vec-operator vecm yields an nq 1 matrix that stacks the columns of M: vecm = While there are several common and incompatible) definitions of matrix derivatives, the derivative of a n 1 vector f with respect to a m 1 vector x is almost universally defined as m 1 m m q f 1 Dfx) f x 1 x T = f n x 1 The matrix derivative of an m p matrix function Q of an n q matrix of variables M contains mnpq partial derivatives, and the matrix derivative arranges these partial derivatives into a matrix We define the matrix f 1 x m f n x m 1
2 derivative by coercing matrices into vectors, and using the above definition of the vector derivative, [QM)] 11 DQM) vecqm) vecm) T = [QM)] mp [QM)] 11 [QM)] mp which is an mp nq matrix of partial derivatives This definition encompasses vector and scalar derivatives as special cases The advantages of this formulation include i) unambiguous definitions for the product and chain rules, and ii) we can easily convert DQM) to the more common definition of the matrix derivative, M = m n1 m 1q via the first identification theorem [, ch 9] We additionally require the matrix chain [, pg 11] and matrix product [1] rules: Definition 1 Matrix Chain Rule) Given functions ϕ : R n q R m p, ϕ 1 : R l r R m p, and ϕ : R n q R l r the derivative of ϕm) = ϕ 1 Y ), where Y = ϕ M), is where A B is the matrix product DϕZ) = Dϕ 1 Y ) Dϕ Z) Definition Matrix Product Rule) Given an m p matrix ϕ 1 M) and a p r matrix ϕ M) the derivative of ϕ 1 M)ϕ M) with respect to M, Dϕ 1 ϕ )M) is where A B is the Kronecker product,, Dϕ 1 ϕ )Z) = ϕ M) T I m ) Dϕ 1 M) + I r ϕ 1 M)) Dϕ M) A Computing the Gradient To compute the derivative of Equation 1 with respect to U, V, and Z we require the following three lemmas: Lemma 1 For any differentiable function F : R m n R F 1 UV T ) = f 1 UV T )V, F V Z T ) = f V Z T )Z, F 1 UV T ) = f 1 UV T ) T U F V Z T ) = f V Z T ) T V Proof We only derive the result for ϕu) = F 1 UV T ) The proof is similar for the other three cases ϕu) can be expressed as the composition of functions: ϕu) = F 1 Y ), Y = ϕ U), ϕ U) = UV T Using the matrix chain rule DϕU) = DF 1 Y ) Dϕ U), we note that 4) Using the matrix product rule 5) DF 1 Y ) = vecf 1UV T ) vecuv T ) T = f 1 UV T ) Dϕ U) = V I m ) vecu vecu) T + I n U) vecv T vec U) T = V I m ) Combining equations 4 and 5 using the matrix chain rule yields DϕU) = vecf 1 UV T )V ) T The result follows immediately from the first identification theorem
3 Lemma Given that the entries of U and V are distinct, X UV T ) Y V Z T ) = XV, = Y Z, X UV T ) = X T U Y V Z T ) = Y T V Proof We derive the result for X UV T )/, the other three derivations are similar To avoid a long digression into matrix differentials, we prove the result by element-wise differentiation Noting that X UV T = x ji u jk v ik, i j we compute the derivative with respect to v pq k X UV T ) v pq = i = j j x ji u jk v ik v pq k x jp u jq = X T U) pq Since the result holds for all p {1,,n} and q {1,,k} it follows that X UV T )/ = X T U Lemma 3 For the l regularizers where a, b, c > 0 controls the strength of regularizers larger values = weaker regularization): GU) = a U Fro the derivatives for the convex conjugates are G U), HV ) = b V Fro = U a = A, H V ), IZ) = c Z Fro, = V b = B, I Z) = Z c = C Proof The result is easily proven by finding the convex conjugate and differentiating it Combining Lemmas 1,, and 3, and denoting the Hadamard element-wise) product of matrices A B, we have that 6) 7) 8) LU, V, Z) = α W f 1 UV T ) X )) V + A, LU, V, Z) = α W f 1 UV T ) X )) T U + 1 α) W f V Z T ) Y )) Z + B, LU, V, Z) = 1 α) W f V Z T ) Y )) T V + C Setting the gradient equal to zero yields update equations either for A, B, C or for U, V, Z An advantage of using a gradient update is that we can relax the requirement that the links are differentiable, replacing gradients with subgradients in the prequel A3 Computing the Newton Update One may be satisfied with a gradient step However, the assumption of a decomposable loss means that most second derivatives are set to zero, and a Newton update can be done efficiently, reducing to row-wise 3
4 optimization of U, V, and Z For the subclass of models where Equations 6-8 are differentiable and the loss is decomposable define, qu i ) = α W i f 1 U i V T ) X i )) V + Ai qv i ) = α W i f 1 UV T i ) X i )) T U + 1 α) Wi f V i Z T ) Y i ) ) Z + B i qz i ) = 1 α) W i f V Z T i ) Y i) ) T V + Ci Any local optimum of the loss corresponds to roots of the equations {qu i )} m i=1, {qv i )} n i=1, and {qz i )} r i=1 Using a Newton step, the update for U i is 9) U new i = U i η [qu i )] [q U i )] 1, where η [0, 1] is the step length, chosen using line search with the Armijo criterion [3, ch 3] The Newton steps for V i and Z i are analogous To describe the Hessians of the loss, q, we introduce the following notation for the Hessians of G, H and I : G i diag G U i )), H i diag H V i )), I i diag I Z i )) For case of l -regularization G i = diaga 1 1) For conciseness we also introduce the following terms: D 1,i diagw i f 1U i V T )), D,i diagw i f 1UV T i )), D 3,i diag W i f V T i Z)), D 4,i diag W i f V ZT i )) This allows us to describe the Hessians of Equation 1 with respect to each row of the parameter matrices Lemma 4 The Hessians of Equation 1 with respect to U i, V i, and Z i are q U i ) qu i ) i q Z i ) qz i ) i q V i ) qv i ) i = αv T D 1,i V + G i = 1 α)v T D 4,i V + I i = αu T D,i U + 1 α)z T D 3,i Z + H i Proof We prove the result for q U i ), noting that the other derivations are similar Since q ) and its argument U i are both vectors, DqU i ) is identical to qu i )/ i Ignoring the terms that do not vary with U i, DqU i ) = D [ αw i fu i V T ))V + A i ] = αd [ W i fu i V T ))V ] + DA i { = α V T I 1 ) vecw i fu i V T )) + I k V ) vecv } + vec G U i ) vecu i vecu i vecu i = α { V T I 1 ) D 1,i V + I k V ) 0 } + G i = αv T D 1,i V + G i Plugging the gradient qu i ) and the Hessian q U i ) into Equation 9 yields 10) U new i q U i ) = U i αv T D 1,i V + G i ) + αη W i X i fu i V T ) )) V ηa i = αu i V T D 1,i V + αη W i X i fu i V T ) )) D 1 1,i D 1,iV + U i G i ηa i = α U i V T + η W i X i fu i V T)) D 1 1,i ) D1,i V + U i G i ηa i 4
5 Likewise for Z i, 11) Z new i q Z i ) = 1 α) Z i V T + η W i Y i fv Zi T ) )) ) T D 1 4,i D 4,i V + Z i I i ηc i The derivation of the update for V i is similar, since LU, V, Z W, W) is a linear combination and the differential operator is linear: { Vi new q V i ) = α V i U T + η W i ) } X i fuvi T ))) T D 1 1) i D,i U + { 1 α) V i Z T + η Wi Y i fv i Z T ) )) ) } D 1 3,i D 3,i Z + V i H i ηb i While D {D j,i } 4 j=1 may not be invertible, ie, a diagonal entry is zero when the corresponding weight is zero, the form of the update equations shows that this does not matter If a diagonal entry in D is zero, replacing its corresponding entry in D 1 by any nonzero value does not change the result of Equations 10, 11, and 1, as the zero weight cancels it out References [1] J R Magnus and K M Abadir On some definitions in matrix algebra Working Paper CIRJE-F-46, University of Tokyo, Feb 007 [] J R Magnus and H Neudecker Matrix Differential Calculus with Applications in Statistics and Econometrics John Wiley, 007 [3] J Nocedal and S J Wright Numerical Optimization Springer Series in Operations Research Springer,
Secrets of Matrix Factorization: Supplementary Document
Secrets of Matrix Factorization: Supplementary Document Je Hyeong Hong University of Cambridge jhh37@cam.ac.uk Andrew Fitzgibbon Microsoft, Cambridge, UK awf@microsoft.com Abstract This document consists
More informationRelational Learning via Collective Matrix Factorization. June 2008 CMU-ML
Relational Learning via Collective Matrix Factorization Ajit P. Singh Geoffrey J. Gordon June 2008 CMU-ML-08-109 Relational Learning via Collective Matrix Factorization Ajit P. Singh Geoffrey J. Gordon
More informationMatrices. Math 240 Calculus III. Wednesday, July 10, Summer 2013, Session II. Matrices. Math 240. Definitions and Notation.
function Matrices Calculus III Summer 2013, Session II Wednesday, July 10, 2013 Agenda function 1. 2. function function Definition An m n matrix is a rectangular array of numbers arranged in m horizontal
More informationB553 Lecture 5: Matrix Algebra Review
B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More informationMatrix Arithmetic. j=1
An m n matrix is an array A = Matrix Arithmetic a 11 a 12 a 1n a 21 a 22 a 2n a m1 a m2 a mn of real numbers a ij An m n matrix has m rows and n columns a ij is the entry in the i-th row and j-th column
More informationLinear Algebra Solutions 1
Math Camp 1 Do the following: Linear Algebra Solutions 1 1. Let A = and B = 3 8 5 A B = 3 5 9 A + B = 9 11 14 4 AB = 69 3 16 BA = 1 4 ( 1 3. Let v = and u = 5 uv = 13 u v = 13 v u = 13 Math Camp 1 ( 7
More informationTensor Analysis in Euclidean Space
Tensor Analysis in Euclidean Space James Emery Edited: 8/5/2016 Contents 1 Classical Tensor Notation 2 2 Multilinear Functionals 4 3 Operations With Tensors 5 4 The Directional Derivative 5 5 Curvilinear
More informationMatrices: 2.1 Operations with Matrices
Goals In this chapter and section we study matrix operations: Define matrix addition Define multiplication of matrix by a scalar, to be called scalar multiplication. Define multiplication of two matrices,
More informationECE580 Fall 2015 Solution to Midterm Exam 1 October 23, Please leave fractions as fractions, but simplify them, etc.
ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, 2015 1 Name: Solution Score: /100 This exam is closed-book. You must show ALL of your work for full credit. Please read the questions carefully.
More informationLinear Algebra (Review) Volker Tresp 2018
Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c
More informationr=1 r=1 argmin Q Jt (20) After computing the descent direction d Jt 2 dt H t d + P (x + d) d i = 0, i / J
7 Appendix 7. Proof of Theorem Proof. There are two main difficulties in proving the convergence of our algorithm, and none of them is addressed in previous works. First, the Hessian matrix H is a block-structured
More informationMultilinear Algebra For the Undergraduate Algebra Student
Multilinear Algebra For the Undergraduate Algebra Student Davis Shurbert Department of Mathematics and Computer Science University of Puget Sound April 4, 204 Introduction When working in the field of
More informationMatrix representation of a linear map
Matrix representation of a linear map As before, let e i = (0,..., 0, 1, 0,..., 0) T, with 1 in the i th place and 0 elsewhere, be standard basis vectors. Given linear map f : R n R m we get n column vectors
More informationCS 246 Review of Linear Algebra 01/17/19
1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector
More informationJim Lambers MAT 610 Summer Session Lecture 1 Notes
Jim Lambers MAT 60 Summer Session 2009-0 Lecture Notes Introduction This course is about numerical linear algebra, which is the study of the approximate solution of fundamental problems from linear algebra
More informationMATRICES. a m,1 a m,n A =
MATRICES Matrices are rectangular arrays of real or complex numbers With them, we define arithmetic operations that are generalizations of those for real and complex numbers The general form a matrix of
More informationLecture: Algorithms for LP, SOCP and SDP
1/53 Lecture: Algorithms for LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:
More informationChapter 2. Matrix Arithmetic. Chapter 2
Matrix Arithmetic Matrix Addition and Subtraction Addition and subtraction act element-wise on matrices. In order for the addition/subtraction (A B) to be possible, the two matrices A and B must have the
More informationFinal Project # 5 The Cartan matrix of a Root System
8.099 Final Project # 5 The Cartan matrix of a Root System Thomas R. Covert July 30, 00 A root system in a Euclidean space V with a symmetric positive definite inner product, is a finite set of elements
More informationMatrix Arithmetic. (Multidimensional Math) Institute for Advanced Analytics. Shaina Race, PhD
Matrix Arithmetic Multidimensional Math) Shaina Race, PhD Institute for Advanced Analytics. TABLE OF CONTENTS Element-wise Operations Linear Combinations of Matrices and Vectors. Vector Multiplication
More informationFinite Elements. Colin Cotter. February 22, Colin Cotter FEM
Finite Elements February 22, 2019 In the previous sections, we introduced the concept of finite element spaces, which contain certain functions defined on a domain. Finite element spaces are examples of
More information. a m1 a mn. a 1 a 2 a = a n
Biostat 140655, 2008: Matrix Algebra Review 1 Definition: An m n matrix, A m n, is a rectangular array of real numbers with m rows and n columns Element in the i th row and the j th column is denoted by
More informationBare minimum on matrix algebra. Psychology 588: Covariance structure and factor models
Bare minimum on matrix algebra Psychology 588: Covariance structure and factor models Matrix multiplication 2 Consider three notations for linear combinations y11 y1 m x11 x 1p b11 b 1m y y x x b b n1
More informationMatrix Factorization and Analysis
Chapter 7 Matrix Factorization and Analysis Matrix factorizations are an important part of the practice and analysis of signal processing. They are at the heart of many signal-processing algorithms. Their
More information= 1 and 2 1. T =, and so det A b d
Chapter 8 Determinants The founder of the theory of determinants is usually taken to be Gottfried Wilhelm Leibniz (1646 1716, who also shares the credit for inventing calculus with Sir Isaac Newton (1643
More informationLinear Algebra and Matrix Inversion
Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much
More informationMAT 610: Numerical Linear Algebra. James V. Lambers
MAT 610: Numerical Linear Algebra James V Lambers January 16, 2017 2 Contents 1 Matrix Multiplication Problems 7 11 Introduction 7 111 Systems of Linear Equations 7 112 The Eigenvalue Problem 8 12 Basic
More information2: LINEAR TRANSFORMATIONS AND MATRICES
2: LINEAR TRANSFORMATIONS AND MATRICES STEVEN HEILMAN Contents 1. Review 1 2. Linear Transformations 1 3. Null spaces, range, coordinate bases 2 4. Linear Transformations and Bases 4 5. Matrix Representation,
More informationSolvability of Linear Matrix Equations in a Symmetric Matrix Variable
Solvability of Linear Matrix Equations in a Symmetric Matrix Variable Maurcio C. de Oliveira J. William Helton Abstract We study the solvability of generalized linear matrix equations of the Lyapunov type
More informationQUATERNIONS AND ROTATIONS
QUATERNIONS AND ROTATIONS SVANTE JANSON 1. Introduction The purpose of this note is to show some well-known relations between quaternions and the Lie groups SO(3) and SO(4) (rotations in R 3 and R 4 )
More informationLinear Algebra review Powers of a diagonalizable matrix Spectral decomposition
Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2018 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 2: Orthogonal Vectors and Matrices; Vector Norms Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 11 Outline 1 Orthogonal
More informationBasics of Calculus and Algebra
Monika Department of Economics ISCTE-IUL September 2012 Basics of linear algebra Real valued Functions Differential Calculus Integral Calculus Optimization Introduction I A matrix is a rectangular array
More informationBasic Concepts in Linear Algebra
Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University February 2, 2015 Grady B Wright Linear Algebra Basics February 2, 2015 1 / 39 Numerical Linear Algebra Linear
More informationMatrix Derivatives and Descent Optimization Methods
Matrix Derivatives and Descent Optimization Methods 1 Qiang Ning Department of Electrical and Computer Engineering Beckman Institute for Advanced Science and Techonology University of Illinois at Urbana-Champaign
More informationI = i 0,
Special Types of Matrices Certain matrices, such as the identity matrix 0 0 0 0 0 0 I = 0 0 0, 0 0 0 have a special shape, which endows the matrix with helpful properties The identity matrix is an example
More informationRANK AND PERIMETER PRESERVER OF RANK-1 MATRICES OVER MAX ALGEBRA
Discussiones Mathematicae General Algebra and Applications 23 (2003 ) 125 137 RANK AND PERIMETER PRESERVER OF RANK-1 MATRICES OVER MAX ALGEBRA Seok-Zun Song and Kyung-Tae Kang Department of Mathematics,
More informationFunctions of Several Variables
Functions of Several Variables The Unconstrained Minimization Problem where In n dimensions the unconstrained problem is stated as f() x variables. minimize f()x x, is a scalar objective function of vector
More informationMATH 2030: MATRICES. Example 0.2. Q:Define A 1 =, A. 3 4 A: We wish to find c 1, c 2, and c 3 such that. c 1 + c c
MATH 2030: MATRICES Matrix Algebra As with vectors, we may use the algebra of matrices to simplify calculations. However, matrices have operations that vectors do not possess, and so it will be of interest
More informationLinear and Multilinear Algebra. Linear maps preserving rank of tensor products of matrices
Linear maps preserving rank of tensor products of matrices Journal: Manuscript ID: GLMA-0-0 Manuscript Type: Original Article Date Submitted by the Author: -Aug-0 Complete List of Authors: Zheng, Baodong;
More informationLinear Algebra review Powers of a diagonalizable matrix Spectral decomposition
Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2016 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing
More informationMatrices. 1 a a2 1 b b 2 1 c c π
Matrices 2-3-207 A matrix is a rectangular array of numbers: 2 π 4 37 42 0 3 a a2 b b 2 c c 2 Actually, the entries can be more general than numbers, but you can think of the entries as numbers to start
More informationMath 240 Calculus III
The Calculus III Summer 2015, Session II Wednesday, July 8, 2015 Agenda 1. of the determinant 2. determinants 3. of determinants What is the determinant? Yesterday: Ax = b has a unique solution when A
More informationNumerical Optimization
Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function
More informationReview of Basic Concepts in Linear Algebra
Review of Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University September 7, 2017 Math 565 Linear Algebra Review September 7, 2017 1 / 40 Numerical Linear Algebra
More information5.6. PSEUDOINVERSES 101. A H w.
5.6. PSEUDOINVERSES 0 Corollary 5.6.4. If A is a matrix such that A H A is invertible, then the least-squares solution to Av = w is v = A H A ) A H w. The matrix A H A ) A H is the left inverse of A and
More informationEE364a Review Session 7
EE364a Review Session 7 EE364a Review session outline: derivatives and chain rule (Appendix A.4) numerical linear algebra (Appendix C) factor and solve method exploiting structure and sparsity 1 Derivative
More informationReview of Linear Algebra
Review of Linear Algebra VBS/MRC Review of Linear Algebra 0 Ok, the Questions Why does algebra sound like cobra? Why do they say abstract algebra? Why bother about all this abstract stuff? What is a vector,
More informationAnnouncements Wednesday, October 10
Announcements Wednesday, October 10 The second midterm is on Friday, October 19 That is one week from this Friday The exam covers 35, 36, 37, 39, 41, 42, 43, 44 (through today s material) WeBWorK 42, 43
More informationPOSITIVE MAP AS DIFFERENCE OF TWO COMPLETELY POSITIVE OR SUPER-POSITIVE MAPS
Adv. Oper. Theory 3 (2018), no. 1, 53 60 http://doi.org/10.22034/aot.1702-1129 ISSN: 2538-225X (electronic) http://aot-math.org POSITIVE MAP AS DIFFERENCE OF TWO COMPLETELY POSITIVE OR SUPER-POSITIVE MAPS
More informationMatrix Multiplication
3.2 Matrix Algebra Matrix Multiplication Example Foxboro Stadium has three main concession stands, located behind the south, north and west stands. The top-selling items are peanuts, hot dogs and soda.
More informationLinear Algebra Review
Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite
More informationAnnouncements Monday, October 02
Announcements Monday, October 02 Please fill out the mid-semester survey under Quizzes on Canvas WeBWorK 18, 19 are due Wednesday at 11:59pm The quiz on Friday covers 17, 18, and 19 My office is Skiles
More informationMatrix Algebra: Summary
May, 27 Appendix E Matrix Algebra: Summary ontents E. Vectors and Matrtices.......................... 2 E.. Notation.................................. 2 E..2 Special Types of Vectors.........................
More informationSufficiency of Signed Principal Minors for Semidefiniteness: A Relatively Easy Proof
Sufficiency of Signed Principal Minors for Semidefiniteness: A Relatively Easy Proof David M. Mandy Department of Economics University of Missouri 118 Professional Building Columbia, MO 65203 USA mandyd@missouri.edu
More informationLecture 7 Spectral methods
CSE 291: Unsupervised learning Spring 2008 Lecture 7 Spectral methods 7.1 Linear algebra review 7.1.1 Eigenvalues and eigenvectors Definition 1. A d d matrix M has eigenvalue λ if there is a d-dimensional
More information18.S34 linear algebra problems (2007)
18.S34 linear algebra problems (2007) Useful ideas for evaluating determinants 1. Row reduction, expanding by minors, or combinations thereof; sometimes these are useful in combination with an induction
More informationCheng Soon Ong & Christian Walder. Canberra February June 2017
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2017 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 141 Part III
More informationOn Expected Gaussian Random Determinants
On Expected Gaussian Random Determinants Moo K. Chung 1 Department of Statistics University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 Abstract The expectation of random determinants whose
More informationMathematical Foundations of Applied Statistics: Matrix Algebra
Mathematical Foundations of Applied Statistics: Matrix Algebra Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/105 Literature Seber, G.
More informationNear-Isometry by Relaxation: Supplement
000 00 00 003 004 005 006 007 008 009 00 0 0 03 04 05 06 07 08 09 00 0 0 03 04 05 06 07 08 09 030 03 03 033 034 035 036 037 038 039 040 04 04 043 044 045 046 047 048 049 050 05 05 053 Near-Isometry by
More informationb 1 b 2.. b = b m A = [a 1,a 2,...,a n ] where a 1,j a 2,j a j = a m,j Let A R m n and x 1 x 2 x = x n
Lectures -2: Linear Algebra Background Almost all linear and nonlinear problems in scientific computation require the use of linear algebra These lectures review basic concepts in a way that has proven
More informationMAC Module 2 Systems of Linear Equations and Matrices II. Learning Objectives. Upon completing this module, you should be able to :
MAC 0 Module Systems of Linear Equations and Matrices II Learning Objectives Upon completing this module, you should be able to :. Find the inverse of a square matrix.. Determine whether a matrix is invertible..
More informationVector, Matrix, and Tensor Derivatives
Vector, Matrix, and Tensor Derivatives Erik Learned-Miller The purpose of this document is to help you learn to take derivatives of vectors, matrices, and higher order tensors (arrays with three dimensions
More informationMatrix Algebra. Matrix Algebra. Chapter 8 - S&B
Chapter 8 - S&B Algebraic operations Matrix: The size of a matrix is indicated by the number of its rows and the number of its columns. A matrix with k rows and n columns is called a k n matrix. The number
More informationInstitute for Computational Mathematics Hong Kong Baptist University
Institute for Computational Mathematics Hong Kong Baptist University ICM Research Report 08-0 How to find a good submatrix S. A. Goreinov, I. V. Oseledets, D. V. Savostyanov, E. E. Tyrtyshnikov, N. L.
More informationMATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018
Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry
More informationLinear algebra 2. Yoav Zemel. March 1, 2012
Linear algebra 2 Yoav Zemel March 1, 2012 These notes were written by Yoav Zemel. The lecturer, Shmuel Berger, should not be held responsible for any mistake. Any comments are welcome at zamsh7@gmail.com.
More informationElementary maths for GMT
Elementary maths for GMT Linear Algebra Part 2: Matrices, Elimination and Determinant m n matrices The system of m linear equations in n variables x 1, x 2,, x n a 11 x 1 + a 12 x 2 + + a 1n x n = b 1
More informationIEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 8, AUGUST /$ IEEE
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 8, AUGUST 2009 3613 Hessian and Concavity of Mutual Information, Differential Entropy, and Entropy Power in Linear Vector Gaussian Channels Miquel
More informationA new product on 2 2 matrices
A new product on 2 2 matrices L Kramer, P Kramer, and V Man ko October 17, 2018 arxiv:1802.03048v1 [math-ph] 8 Feb 2018 Abstract We study a bilinear multiplication rule on 2 2 matrices which is intermediate
More informationA FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic
A FIRST COURSE IN LINEAR ALGEBRA An Open Text by Ken Kuttler Matrix Arithmetic Lecture Notes by Karen Seyffarth Adapted by LYRYX SERVICE COURSE SOLUTION Attribution-NonCommercial-ShareAlike (CC BY-NC-SA)
More informationSparse Levenberg-Marquardt algorithm.
Sparse Levenberg-Marquardt algorithm. R. I. Hartley and A. Zisserman: Multiple View Geometry in Computer Vision. Cambridge University Press, second edition, 2004. Appendix 6 was used in part. The Levenberg-Marquardt
More informationOptimization Tutorial 1. Basic Gradient Descent
E0 270 Machine Learning Jan 16, 2015 Optimization Tutorial 1 Basic Gradient Descent Lecture by Harikrishna Narasimhan Note: This tutorial shall assume background in elementary calculus and linear algebra.
More informationThe Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)
Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will
More informationTopics. Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij
Topics Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij or a ij lives in row i and column j Definition of a matrix
More informationMatrices and Vectors
Matrices and Vectors James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 11, 2013 Outline 1 Matrices and Vectors 2 Vector Details 3 Matrix
More informationARTICLE IN PRESS. Journal of Multivariate Analysis ( ) Contents lists available at ScienceDirect. Journal of Multivariate Analysis
Journal of Multivariate Analysis ( ) Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Marginal parameterizations of discrete models
More informationMatrix Operations: Determinant
Matrix Operations: Determinant Determinants Determinants are only applicable for square matrices. Determinant of the square matrix A is denoted as: det(a) or A Recall that the absolute value of the determinant
More informationAccelerated Block-Coordinate Relaxation for Regularized Optimization
Accelerated Block-Coordinate Relaxation for Regularized Optimization Stephen J. Wright Computer Sciences University of Wisconsin, Madison October 09, 2012 Problem descriptions Consider where f is smooth
More information(Part 1) High-dimensional statistics May / 41
Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2
More informationStrongly Regular Decompositions of the Complete Graph
Journal of Algebraic Combinatorics, 17, 181 201, 2003 c 2003 Kluwer Academic Publishers. Manufactured in The Netherlands. Strongly Regular Decompositions of the Complete Graph EDWIN R. VAN DAM Edwin.vanDam@uvt.nl
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationLinear Algebra V = T = ( 4 3 ).
Linear Algebra Vectors A column vector is a list of numbers stored vertically The dimension of a column vector is the number of values in the vector W is a -dimensional column vector and V is a 5-dimensional
More informationLinear Algebra Review. Vectors
Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors
More informationDOMINO TILING. Contents 1. Introduction 1 2. Rectangular Grids 2 Acknowledgments 10 References 10
DOMINO TILING KASPER BORYS Abstract In this paper we explore the problem of domino tiling: tessellating a region with x2 rectangular dominoes First we address the question of existence for domino tilings
More informationTHE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION
THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION MANTAS MAŽEIKA Abstract. The purpose of this paper is to present a largely self-contained proof of the singular value decomposition (SVD), and
More informationApril 26, Applied mathematics PhD candidate, physics MA UC Berkeley. Lecture 4/26/2013. Jed Duersch. Spd matrices. Cholesky decomposition
Applied mathematics PhD candidate, physics MA UC Berkeley April 26, 2013 UCB 1/19 Symmetric positive-definite I Definition A symmetric matrix A R n n is positive definite iff x T Ax > 0 holds x 0 R n.
More informationMa/CS 6b Class 20: Spectral Graph Theory
Ma/CS 6b Class 20: Spectral Graph Theory By Adam Sheffer Eigenvalues and Eigenvectors A an n n matrix of real numbers. The eigenvalues of A are the numbers λ such that Ax = λx for some nonzero vector x
More informationCSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming
CSC2411 - Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming Notes taken by Mike Jamieson March 28, 2005 Summary: In this lecture, we introduce semidefinite programming
More informationElementary Row Operations on Matrices
King Saud University September 17, 018 Table of contents 1 Definition A real matrix is a rectangular array whose entries are real numbers. These numbers are organized on rows and columns. An m n matrix
More informationMatrix Differential Calculus with Applications in Statistics and Econometrics
Matrix Differential Calculus with Applications in Statistics and Econometrics Revised Edition JAN. R. MAGNUS CentERjor Economic Research, Tilburg University and HEINZ NEUDECKER Cesaro, Schagen JOHN WILEY
More informationICS 6N Computational Linear Algebra Matrix Algebra
ICS 6N Computational Linear Algebra Matrix Algebra Xiaohui Xie University of California, Irvine xhx@uci.edu February 2, 2017 Xiaohui Xie (UCI) ICS 6N February 2, 2017 1 / 24 Matrix Consider an m n matrix
More informationLinear Algebra Review
Chapter 1 Linear Algebra Review It is assumed that you have had a beginning course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc I will review some of these terms here,
More informationMathematics 13: Lecture 10
Mathematics 13: Lecture 10 Matrices Dan Sloughter Furman University January 25, 2008 Dan Sloughter (Furman University) Mathematics 13: Lecture 10 January 25, 2008 1 / 19 Matrices Recall: A matrix is a
More informationMultivariable Calculus
2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)
More information1 Matrices and Systems of Linear Equations
March 3, 203 6-6. Systems of Linear Equations Matrices and Systems of Linear Equations An m n matrix is an array A = a ij of the form a a n a 2 a 2n... a m a mn where each a ij is a real or complex number.
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)
AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 1: Course Overview; Matrix Multiplication Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical
More information