A Derivation of the 3-entity-type model

Size: px
Start display at page:

Download "A Derivation of the 3-entity-type model"

Transcription

1 A Derivation of the 3-entity-type model The 3-entity-type model consists of two relationship matrices, a m n matrix X and a n r matrix Y In the spirit of generalized linear models we define the low rank representations of the relationship matrices to be X f 1 UV T ) and Y f V Z T ) where f 1 : R m n R m n and f : R n r R n r are the prediction links, and U R m k, V R n k, and Z R r k are the parameters of the model for k Z We further define parameter transformations G : R m k R m k, H : R n k R n k, I : R r k R r k, to model prior knowledge about our parameters, eg, regularizers We additionally required the convex conjugate, G U) = sup [U A GA)], U doma) where U A = tru T A) = ij U ija ij is the matrix dot product The overall loss function for our model is 1) LU, V, Z W, W) = αl 1 U, V W) + 1 α)l V, Z W) where we introduce fixed weight matrices for the observations, W R m n and W R n r The individual objectives on the reconstruction of X and Y are, respectively, ) 3) L 1 U, V W) = W F 1 UV T ) X UV T) + G U) + H V ), L V, Z W) = W F V Z T ) Y V Z T) + H V ) + I Z) The objective LU, V, Z W, W) is convex in any one of its arguments, but is in general non-convex in all its arguments As such, we use an alternating minimization scheme that optimizes one factor U, V, or Z at a time This appendix describes the derivation of both gradient and Newton update rules for U, V, and Z For completeness, Section A1 reviews useful definitions from matrix calculus The gradient of the objective, with respect to each argument is derived in Section A Finally, by assuming the loss is decomposable, we derive the Newton update in Section A3, whose additional cost over its gradient analogue is essentially a factor of k times more expensive A1 Matrix Calculus For the sake of completeness we define matrix derivatives, which generalizes both scalar and vector derivatives Using this definition of matrix derivatives, we also generalize the scalar chain and product rules to matrices The discussion herein is based on Magnus et al [1, ] Let M be an n q matrix of variables, where m j denotes the j-th column of M The vec-operator vecm yields an nq 1 matrix that stacks the columns of M: vecm = While there are several common and incompatible) definitions of matrix derivatives, the derivative of a n 1 vector f with respect to a m 1 vector x is almost universally defined as m 1 m m q f 1 Dfx) f x 1 x T = f n x 1 The matrix derivative of an m p matrix function Q of an n q matrix of variables M contains mnpq partial derivatives, and the matrix derivative arranges these partial derivatives into a matrix We define the matrix f 1 x m f n x m 1

2 derivative by coercing matrices into vectors, and using the above definition of the vector derivative, [QM)] 11 DQM) vecqm) vecm) T = [QM)] mp [QM)] 11 [QM)] mp which is an mp nq matrix of partial derivatives This definition encompasses vector and scalar derivatives as special cases The advantages of this formulation include i) unambiguous definitions for the product and chain rules, and ii) we can easily convert DQM) to the more common definition of the matrix derivative, M = m n1 m 1q via the first identification theorem [, ch 9] We additionally require the matrix chain [, pg 11] and matrix product [1] rules: Definition 1 Matrix Chain Rule) Given functions ϕ : R n q R m p, ϕ 1 : R l r R m p, and ϕ : R n q R l r the derivative of ϕm) = ϕ 1 Y ), where Y = ϕ M), is where A B is the matrix product DϕZ) = Dϕ 1 Y ) Dϕ Z) Definition Matrix Product Rule) Given an m p matrix ϕ 1 M) and a p r matrix ϕ M) the derivative of ϕ 1 M)ϕ M) with respect to M, Dϕ 1 ϕ )M) is where A B is the Kronecker product,, Dϕ 1 ϕ )Z) = ϕ M) T I m ) Dϕ 1 M) + I r ϕ 1 M)) Dϕ M) A Computing the Gradient To compute the derivative of Equation 1 with respect to U, V, and Z we require the following three lemmas: Lemma 1 For any differentiable function F : R m n R F 1 UV T ) = f 1 UV T )V, F V Z T ) = f V Z T )Z, F 1 UV T ) = f 1 UV T ) T U F V Z T ) = f V Z T ) T V Proof We only derive the result for ϕu) = F 1 UV T ) The proof is similar for the other three cases ϕu) can be expressed as the composition of functions: ϕu) = F 1 Y ), Y = ϕ U), ϕ U) = UV T Using the matrix chain rule DϕU) = DF 1 Y ) Dϕ U), we note that 4) Using the matrix product rule 5) DF 1 Y ) = vecf 1UV T ) vecuv T ) T = f 1 UV T ) Dϕ U) = V I m ) vecu vecu) T + I n U) vecv T vec U) T = V I m ) Combining equations 4 and 5 using the matrix chain rule yields DϕU) = vecf 1 UV T )V ) T The result follows immediately from the first identification theorem

3 Lemma Given that the entries of U and V are distinct, X UV T ) Y V Z T ) = XV, = Y Z, X UV T ) = X T U Y V Z T ) = Y T V Proof We derive the result for X UV T )/, the other three derivations are similar To avoid a long digression into matrix differentials, we prove the result by element-wise differentiation Noting that X UV T = x ji u jk v ik, i j we compute the derivative with respect to v pq k X UV T ) v pq = i = j j x ji u jk v ik v pq k x jp u jq = X T U) pq Since the result holds for all p {1,,n} and q {1,,k} it follows that X UV T )/ = X T U Lemma 3 For the l regularizers where a, b, c > 0 controls the strength of regularizers larger values = weaker regularization): GU) = a U Fro the derivatives for the convex conjugates are G U), HV ) = b V Fro = U a = A, H V ), IZ) = c Z Fro, = V b = B, I Z) = Z c = C Proof The result is easily proven by finding the convex conjugate and differentiating it Combining Lemmas 1,, and 3, and denoting the Hadamard element-wise) product of matrices A B, we have that 6) 7) 8) LU, V, Z) = α W f 1 UV T ) X )) V + A, LU, V, Z) = α W f 1 UV T ) X )) T U + 1 α) W f V Z T ) Y )) Z + B, LU, V, Z) = 1 α) W f V Z T ) Y )) T V + C Setting the gradient equal to zero yields update equations either for A, B, C or for U, V, Z An advantage of using a gradient update is that we can relax the requirement that the links are differentiable, replacing gradients with subgradients in the prequel A3 Computing the Newton Update One may be satisfied with a gradient step However, the assumption of a decomposable loss means that most second derivatives are set to zero, and a Newton update can be done efficiently, reducing to row-wise 3

4 optimization of U, V, and Z For the subclass of models where Equations 6-8 are differentiable and the loss is decomposable define, qu i ) = α W i f 1 U i V T ) X i )) V + Ai qv i ) = α W i f 1 UV T i ) X i )) T U + 1 α) Wi f V i Z T ) Y i ) ) Z + B i qz i ) = 1 α) W i f V Z T i ) Y i) ) T V + Ci Any local optimum of the loss corresponds to roots of the equations {qu i )} m i=1, {qv i )} n i=1, and {qz i )} r i=1 Using a Newton step, the update for U i is 9) U new i = U i η [qu i )] [q U i )] 1, where η [0, 1] is the step length, chosen using line search with the Armijo criterion [3, ch 3] The Newton steps for V i and Z i are analogous To describe the Hessians of the loss, q, we introduce the following notation for the Hessians of G, H and I : G i diag G U i )), H i diag H V i )), I i diag I Z i )) For case of l -regularization G i = diaga 1 1) For conciseness we also introduce the following terms: D 1,i diagw i f 1U i V T )), D,i diagw i f 1UV T i )), D 3,i diag W i f V T i Z)), D 4,i diag W i f V ZT i )) This allows us to describe the Hessians of Equation 1 with respect to each row of the parameter matrices Lemma 4 The Hessians of Equation 1 with respect to U i, V i, and Z i are q U i ) qu i ) i q Z i ) qz i ) i q V i ) qv i ) i = αv T D 1,i V + G i = 1 α)v T D 4,i V + I i = αu T D,i U + 1 α)z T D 3,i Z + H i Proof We prove the result for q U i ), noting that the other derivations are similar Since q ) and its argument U i are both vectors, DqU i ) is identical to qu i )/ i Ignoring the terms that do not vary with U i, DqU i ) = D [ αw i fu i V T ))V + A i ] = αd [ W i fu i V T ))V ] + DA i { = α V T I 1 ) vecw i fu i V T )) + I k V ) vecv } + vec G U i ) vecu i vecu i vecu i = α { V T I 1 ) D 1,i V + I k V ) 0 } + G i = αv T D 1,i V + G i Plugging the gradient qu i ) and the Hessian q U i ) into Equation 9 yields 10) U new i q U i ) = U i αv T D 1,i V + G i ) + αη W i X i fu i V T ) )) V ηa i = αu i V T D 1,i V + αη W i X i fu i V T ) )) D 1 1,i D 1,iV + U i G i ηa i = α U i V T + η W i X i fu i V T)) D 1 1,i ) D1,i V + U i G i ηa i 4

5 Likewise for Z i, 11) Z new i q Z i ) = 1 α) Z i V T + η W i Y i fv Zi T ) )) ) T D 1 4,i D 4,i V + Z i I i ηc i The derivation of the update for V i is similar, since LU, V, Z W, W) is a linear combination and the differential operator is linear: { Vi new q V i ) = α V i U T + η W i ) } X i fuvi T ))) T D 1 1) i D,i U + { 1 α) V i Z T + η Wi Y i fv i Z T ) )) ) } D 1 3,i D 3,i Z + V i H i ηb i While D {D j,i } 4 j=1 may not be invertible, ie, a diagonal entry is zero when the corresponding weight is zero, the form of the update equations shows that this does not matter If a diagonal entry in D is zero, replacing its corresponding entry in D 1 by any nonzero value does not change the result of Equations 10, 11, and 1, as the zero weight cancels it out References [1] J R Magnus and K M Abadir On some definitions in matrix algebra Working Paper CIRJE-F-46, University of Tokyo, Feb 007 [] J R Magnus and H Neudecker Matrix Differential Calculus with Applications in Statistics and Econometrics John Wiley, 007 [3] J Nocedal and S J Wright Numerical Optimization Springer Series in Operations Research Springer,

Secrets of Matrix Factorization: Supplementary Document

Secrets of Matrix Factorization: Supplementary Document Secrets of Matrix Factorization: Supplementary Document Je Hyeong Hong University of Cambridge jhh37@cam.ac.uk Andrew Fitzgibbon Microsoft, Cambridge, UK awf@microsoft.com Abstract This document consists

More information

Relational Learning via Collective Matrix Factorization. June 2008 CMU-ML

Relational Learning via Collective Matrix Factorization. June 2008 CMU-ML Relational Learning via Collective Matrix Factorization Ajit P. Singh Geoffrey J. Gordon June 2008 CMU-ML-08-109 Relational Learning via Collective Matrix Factorization Ajit P. Singh Geoffrey J. Gordon

More information

Matrices. Math 240 Calculus III. Wednesday, July 10, Summer 2013, Session II. Matrices. Math 240. Definitions and Notation.

Matrices. Math 240 Calculus III. Wednesday, July 10, Summer 2013, Session II. Matrices. Math 240. Definitions and Notation. function Matrices Calculus III Summer 2013, Session II Wednesday, July 10, 2013 Agenda function 1. 2. function function Definition An m n matrix is a rectangular array of numbers arranged in m horizontal

More information

B553 Lecture 5: Matrix Algebra Review

B553 Lecture 5: Matrix Algebra Review B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

Matrix Arithmetic. j=1

Matrix Arithmetic. j=1 An m n matrix is an array A = Matrix Arithmetic a 11 a 12 a 1n a 21 a 22 a 2n a m1 a m2 a mn of real numbers a ij An m n matrix has m rows and n columns a ij is the entry in the i-th row and j-th column

More information

Linear Algebra Solutions 1

Linear Algebra Solutions 1 Math Camp 1 Do the following: Linear Algebra Solutions 1 1. Let A = and B = 3 8 5 A B = 3 5 9 A + B = 9 11 14 4 AB = 69 3 16 BA = 1 4 ( 1 3. Let v = and u = 5 uv = 13 u v = 13 v u = 13 Math Camp 1 ( 7

More information

Tensor Analysis in Euclidean Space

Tensor Analysis in Euclidean Space Tensor Analysis in Euclidean Space James Emery Edited: 8/5/2016 Contents 1 Classical Tensor Notation 2 2 Multilinear Functionals 4 3 Operations With Tensors 5 4 The Directional Derivative 5 5 Curvilinear

More information

Matrices: 2.1 Operations with Matrices

Matrices: 2.1 Operations with Matrices Goals In this chapter and section we study matrix operations: Define matrix addition Define multiplication of matrix by a scalar, to be called scalar multiplication. Define multiplication of two matrices,

More information

ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, Please leave fractions as fractions, but simplify them, etc.

ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, Please leave fractions as fractions, but simplify them, etc. ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, 2015 1 Name: Solution Score: /100 This exam is closed-book. You must show ALL of your work for full credit. Please read the questions carefully.

More information

Linear Algebra (Review) Volker Tresp 2018

Linear Algebra (Review) Volker Tresp 2018 Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c

More information

r=1 r=1 argmin Q Jt (20) After computing the descent direction d Jt 2 dt H t d + P (x + d) d i = 0, i / J

r=1 r=1 argmin Q Jt (20) After computing the descent direction d Jt 2 dt H t d + P (x + d) d i = 0, i / J 7 Appendix 7. Proof of Theorem Proof. There are two main difficulties in proving the convergence of our algorithm, and none of them is addressed in previous works. First, the Hessian matrix H is a block-structured

More information

Multilinear Algebra For the Undergraduate Algebra Student

Multilinear Algebra For the Undergraduate Algebra Student Multilinear Algebra For the Undergraduate Algebra Student Davis Shurbert Department of Mathematics and Computer Science University of Puget Sound April 4, 204 Introduction When working in the field of

More information

Matrix representation of a linear map

Matrix representation of a linear map Matrix representation of a linear map As before, let e i = (0,..., 0, 1, 0,..., 0) T, with 1 in the i th place and 0 elsewhere, be standard basis vectors. Given linear map f : R n R m we get n column vectors

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Jim Lambers MAT 610 Summer Session Lecture 1 Notes

Jim Lambers MAT 610 Summer Session Lecture 1 Notes Jim Lambers MAT 60 Summer Session 2009-0 Lecture Notes Introduction This course is about numerical linear algebra, which is the study of the approximate solution of fundamental problems from linear algebra

More information

MATRICES. a m,1 a m,n A =

MATRICES. a m,1 a m,n A = MATRICES Matrices are rectangular arrays of real or complex numbers With them, we define arithmetic operations that are generalizations of those for real and complex numbers The general form a matrix of

More information

Lecture: Algorithms for LP, SOCP and SDP

Lecture: Algorithms for LP, SOCP and SDP 1/53 Lecture: Algorithms for LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:

More information

Chapter 2. Matrix Arithmetic. Chapter 2

Chapter 2. Matrix Arithmetic. Chapter 2 Matrix Arithmetic Matrix Addition and Subtraction Addition and subtraction act element-wise on matrices. In order for the addition/subtraction (A B) to be possible, the two matrices A and B must have the

More information

Final Project # 5 The Cartan matrix of a Root System

Final Project # 5 The Cartan matrix of a Root System 8.099 Final Project # 5 The Cartan matrix of a Root System Thomas R. Covert July 30, 00 A root system in a Euclidean space V with a symmetric positive definite inner product, is a finite set of elements

More information

Matrix Arithmetic. (Multidimensional Math) Institute for Advanced Analytics. Shaina Race, PhD

Matrix Arithmetic. (Multidimensional Math) Institute for Advanced Analytics. Shaina Race, PhD Matrix Arithmetic Multidimensional Math) Shaina Race, PhD Institute for Advanced Analytics. TABLE OF CONTENTS Element-wise Operations Linear Combinations of Matrices and Vectors. Vector Multiplication

More information

Finite Elements. Colin Cotter. February 22, Colin Cotter FEM

Finite Elements. Colin Cotter. February 22, Colin Cotter FEM Finite Elements February 22, 2019 In the previous sections, we introduced the concept of finite element spaces, which contain certain functions defined on a domain. Finite element spaces are examples of

More information

. a m1 a mn. a 1 a 2 a = a n

. a m1 a mn. a 1 a 2 a = a n Biostat 140655, 2008: Matrix Algebra Review 1 Definition: An m n matrix, A m n, is a rectangular array of real numbers with m rows and n columns Element in the i th row and the j th column is denoted by

More information

Bare minimum on matrix algebra. Psychology 588: Covariance structure and factor models

Bare minimum on matrix algebra. Psychology 588: Covariance structure and factor models Bare minimum on matrix algebra Psychology 588: Covariance structure and factor models Matrix multiplication 2 Consider three notations for linear combinations y11 y1 m x11 x 1p b11 b 1m y y x x b b n1

More information

Matrix Factorization and Analysis

Matrix Factorization and Analysis Chapter 7 Matrix Factorization and Analysis Matrix factorizations are an important part of the practice and analysis of signal processing. They are at the heart of many signal-processing algorithms. Their

More information

= 1 and 2 1. T =, and so det A b d

= 1 and 2 1. T =, and so det A b d Chapter 8 Determinants The founder of the theory of determinants is usually taken to be Gottfried Wilhelm Leibniz (1646 1716, who also shares the credit for inventing calculus with Sir Isaac Newton (1643

More information

Linear Algebra and Matrix Inversion

Linear Algebra and Matrix Inversion Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much

More information

MAT 610: Numerical Linear Algebra. James V. Lambers

MAT 610: Numerical Linear Algebra. James V. Lambers MAT 610: Numerical Linear Algebra James V Lambers January 16, 2017 2 Contents 1 Matrix Multiplication Problems 7 11 Introduction 7 111 Systems of Linear Equations 7 112 The Eigenvalue Problem 8 12 Basic

More information

2: LINEAR TRANSFORMATIONS AND MATRICES

2: LINEAR TRANSFORMATIONS AND MATRICES 2: LINEAR TRANSFORMATIONS AND MATRICES STEVEN HEILMAN Contents 1. Review 1 2. Linear Transformations 1 3. Null spaces, range, coordinate bases 2 4. Linear Transformations and Bases 4 5. Matrix Representation,

More information

Solvability of Linear Matrix Equations in a Symmetric Matrix Variable

Solvability of Linear Matrix Equations in a Symmetric Matrix Variable Solvability of Linear Matrix Equations in a Symmetric Matrix Variable Maurcio C. de Oliveira J. William Helton Abstract We study the solvability of generalized linear matrix equations of the Lyapunov type

More information

QUATERNIONS AND ROTATIONS

QUATERNIONS AND ROTATIONS QUATERNIONS AND ROTATIONS SVANTE JANSON 1. Introduction The purpose of this note is to show some well-known relations between quaternions and the Lie groups SO(3) and SO(4) (rotations in R 3 and R 4 )

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2018 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 2: Orthogonal Vectors and Matrices; Vector Norms Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 11 Outline 1 Orthogonal

More information

Basics of Calculus and Algebra

Basics of Calculus and Algebra Monika Department of Economics ISCTE-IUL September 2012 Basics of linear algebra Real valued Functions Differential Calculus Integral Calculus Optimization Introduction I A matrix is a rectangular array

More information

Basic Concepts in Linear Algebra

Basic Concepts in Linear Algebra Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University February 2, 2015 Grady B Wright Linear Algebra Basics February 2, 2015 1 / 39 Numerical Linear Algebra Linear

More information

Matrix Derivatives and Descent Optimization Methods

Matrix Derivatives and Descent Optimization Methods Matrix Derivatives and Descent Optimization Methods 1 Qiang Ning Department of Electrical and Computer Engineering Beckman Institute for Advanced Science and Techonology University of Illinois at Urbana-Champaign

More information

I = i 0,

I = i 0, Special Types of Matrices Certain matrices, such as the identity matrix 0 0 0 0 0 0 I = 0 0 0, 0 0 0 have a special shape, which endows the matrix with helpful properties The identity matrix is an example

More information

RANK AND PERIMETER PRESERVER OF RANK-1 MATRICES OVER MAX ALGEBRA

RANK AND PERIMETER PRESERVER OF RANK-1 MATRICES OVER MAX ALGEBRA Discussiones Mathematicae General Algebra and Applications 23 (2003 ) 125 137 RANK AND PERIMETER PRESERVER OF RANK-1 MATRICES OVER MAX ALGEBRA Seok-Zun Song and Kyung-Tae Kang Department of Mathematics,

More information

Functions of Several Variables

Functions of Several Variables Functions of Several Variables The Unconstrained Minimization Problem where In n dimensions the unconstrained problem is stated as f() x variables. minimize f()x x, is a scalar objective function of vector

More information

MATH 2030: MATRICES. Example 0.2. Q:Define A 1 =, A. 3 4 A: We wish to find c 1, c 2, and c 3 such that. c 1 + c c

MATH 2030: MATRICES. Example 0.2. Q:Define A 1 =, A. 3 4 A: We wish to find c 1, c 2, and c 3 such that. c 1 + c c MATH 2030: MATRICES Matrix Algebra As with vectors, we may use the algebra of matrices to simplify calculations. However, matrices have operations that vectors do not possess, and so it will be of interest

More information

Linear and Multilinear Algebra. Linear maps preserving rank of tensor products of matrices

Linear and Multilinear Algebra. Linear maps preserving rank of tensor products of matrices Linear maps preserving rank of tensor products of matrices Journal: Manuscript ID: GLMA-0-0 Manuscript Type: Original Article Date Submitted by the Author: -Aug-0 Complete List of Authors: Zheng, Baodong;

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2016 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

Matrices. 1 a a2 1 b b 2 1 c c π

Matrices. 1 a a2 1 b b 2 1 c c π Matrices 2-3-207 A matrix is a rectangular array of numbers: 2 π 4 37 42 0 3 a a2 b b 2 c c 2 Actually, the entries can be more general than numbers, but you can think of the entries as numbers to start

More information

Math 240 Calculus III

Math 240 Calculus III The Calculus III Summer 2015, Session II Wednesday, July 8, 2015 Agenda 1. of the determinant 2. determinants 3. of determinants What is the determinant? Yesterday: Ax = b has a unique solution when A

More information

Numerical Optimization

Numerical Optimization Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function

More information

Review of Basic Concepts in Linear Algebra

Review of Basic Concepts in Linear Algebra Review of Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University September 7, 2017 Math 565 Linear Algebra Review September 7, 2017 1 / 40 Numerical Linear Algebra

More information

5.6. PSEUDOINVERSES 101. A H w.

5.6. PSEUDOINVERSES 101. A H w. 5.6. PSEUDOINVERSES 0 Corollary 5.6.4. If A is a matrix such that A H A is invertible, then the least-squares solution to Av = w is v = A H A ) A H w. The matrix A H A ) A H is the left inverse of A and

More information

EE364a Review Session 7

EE364a Review Session 7 EE364a Review Session 7 EE364a Review session outline: derivatives and chain rule (Appendix A.4) numerical linear algebra (Appendix C) factor and solve method exploiting structure and sparsity 1 Derivative

More information

Review of Linear Algebra

Review of Linear Algebra Review of Linear Algebra VBS/MRC Review of Linear Algebra 0 Ok, the Questions Why does algebra sound like cobra? Why do they say abstract algebra? Why bother about all this abstract stuff? What is a vector,

More information

Announcements Wednesday, October 10

Announcements Wednesday, October 10 Announcements Wednesday, October 10 The second midterm is on Friday, October 19 That is one week from this Friday The exam covers 35, 36, 37, 39, 41, 42, 43, 44 (through today s material) WeBWorK 42, 43

More information

POSITIVE MAP AS DIFFERENCE OF TWO COMPLETELY POSITIVE OR SUPER-POSITIVE MAPS

POSITIVE MAP AS DIFFERENCE OF TWO COMPLETELY POSITIVE OR SUPER-POSITIVE MAPS Adv. Oper. Theory 3 (2018), no. 1, 53 60 http://doi.org/10.22034/aot.1702-1129 ISSN: 2538-225X (electronic) http://aot-math.org POSITIVE MAP AS DIFFERENCE OF TWO COMPLETELY POSITIVE OR SUPER-POSITIVE MAPS

More information

Matrix Multiplication

Matrix Multiplication 3.2 Matrix Algebra Matrix Multiplication Example Foxboro Stadium has three main concession stands, located behind the south, north and west stands. The top-selling items are peanuts, hot dogs and soda.

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite

More information

Announcements Monday, October 02

Announcements Monday, October 02 Announcements Monday, October 02 Please fill out the mid-semester survey under Quizzes on Canvas WeBWorK 18, 19 are due Wednesday at 11:59pm The quiz on Friday covers 17, 18, and 19 My office is Skiles

More information

Matrix Algebra: Summary

Matrix Algebra: Summary May, 27 Appendix E Matrix Algebra: Summary ontents E. Vectors and Matrtices.......................... 2 E.. Notation.................................. 2 E..2 Special Types of Vectors.........................

More information

Sufficiency of Signed Principal Minors for Semidefiniteness: A Relatively Easy Proof

Sufficiency of Signed Principal Minors for Semidefiniteness: A Relatively Easy Proof Sufficiency of Signed Principal Minors for Semidefiniteness: A Relatively Easy Proof David M. Mandy Department of Economics University of Missouri 118 Professional Building Columbia, MO 65203 USA mandyd@missouri.edu

More information

Lecture 7 Spectral methods

Lecture 7 Spectral methods CSE 291: Unsupervised learning Spring 2008 Lecture 7 Spectral methods 7.1 Linear algebra review 7.1.1 Eigenvalues and eigenvectors Definition 1. A d d matrix M has eigenvalue λ if there is a d-dimensional

More information

18.S34 linear algebra problems (2007)

18.S34 linear algebra problems (2007) 18.S34 linear algebra problems (2007) Useful ideas for evaluating determinants 1. Row reduction, expanding by minors, or combinations thereof; sometimes these are useful in combination with an induction

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2017

Cheng Soon Ong & Christian Walder. Canberra February June 2017 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2017 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 141 Part III

More information

On Expected Gaussian Random Determinants

On Expected Gaussian Random Determinants On Expected Gaussian Random Determinants Moo K. Chung 1 Department of Statistics University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 Abstract The expectation of random determinants whose

More information

Mathematical Foundations of Applied Statistics: Matrix Algebra

Mathematical Foundations of Applied Statistics: Matrix Algebra Mathematical Foundations of Applied Statistics: Matrix Algebra Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/105 Literature Seber, G.

More information

Near-Isometry by Relaxation: Supplement

Near-Isometry by Relaxation: Supplement 000 00 00 003 004 005 006 007 008 009 00 0 0 03 04 05 06 07 08 09 00 0 0 03 04 05 06 07 08 09 030 03 03 033 034 035 036 037 038 039 040 04 04 043 044 045 046 047 048 049 050 05 05 053 Near-Isometry by

More information

b 1 b 2.. b = b m A = [a 1,a 2,...,a n ] where a 1,j a 2,j a j = a m,j Let A R m n and x 1 x 2 x = x n

b 1 b 2.. b = b m A = [a 1,a 2,...,a n ] where a 1,j a 2,j a j = a m,j Let A R m n and x 1 x 2 x = x n Lectures -2: Linear Algebra Background Almost all linear and nonlinear problems in scientific computation require the use of linear algebra These lectures review basic concepts in a way that has proven

More information

MAC Module 2 Systems of Linear Equations and Matrices II. Learning Objectives. Upon completing this module, you should be able to :

MAC Module 2 Systems of Linear Equations and Matrices II. Learning Objectives. Upon completing this module, you should be able to : MAC 0 Module Systems of Linear Equations and Matrices II Learning Objectives Upon completing this module, you should be able to :. Find the inverse of a square matrix.. Determine whether a matrix is invertible..

More information

Vector, Matrix, and Tensor Derivatives

Vector, Matrix, and Tensor Derivatives Vector, Matrix, and Tensor Derivatives Erik Learned-Miller The purpose of this document is to help you learn to take derivatives of vectors, matrices, and higher order tensors (arrays with three dimensions

More information

Matrix Algebra. Matrix Algebra. Chapter 8 - S&B

Matrix Algebra. Matrix Algebra. Chapter 8 - S&B Chapter 8 - S&B Algebraic operations Matrix: The size of a matrix is indicated by the number of its rows and the number of its columns. A matrix with k rows and n columns is called a k n matrix. The number

More information

Institute for Computational Mathematics Hong Kong Baptist University

Institute for Computational Mathematics Hong Kong Baptist University Institute for Computational Mathematics Hong Kong Baptist University ICM Research Report 08-0 How to find a good submatrix S. A. Goreinov, I. V. Oseledets, D. V. Savostyanov, E. E. Tyrtyshnikov, N. L.

More information

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018 Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry

More information

Linear algebra 2. Yoav Zemel. March 1, 2012

Linear algebra 2. Yoav Zemel. March 1, 2012 Linear algebra 2 Yoav Zemel March 1, 2012 These notes were written by Yoav Zemel. The lecturer, Shmuel Berger, should not be held responsible for any mistake. Any comments are welcome at zamsh7@gmail.com.

More information

Elementary maths for GMT

Elementary maths for GMT Elementary maths for GMT Linear Algebra Part 2: Matrices, Elimination and Determinant m n matrices The system of m linear equations in n variables x 1, x 2,, x n a 11 x 1 + a 12 x 2 + + a 1n x n = b 1

More information

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 8, AUGUST /$ IEEE

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 8, AUGUST /$ IEEE IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 8, AUGUST 2009 3613 Hessian and Concavity of Mutual Information, Differential Entropy, and Entropy Power in Linear Vector Gaussian Channels Miquel

More information

A new product on 2 2 matrices

A new product on 2 2 matrices A new product on 2 2 matrices L Kramer, P Kramer, and V Man ko October 17, 2018 arxiv:1802.03048v1 [math-ph] 8 Feb 2018 Abstract We study a bilinear multiplication rule on 2 2 matrices which is intermediate

More information

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic A FIRST COURSE IN LINEAR ALGEBRA An Open Text by Ken Kuttler Matrix Arithmetic Lecture Notes by Karen Seyffarth Adapted by LYRYX SERVICE COURSE SOLUTION Attribution-NonCommercial-ShareAlike (CC BY-NC-SA)

More information

Sparse Levenberg-Marquardt algorithm.

Sparse Levenberg-Marquardt algorithm. Sparse Levenberg-Marquardt algorithm. R. I. Hartley and A. Zisserman: Multiple View Geometry in Computer Vision. Cambridge University Press, second edition, 2004. Appendix 6 was used in part. The Levenberg-Marquardt

More information

Optimization Tutorial 1. Basic Gradient Descent

Optimization Tutorial 1. Basic Gradient Descent E0 270 Machine Learning Jan 16, 2015 Optimization Tutorial 1 Basic Gradient Descent Lecture by Harikrishna Narasimhan Note: This tutorial shall assume background in elementary calculus and linear algebra.

More information

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Topics. Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij

Topics. Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij Topics Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij or a ij lives in row i and column j Definition of a matrix

More information

Matrices and Vectors

Matrices and Vectors Matrices and Vectors James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 11, 2013 Outline 1 Matrices and Vectors 2 Vector Details 3 Matrix

More information

ARTICLE IN PRESS. Journal of Multivariate Analysis ( ) Contents lists available at ScienceDirect. Journal of Multivariate Analysis

ARTICLE IN PRESS. Journal of Multivariate Analysis ( ) Contents lists available at ScienceDirect. Journal of Multivariate Analysis Journal of Multivariate Analysis ( ) Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Marginal parameterizations of discrete models

More information

Matrix Operations: Determinant

Matrix Operations: Determinant Matrix Operations: Determinant Determinants Determinants are only applicable for square matrices. Determinant of the square matrix A is denoted as: det(a) or A Recall that the absolute value of the determinant

More information

Accelerated Block-Coordinate Relaxation for Regularized Optimization

Accelerated Block-Coordinate Relaxation for Regularized Optimization Accelerated Block-Coordinate Relaxation for Regularized Optimization Stephen J. Wright Computer Sciences University of Wisconsin, Madison October 09, 2012 Problem descriptions Consider where f is smooth

More information

(Part 1) High-dimensional statistics May / 41

(Part 1) High-dimensional statistics May / 41 Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2

More information

Strongly Regular Decompositions of the Complete Graph

Strongly Regular Decompositions of the Complete Graph Journal of Algebraic Combinatorics, 17, 181 201, 2003 c 2003 Kluwer Academic Publishers. Manufactured in The Netherlands. Strongly Regular Decompositions of the Complete Graph EDWIN R. VAN DAM Edwin.vanDam@uvt.nl

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

Linear Algebra V = T = ( 4 3 ).

Linear Algebra V = T = ( 4 3 ). Linear Algebra Vectors A column vector is a list of numbers stored vertically The dimension of a column vector is the number of values in the vector W is a -dimensional column vector and V is a 5-dimensional

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

DOMINO TILING. Contents 1. Introduction 1 2. Rectangular Grids 2 Acknowledgments 10 References 10

DOMINO TILING. Contents 1. Introduction 1 2. Rectangular Grids 2 Acknowledgments 10 References 10 DOMINO TILING KASPER BORYS Abstract In this paper we explore the problem of domino tiling: tessellating a region with x2 rectangular dominoes First we address the question of existence for domino tilings

More information

THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION

THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION THE SINGULAR VALUE DECOMPOSITION AND LOW RANK APPROXIMATION MANTAS MAŽEIKA Abstract. The purpose of this paper is to present a largely self-contained proof of the singular value decomposition (SVD), and

More information

April 26, Applied mathematics PhD candidate, physics MA UC Berkeley. Lecture 4/26/2013. Jed Duersch. Spd matrices. Cholesky decomposition

April 26, Applied mathematics PhD candidate, physics MA UC Berkeley. Lecture 4/26/2013. Jed Duersch. Spd matrices. Cholesky decomposition Applied mathematics PhD candidate, physics MA UC Berkeley April 26, 2013 UCB 1/19 Symmetric positive-definite I Definition A symmetric matrix A R n n is positive definite iff x T Ax > 0 holds x 0 R n.

More information

Ma/CS 6b Class 20: Spectral Graph Theory

Ma/CS 6b Class 20: Spectral Graph Theory Ma/CS 6b Class 20: Spectral Graph Theory By Adam Sheffer Eigenvalues and Eigenvectors A an n n matrix of real numbers. The eigenvalues of A are the numbers λ such that Ax = λx for some nonzero vector x

More information

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming CSC2411 - Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming Notes taken by Mike Jamieson March 28, 2005 Summary: In this lecture, we introduce semidefinite programming

More information

Elementary Row Operations on Matrices

Elementary Row Operations on Matrices King Saud University September 17, 018 Table of contents 1 Definition A real matrix is a rectangular array whose entries are real numbers. These numbers are organized on rows and columns. An m n matrix

More information

Matrix Differential Calculus with Applications in Statistics and Econometrics

Matrix Differential Calculus with Applications in Statistics and Econometrics Matrix Differential Calculus with Applications in Statistics and Econometrics Revised Edition JAN. R. MAGNUS CentERjor Economic Research, Tilburg University and HEINZ NEUDECKER Cesaro, Schagen JOHN WILEY

More information

ICS 6N Computational Linear Algebra Matrix Algebra

ICS 6N Computational Linear Algebra Matrix Algebra ICS 6N Computational Linear Algebra Matrix Algebra Xiaohui Xie University of California, Irvine xhx@uci.edu February 2, 2017 Xiaohui Xie (UCI) ICS 6N February 2, 2017 1 / 24 Matrix Consider an m n matrix

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a beginning course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc I will review some of these terms here,

More information

Mathematics 13: Lecture 10

Mathematics 13: Lecture 10 Mathematics 13: Lecture 10 Matrices Dan Sloughter Furman University January 25, 2008 Dan Sloughter (Furman University) Mathematics 13: Lecture 10 January 25, 2008 1 / 19 Matrices Recall: A matrix is a

More information

Multivariable Calculus

Multivariable Calculus 2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)

More information

1 Matrices and Systems of Linear Equations

1 Matrices and Systems of Linear Equations March 3, 203 6-6. Systems of Linear Equations Matrices and Systems of Linear Equations An m n matrix is an array A = a ij of the form a a n a 2 a 2n... a m a mn where each a ij is a real or complex number.

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 1: Course Overview; Matrix Multiplication Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information