Orthogonal tensor decomposition

Size: px
Start display at page:

Download "Orthogonal tensor decomposition"

Transcription

1 Orthogonal tensor decomposition Daniel Hsu Columbia University Largely based on 2012 arxiv report Tensor decompositions for learning latent variable models, with Anandkumar, Ge, Kakade, and Telgarsky. 1

2 The basic decomposition problem Notation: For a vector x = (x 1, x 2,..., x n ) R n, x x x denotes the 3-way array (call it a tensor ) in R n n n (i, j, k) th entry is x i x j x k. whose Problem: Given T R n n n with the promise that T = n λ t v t v t v t for some orthonormal basis { v t } of R n (w.r.t. standard inner product) and positive scalars {λ t > 0}, approximately find {( v t, λ t )} (up to some desired precision). 3

3 Basic questions 1. Is {( v t, λ t )} uniquely determined? 2. If so, is there an efficient algorithm for finding the decomposition? 3. What if T is perturbed by some small amount? Perturbed problem: Same as the original problem, except instead of T, we are given T + E for some error tensor E. How large can E be if we want ε precision? 4

4 Analogous matrix problem Matrix problem: Given M R n n with the promise that M = n λ t v t v t for some orthonormal basis { v t } of R n (w.r.t. standard inner product) and positive scalars {λ t > 0}, approximately find {( v t, λ t )} (up to some desired precision). 5

5 Analogous matrix problem We re promised that M is symmetric and positive definite, so requested decomposition is an eigendecomposition. In this case, an eigendecomposition always exists, and can be found efficiently. It is unique if and only if the {λ i } are distinct. What if M is perturbed by some small amount? Perturbed matrix problem: Same as the original problem, except instead of M, we are given M + E for some error matrix E (assume to be symmetric). Answer provided by matrix perturbation theory (e.g., Davis-Kahan), which requires E 2 < min i j λ i λ j. 6

6 Back to the original problem Problem: Given T R n n n with the promise that T = n λ t v t v t v t for some orthonormal basis { v t } of R n (w.r.t. standard inner product) and positive scalars {λ t > 0}, approximately find {( v t, λ t )} (up to some desired precision). Such decompositions do not necessarily exist, even for symmetric tensors. Where the decompositions do exist, the Perturbed problem asks if they are robust. 7

7 Main ideas Easy claim: Repeated application of a certain quadratic operator based on T (a power iteration ) recovers a single ( v t, λ t ) up to any desired precision. Self-reduction: Replace T with T λ t v t v t v t. Why?: T λ t v t v t v t = τ t λ τ v τ v τ v τ. Catch: We don t recover ( v t, λ t ) exactly, so we actually can only replace T with for some error tensor E t. T λ t v t v t v t + E t Therefore, must anyway deal with perturbations. 8

8 Rest of this talk 1. Identifiability of decomposition {( v t, λ t )} from T. 2. A decomposition algorithm based on tensor power iteration. 3. Error analysis of decomposition algorithm. 9

9 Identifiability of the decomposition Orthonormal basis { v t } of R n, positive scalars {λ t > 0}: T = n λ t v t v t v t In what sense is {( v t, λ t )} uniquely determined? Claim: { v t } are the n isolated local maximizers of certain cubic form f T : B n R, and f T ( v t ) = λ t. 11

10 Aside: multilinear form There is a natural trilinear form associated with T : ( x, y, z) i,j,k T i,j,k x i y j z k. For matrices M, it looks like ( x, y) i,j M i,j x i y j = x M y. 12

11 Review: Rayleigh quotient Recall Rayleigh quotient for matrix M := n λ t v t v t (assuming x S n 1 ): R M ( x) := x M x = n λ t ( v t x) 2. Every v t such that λ t = max! is a maximizer of R M. (These are also the only local maximizers.) 13

12 The natural cubic form Consider the function f T : B n R given by x f T ( x) = i,j,k T i,j,k x i x j x k. For our promised T = n λ t v t v t v t, f T becomes f T ( x) = = = n n λ t λ t ( ) v t v t v t i,j,k i,j,k x ix j x k ( v t ) i ( v t ) j ( v t ) k x i x j x k i,j,k n λ t ( v t x) 3. Observation: f T ( v t ) = λ t. 14

13 Variational characterization Claim: Isolated local maximizers of f T on B n are { v t }. Objective function (with constraint): x inf λ 0 n λ t ( v t x) 3 1.5λ( x 2 2 1). First-order condition for local maxima: n λ t ( v t x) 2 v t = λ x. Second-order condition for isolated local maxima: w (2 n ) λ t ( v t x) v t v t λi w < 0, w x. 15

14 Intuition behind variational characterization May as well assume v t is t th coordinate basis vector, so max f T ( x) = x R n n λ t xt 3 s.t. n xt 2 1. Intuition: Suppose supp( x) = {1, 2}, and x 1, x 2 > 0. f T ( x) = λ 1 x λ 2x 3 2 < λ 1x λ 2x 2 2 max{λ 1, λ 2 }. Better to have supp( x) = 1, i.e., picking x to be a coordinate basis vector. 16

15 Aside: canonical polyadic decomposition Rank-K canonical polyadic decomposition (CPD) of T (also called PARAFAC, CANDECOMP, or CP): T = K σ i u i v i w i. i=1 [Harshman/Jennrich, 1970; Kruskal, 1977; Leurgans et al., 1993]. Number of parameters: K (3n + 1) (compared to n 3 in general). Fact: Our promised T has a rank-n CPD. N.B.: Overcomplete (K > n) CPD is also interesting and a possibility as long as K (3n + 1) n 3. 17

16 The quadratic operator Easy claim: Repeated application of a certain quadratic operator (based on T ) recovers a single (λ t, v t ) up to any desired precision. For any T R n n n and x = (x 1, x 2,..., x n ) R n, define the quadratic operator φ T ( x) := i,j,k T i,j,k x j x k e i R n where e i R n is the i th coordinate basis vector. n n ( If T = λ t v t v t v t, then φ T ( x) = λ t v t x )2 v t. 19

17 An algorithm? Recall: First-order condition for local maxima of f T ( x) = n λ t ( v t x) 3 for x B n : φ T ( x) = n λ t ( v t x) 2 v t = λ x i.e., eigenvector -like condition. Algorithm: Find x B n fixed under x φ T ( x)/ φ T ( x). (Ignoring numerical issues, can just repeatedly apply φ T and defer normalization until later.) N.B.: Gradient ascent also works [Kolda & Mayo, 11]. 20

18 Tensor power iteration [De Lathauwer et al, 2000] Start with some x (0), and for j = 1, 2,... : x (j) := φ T ( x (j 1)) = n ( λ t v t x (j 1))2 v t. Claim: For almost all initial x (0), the sequence ( x (j) / x (j) ) j=1 converges quadratically fast to some v t. 21

19 Review: matrix power iteration Recall matrix power iteration for matrix M := n λ t v t v t : Start with some x (0), and for j = 1, 2,... : x (i) := M x (j 1) = n ( λ t v t x (j 1)) v t. i.e., component in v t direction is scaled by λ t. If λ 1 > λ 2, then ( v 1 x (j)) 2 ( ) 2j n ( v t x (j)) 2 1 k λ2. λ 1 i.e., converges linearly to v 1 (assuming gap λ 2 /λ 1 < 1). 22

20 Tensor power iteration convergence analysis Let c t := v t x (0) (initial component in v t direction); assume WLOG λ 1 c 1 > λ 2 c 2 λ 3 c 3. Then n x (1) ( = λ t n v t x (0))2 v t = λ t ct 2 v t i.e., component in v t direction is squared then scaled by λ t. Easy to show ( v 1 x (j)) 2 ( ) ( v t x (j)) 2 1 k λ 2 1 λ 2 c 2 max t 1 λ t λ 1 c 1 n 2j+1. 23

21 Example n = 1024, λ t u.a.r. [0, 1] Value of ( v t x (0) ) 2 for t = 1, 2,...,

22 Example n = 1024, λ t u.a.r. [0, 1] Value of ( v t x (1) ) 2 for t = 1, 2,...,

23 Example n = 1024, λ t u.a.r. [0, 1] Value of ( v t x (2) ) 2 for t = 1, 2,...,

24 Example n = 1024, λ t u.a.r. [0, 1] Value of ( v t x (3) ) 2 for t = 1, 2,...,

25 Example n = 1024, λ t u.a.r. [0, 1] Value of ( v t x (4) ) 2 for t = 1, 2,...,

26 Example n = 1024, λ t u.a.r. [0, 1] Value of ( v t x (5) ) 2 for t = 1, 2,...,

27 Matrix vs. tensor power iteration Matrix power iteration: 1. Requires gap between largest and second-largest λ t. (Property of the matrix only.) 2. Converges to top v t. 3. Linear convergence. (Need O(log(1/ɛ)) iterations.) Tensor power iteration: 1. Requires gap between largest and second-largest λ t c t. (Property of the tensor and initialization x (0).) 2. Converges to v t for which λ t c t = max! (could be any of them). 3. Quadratic convergence. (Need O(log log(1/ɛ)) iterations.) 30

28 Initialization of tensor power iteration Convergence of tensor power iteration requires gap between largest and second-largest λ t v t x (0). Example of bad initialization: Suppose T = t v t v t v t, and x (0) = 1 2 ( v 1 + v 2 ). φ T ( x (0) ) = ( v 1 x (0) ) 2 v 1 + ( v 2 x (0) ) 2 v 2 = 1 2 ( v 1 + v 2 ) = 1 x (0). 2 Fortunately, bad initialization points are atypical. 31

29 Full decomposition algorithm Input: T R n n n. Initialize: T := T. For i = 1, 2,..., n: 1. Pick x (0) S n 1 unif. at random. 2. Run tensor power iteration with T starting from x (0) for N iterations. 3. Set ˆv i := x (N) / x (N) and ˆλ i := f T (ˆv i ). 4. Replace T := T ˆλ i ˆv i ˆv i ˆv i. Output: {(ˆv i, ˆλ i ) : i [n]}. Actually: repeat Steps 1 3 several times, and take results of trial yielding largest ˆλ i. 32

30 Aside: direct minimization Can also consider directly minimizing n 2 T ˆλ t ˆv t ˆv t ˆv t via local optimization (e.g., coord. descent, alternating least squares). Decomposition algorithm via tensor power iteration can be viewed as orthgonal greedy algorithm for minimizing above objective [Zhang & Golub, 01]. F 33

31 Aside: implementation for bag-of-words models Let f (i) be empirical word frequency vector for document i: ( f (i) ) j = # times word j appears in document i length of document i Matrix of word-pair frequencies (from m documents) Pairs 1 m m f (i) f (i) i=1 K µ t µ t. Tensor of word-triple frequencies (from m documents) Triples 1 m m f (i) f (i) f (i) i=1 K µ t µ t µ t. 34

32 Aside: implementation for bag-of-words models Use inner product system given by x, y := x Pairs y. Why?: If Pairs = K µ t µ t, then µ i, µ j = 1 {i=j}. { µ i } are orthonormal under this inner product system. Power iteration step: φ Triples ( x) := 1 m m x, f (i) 2 f (i) = 1 m m ( x Pairs f (i) ) 2 f (i). i=1 i=1 1. First compute y := Pairs x (use low-rank factors of Pairs). 2. Then compute ( y f (i) ) 2 f (i) for all documents i, and add them up (all sparse operations). Final running time # topics (model size + input size). 35

33 Effect of errors in tensor power iterations Suppose we are given T := T + E, with n T = λ t v t v t v t, ε := sup x S n 1 φ E ( x). What can we say about the resulting ˆv i and ˆλ i? 37

34 Perturbation analysis Theorem: If ε O( min t λ t n ), then with high probability, a modified variant of the full decomposition algorithm returns {(ˆv i, ˆλ i ) : i [n]} with ˆv i v i O(ε/λ i ), ˆλ i λ i O(ε), i [n]. Essentially third-order analogue of Wedin s theorem for SVD of matrices, but specific to fixed-point iteration algorithm. Similar analysis holds for variational characterization. 38

35 Effect of errors in tensor power iterations Quadratic operator φ T with T : φ T ( x) = n ( λ t v t x )2 v t + φ E ( x). Claim: If ε O( min t λ t n ) and N Ω(log(n) + log log max t λ t ε ), then N steps of tensor power iteration on T + E (with good initialization) gives ˆv i v i O(ε/λ i ), ˆλ i λ i O(ε). 39

36 Deflation (For simplicity, assume λ 1 = = λ n = 1.) Using tensor power iteration on T := T + E: Approximate (say) v 1 with ˆv 1 up to error v 1 ˆv 1 ε. Deflation danger: To find next v t, use T ˆv 1 ˆv 1 ˆv 1 = n v t v t v t t=2 + E + ( v 1 v 1 v 1 ˆv 1 ˆv 1 ˆv 1 ). Now error seems to be of size 2ε... exponential explosion? 40

37 How do the errors look? E 1 := v 1 v 1 v 1 ˆv 1 ˆv 1 ˆv 1 Take any direction x orthogonal to v 1 : φ E 1 ( x) = ( v 1 x) 2 v 1 (ˆv 1 x) 2ˆv 1 = (ˆv 1 x) 2ˆv 1 = ((ˆv 1 v 1 ) x) 2 ˆv 1 v 1 2 ε 2. Effect of E + E 1 in directions orthogonal to v 1 is just (1 + o(1))ε. 41

38 Deflation analysis Upshot: all errors due to deflation have only lower-order effects on ability to find subsequent v t. Analogous statement for matrix power iteration is not true. 42

39 Recap and remarks Orthogonally diagonalizable tensors have very nice identifiability, computational, and robustness properties. Many analogues to matrix SVD, but also many important differences arising from non-linearity. Greedy algorithm for finding the decomposition can be rigorously analyzed and shown to be effective and efficient. Many other approaches to moment-based estimation (e.g., subspace ID / OOMs, local optimization). 44

40 Other stuff I didn t talk about 1. Overcomplete tensor decomposition: K > n components in R n. K T = λ t v t v t v t. ICA/blind source separation [Cardoso, 1991; Goyal et al, 2014] Mixture models [Bhaskara et al, 2014; Anderson et al, 2014] Dictionary learning [Barak et al, 2014] General Tucker decompositions (CPD is a special case). Exploit other structure (e.g., sparsity) Talk to Anima about this! 45

41 Questions? 46

42 Tensor product of vector spaces What is the tensor product V W of vector spaces V and W? Define objects E v, w for v V and w W. Declare equivalences E v1 + v 2, w E v1, w + E v2, w E v, w1 + w 2 E v, w1 + E v, w2 c E v, w E c v, w E v,c w for c R. Pick any bases B V for V, and B W for W. V W := span of {E v, w : v B V, w B W }, modulo equivalences (eliminating dependence on choice of bases). Can check that V W is a vector space. v w (tensor product of v V and w W ) is the equivalence class of E v, w in V W. 48

43 Tensor algebra perspective From tensor algebra: Since { v t : t [n] } is a basis for R n, { } v i v j v k : i, j, k [n] is a basis for R n R n R n ( denotes the tensor product of vector spaces) Every tensor T R n R n R n has a unique representation in this basis: T = c i,j,k v i v j v k i,j,k N.B.: dim(r n R n R n ) = n 3. 49

44 Aside: general bases for R n R n R n Pick any bases ( { α i }, { β i }, { γ i } ) for R n (not necessary orthonormal). Basis for R n R n R n : { αi β j γ k : 1 i, j, k n }. Every tensor T R n R n R n has a unique representation in this basis: T = c i,j,k α i β j γ k. i,j,k A tensor T such that c i,j,k 0 i = j = k is called diagonal: T = n c i,i,i α i β i γ i. i=1 Claim: A tensor T can be diagonal w.r.t. at most one basis. 50

45 Aside: canonical polyadic decomposition Rank-K canonical polyadic decomposition (CPD) of T (also called PARAFAC, CANDECOMP, or CP): T = K σ i u i v i w i. i=1 Number of parameters: K (3n + 1) (compared to n 3 in general). Fact: If T is diagonal w.r.t. bases then it has a rank-k CPD with K n. Diagonal w.r.t. bases non-overcomplete CPD. N.B.: Overcomplete (K > n) CPD is also interesting and a possibility as long as K (3n + 1) n 3. 51

46 Initialization of tensor power iteration Let t max := arg max t λ t, and draw x (0) S n 1 unif. at random. Most coefficients of x (0) are around 1/ n; largest is around log(n)/n. Almost surely, a gap exists: max t t max λ t v t x (0) λ tmax v tmax x (0) < 1. With probability 1/n 1.2, the gap is non-negligible: max t t max λ t v t x (0) λ tmax v tmax x (0) < 0.9. Try O(n 1.3 ) initializers; chances are at least one is good. (Very conservative estimate only; can be much better than this.) 53

A Robust Form of Kruskal s Identifiability Theorem

A Robust Form of Kruskal s Identifiability Theorem A Robust Form of Kruskal s Identifiability Theorem Aditya Bhaskara (Google NYC) Joint work with Moses Charikar (Princeton University) Aravindan Vijayaraghavan (NYU Northwestern) Background: understanding

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Volodymyr Kuleshov ú Arun Tejasvi Chaganty ú Percy Liang. May 11, 2015

Volodymyr Kuleshov ú Arun Tejasvi Chaganty ú Percy Liang. May 11, 2015 Tensor Factorization via Matrix Factorization Volodymyr Kuleshov ú Arun Tejasvi Chaganty ú Percy Liang Stanford University May 11, 2015 Kuleshov, Chaganty, Liang (Stanford University) Tensor Factorization

More information

Non-convex Robust PCA: Provable Bounds

Non-convex Robust PCA: Provable Bounds Non-convex Robust PCA: Provable Bounds Anima Anandkumar U.C. Irvine Joint work with Praneeth Netrapalli, U.N. Niranjan, Prateek Jain and Sujay Sanghavi. Learning with Big Data High Dimensional Regime Missing

More information

Tensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia

Tensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia Network Feature s Decompositions for Machine Learning 1 1 Department of Computer Science University of British Columbia UBC Machine Learning Group, June 15 2016 1/30 Contact information Network Feature

More information

A concise proof of Kruskal s theorem on tensor decomposition

A concise proof of Kruskal s theorem on tensor decomposition A concise proof of Kruskal s theorem on tensor decomposition John A. Rhodes 1 Department of Mathematics and Statistics University of Alaska Fairbanks PO Box 756660 Fairbanks, AK 99775 Abstract A theorem

More information

Donald Goldfarb IEOR Department Columbia University UCLA Mathematics Department Distinguished Lecture Series May 17 19, 2016

Donald Goldfarb IEOR Department Columbia University UCLA Mathematics Department Distinguished Lecture Series May 17 19, 2016 Optimization for Tensor Models Donald Goldfarb IEOR Department Columbia University UCLA Mathematics Department Distinguished Lecture Series May 17 19, 2016 1 Tensors Matrix Tensor: higher-order matrix

More information

ECE 598: Representation Learning: Algorithms and Models Fall 2017

ECE 598: Representation Learning: Algorithms and Models Fall 2017 ECE 598: Representation Learning: Algorithms and Models Fall 2017 Lecture 1: Tensor Methods in Machine Learning Lecturer: Pramod Viswanathan Scribe: Bharath V Raghavan, Oct 3, 2017 11 Introduction Tensors

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 22 1 / 21 Overview

More information

Dictionary Learning Using Tensor Methods

Dictionary Learning Using Tensor Methods Dictionary Learning Using Tensor Methods Anima Anandkumar U.C. Irvine Joint work with Rong Ge, Majid Janzamin and Furong Huang. Feature learning as cornerstone of ML ML Practice Feature learning as cornerstone

More information

Guaranteed Learning of Latent Variable Models through Spectral and Tensor Methods

Guaranteed Learning of Latent Variable Models through Spectral and Tensor Methods Guaranteed Learning of Latent Variable Models through Spectral and Tensor Methods Anima Anandkumar U.C. Irvine Application 1: Clustering Basic operation of grouping data points. Hypothesis: each data point

More information

TENSOR LAYERS FOR COMPRESSION OF DEEP LEARNING NETWORKS. Cris Cecka Senior Research Scientist, NVIDIA GTC 2018

TENSOR LAYERS FOR COMPRESSION OF DEEP LEARNING NETWORKS. Cris Cecka Senior Research Scientist, NVIDIA GTC 2018 TENSOR LAYERS FOR COMPRESSION OF DEEP LEARNING NETWORKS Cris Cecka Senior Research Scientist, NVIDIA GTC 2018 Tensors Computations and the GPU AGENDA Tensor Networks and Decompositions Tensor Layers in

More information

arxiv: v1 [math.ra] 13 Jan 2009

arxiv: v1 [math.ra] 13 Jan 2009 A CONCISE PROOF OF KRUSKAL S THEOREM ON TENSOR DECOMPOSITION arxiv:0901.1796v1 [math.ra] 13 Jan 2009 JOHN A. RHODES Abstract. A theorem of J. Kruskal from 1977, motivated by a latent-class statistical

More information

Lecture 8 : Eigenvalues and Eigenvectors

Lecture 8 : Eigenvalues and Eigenvectors CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with

More information

Linear Algebra (Review) Volker Tresp 2017

Linear Algebra (Review) Volker Tresp 2017 Linear Algebra (Review) Volker Tresp 2017 1 Vectors k is a scalar (a number) c is a column vector. Thus in two dimensions, c = ( c1 c 2 ) (Advanced: More precisely, a vector is defined in a vector space.

More information

Linear Algebra (Review) Volker Tresp 2018

Linear Algebra (Review) Volker Tresp 2018 Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c

More information

https://goo.gl/kfxweg KYOTO UNIVERSITY Statistical Machine Learning Theory Sparsity Hisashi Kashima kashima@i.kyoto-u.ac.jp DEPARTMENT OF INTELLIGENCE SCIENCE AND TECHNOLOGY 1 KYOTO UNIVERSITY Topics:

More information

THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING

THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING Luis Rademacher, Ohio State University, Computer Science and Engineering. Joint work with Mikhail Belkin and James Voss This talk A new approach to multi-way

More information

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric

More information

The Eigenvalue Problem: Perturbation Theory

The Eigenvalue Problem: Perturbation Theory Jim Lambers MAT 610 Summer Session 2009-10 Lecture 13 Notes These notes correspond to Sections 7.2 and 8.1 in the text. The Eigenvalue Problem: Perturbation Theory The Unsymmetric Eigenvalue Problem Just

More information

Structured matrix factorizations. Example: Eigenfaces

Structured matrix factorizations. Example: Eigenfaces Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Definition (T -invariant subspace) Example. Example

Definition (T -invariant subspace) Example. Example Eigenvalues, Eigenvectors, Similarity, and Diagonalization We now turn our attention to linear transformations of the form T : V V. To better understand the effect of T on the vector space V, we begin

More information

Applied Linear Algebra in Geoscience Using MATLAB

Applied Linear Algebra in Geoscience Using MATLAB Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in

More information

An Introduction to Spectral Learning

An Introduction to Spectral Learning An Introduction to Spectral Learning Hanxiao Liu November 8, 2013 Outline 1 Method of Moments 2 Learning topic models using spectral properties 3 Anchor words Preliminaries X 1,, X n p (x; θ), θ = (θ 1,

More information

Least Squares Optimization

Least Squares Optimization Least Squares Optimization The following is a brief review of least squares optimization and constrained optimization techniques. I assume the reader is familiar with basic linear algebra, including the

More information

1. Structured representation of high-order tensors revisited. 2. Multi-linear algebra (MLA) with Kronecker-product data.

1. Structured representation of high-order tensors revisited. 2. Multi-linear algebra (MLA) with Kronecker-product data. Lect. 4. Toward MLA in tensor-product formats B. Khoromskij, Leipzig 2007(L4) 1 Contents of Lecture 4 1. Structured representation of high-order tensors revisited. - Tucker model. - Canonical (PARAFAC)

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 6: Some Other Stuff PD Dr.

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The

More information

A Simpler Approach to Low-Rank Tensor Canonical Polyadic Decomposition

A Simpler Approach to Low-Rank Tensor Canonical Polyadic Decomposition A Simpler Approach to Low-ank Tensor Canonical Polyadic Decomposition Daniel L. Pimentel-Alarcón University of Wisconsin-Madison Abstract In this paper we present a simple and efficient method to compute

More information

Foundations of Computer Vision

Foundations of Computer Vision Foundations of Computer Vision Wesley. E. Snyder North Carolina State University Hairong Qi University of Tennessee, Knoxville Last Edited February 8, 2017 1 3.2. A BRIEF REVIEW OF LINEAR ALGEBRA Apply

More information

Fourier PCA. Navin Goyal (MSR India), Santosh Vempala (Georgia Tech) and Ying Xiao (Georgia Tech)

Fourier PCA. Navin Goyal (MSR India), Santosh Vempala (Georgia Tech) and Ying Xiao (Georgia Tech) Fourier PCA Navin Goyal (MSR India), Santosh Vempala (Georgia Tech) and Ying Xiao (Georgia Tech) Introduction 1. Describe a learning problem. 2. Develop an efficient tensor decomposition. Independent component

More information

18.06SC Final Exam Solutions

18.06SC Final Exam Solutions 18.06SC Final Exam Solutions 1 (4+7=11 pts.) Suppose A is 3 by 4, and Ax = 0 has exactly 2 special solutions: 1 2 x 1 = 1 and x 2 = 1 1 0 0 1 (a) Remembering that A is 3 by 4, find its row reduced echelon

More information

arxiv: v1 [math.na] 5 May 2011

arxiv: v1 [math.na] 5 May 2011 ITERATIVE METHODS FOR COMPUTING EIGENVALUES AND EIGENVECTORS MAYSUM PANJU arxiv:1105.1185v1 [math.na] 5 May 2011 Abstract. We examine some numerical iterative methods for computing the eigenvalues and

More information

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters, transforms,

More information

Introduction to Tensors. 8 May 2014

Introduction to Tensors. 8 May 2014 Introduction to Tensors 8 May 2014 Introduction to Tensors What is a tensor? Basic Operations CP Decompositions and Tensor Rank Matricization and Computing the CP Dear Tullio,! I admire the elegance of

More information

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak, scribe: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters,

More information

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated. Math 504, Homework 5 Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated 1 Find the eigenvalues and the associated eigenspaces

More information

DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING. By T. Tony Cai and Linjun Zhang University of Pennsylvania

DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING. By T. Tony Cai and Linjun Zhang University of Pennsylvania Submitted to the Annals of Statistics DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING By T. Tony Cai and Linjun Zhang University of Pennsylvania We would like to congratulate the

More information

2. Review of Linear Algebra

2. Review of Linear Algebra 2. Review of Linear Algebra ECE 83, Spring 217 In this course we will represent signals as vectors and operators (e.g., filters, transforms, etc) as matrices. This lecture reviews basic concepts from linear

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: PCA and the Power Iteration Method

CS168: The Modern Algorithmic Toolbox Lecture #8: PCA and the Power Iteration Method CS168: The Modern Algorithmic Toolbox Lecture #8: PCA and the Power Iteration Method Tim Roughgarden & Gregory Valiant April 15, 015 This lecture began with an extended recap of Lecture 7. Recall that

More information

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,

More information

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search

More information

Provable Alternating Minimization Methods for Non-convex Optimization

Provable Alternating Minimization Methods for Non-convex Optimization Provable Alternating Minimization Methods for Non-convex Optimization Prateek Jain Microsoft Research, India Joint work with Praneeth Netrapalli, Sujay Sanghavi, Alekh Agarwal, Animashree Anandkumar, Rashish

More information

Linear Algebra- Final Exam Review

Linear Algebra- Final Exam Review Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Power iteration Notes for 2016-10-17 In most introductory linear algebra classes, one computes eigenvalues as roots of a characteristic polynomial. For most problems, this is a bad idea: the roots of

More information

There are two things that are particularly nice about the first basis

There are two things that are particularly nice about the first basis Orthogonality and the Gram-Schmidt Process In Chapter 4, we spent a great deal of time studying the problem of finding a basis for a vector space We know that a basis for a vector space can potentially

More information

Appendix A. Proof to Theorem 1

Appendix A. Proof to Theorem 1 Appendix A Proof to Theorem In this section, we prove the sample complexity bound given in Theorem The proof consists of three main parts In Appendix A, we prove perturbation lemmas that bound the estimation

More information

Tensor Factorization via Matrix Factorization

Tensor Factorization via Matrix Factorization Volodymyr Kuleshov Arun Tejasvi Chaganty Percy Liang Department of Computer Science Stanford University Stanford, CA 94305 Abstract Tensor factorization arises in many machine learning applications, such

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Numerical Methods I Eigenvalue Problems

Numerical Methods I Eigenvalue Problems Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture

More information

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition AM 205: lecture 8 Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition QR Factorization A matrix A R m n, m n, can be factorized

More information

Lecture 8: Linear Algebra Background

Lecture 8: Linear Algebra Background CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected

More information

Applied Machine Learning for Biomedical Engineering. Enrico Grisan

Applied Machine Learning for Biomedical Engineering. Enrico Grisan Applied Machine Learning for Biomedical Engineering Enrico Grisan enrico.grisan@dei.unipd.it Data representation To find a representation that approximates elements of a signal class with a linear combination

More information

MAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction

MAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction MAT4 : Introduction to Applied Linear Algebra Mike Newman fall 7 9. Projections introduction One reason to consider projections is to understand approximate solutions to linear systems. A common example

More information

Lecture 4. CP and KSVD Representations. Charles F. Van Loan

Lecture 4. CP and KSVD Representations. Charles F. Van Loan Structured Matrix Computations from Structured Tensors Lecture 4. CP and KSVD Representations Charles F. Van Loan Cornell University CIME-EMS Summer School June 22-26, 2015 Cetraro, Italy Structured Matrix

More information

Computational Methods. Eigenvalues and Singular Values

Computational Methods. Eigenvalues and Singular Values Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations

More information

Conjugate Gradient (CG) Method

Conjugate Gradient (CG) Method Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous

More information

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

MATH 304 Linear Algebra Lecture 34: Review for Test 2. MATH 304 Linear Algebra Lecture 34: Review for Test 2. Topics for Test 2 Linear transformations (Leon 4.1 4.3) Matrix transformations Matrix of a linear mapping Similar matrices Orthogonality (Leon 5.1

More information

Introduction to gradient descent

Introduction to gradient descent 6-1: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction to gradient descent Derivation and intuitions Hessian 6-2: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction Our

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM 33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM (UPDATED MARCH 17, 2018) The final exam will be cumulative, with a bit more weight on more recent material. This outline covers the what we ve done since the

More information

CS 143 Linear Algebra Review

CS 143 Linear Algebra Review CS 143 Linear Algebra Review Stefan Roth September 29, 2003 Introductory Remarks This review does not aim at mathematical rigor very much, but instead at ease of understanding and conciseness. Please see

More information

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix

More information

ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition

ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition Wing-Kin (Ken) Ma 2017 2018 Term 2 Department of Electronic Engineering The Chinese University

More information

Principal Component Analysis

Principal Component Analysis Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used

More information

Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning

Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning David Steurer Cornell Approximation Algorithms and Hardness, Banff, August 2014 for many problems (e.g., all UG-hard ones): better guarantees

More information

Numerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization

Numerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725 Consider Last time: proximal Newton method min x g(x) + h(x) where g, h convex, g twice differentiable, and h simple. Proximal

More information

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers. Linear Algebra - Test File - Spring Test # For problems - consider the following system of equations. x + y - z = x + y + 4z = x + y + 6z =.) Solve the system without using your calculator..) Find the

More information

Identifiability and Learning of Topic Models: Tensor Decompositions under Structural Constraints

Identifiability and Learning of Topic Models: Tensor Decompositions under Structural Constraints Identifiability and Learning of Topic Models: Tensor Decompositions under Structural Constraints Anima Anandkumar U.C. Irvine Joint work with Daniel Hsu, Majid Janzamin Adel Javanmard and Sham Kakade.

More information

Lecture 2: Numerical linear algebra

Lecture 2: Numerical linear algebra Lecture 2: Numerical linear algebra QR factorization Eigenvalue decomposition Singular value decomposition Conditioning of a problem Floating point arithmetic and stability of an algorithm Linear algebra

More information

TENSOR DECOMPOSITIONS AND THEIR APPLICATIONS

TENSOR DECOMPOSITIONS AND THEIR APPLICATIONS TENSOR DECOMPOSITIONS AND THEIR APPLICATIONS ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY SPEARMAN S HYPOTHESIS Charles Spearman (1904): There are two types of intelligence, educ%ve and reproduc%ve

More information

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013. The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment Two Caramanis/Sanghavi Due: Tuesday, Feb. 19, 2013. Computational

More information

Principal components analysis COMS 4771

Principal components analysis COMS 4771 Principal components analysis COMS 4771 1. Representation learning Useful representations of data Representation learning: Given: raw feature vectors x 1, x 2,..., x n R d. Goal: learn a useful feature

More information

Tensors. Notes by Mateusz Michalek and Bernd Sturmfels for the lecture on June 5, 2018, in the IMPRS Ringvorlesung Introduction to Nonlinear Algebra

Tensors. Notes by Mateusz Michalek and Bernd Sturmfels for the lecture on June 5, 2018, in the IMPRS Ringvorlesung Introduction to Nonlinear Algebra Tensors Notes by Mateusz Michalek and Bernd Sturmfels for the lecture on June 5, 2018, in the IMPRS Ringvorlesung Introduction to Nonlinear Algebra This lecture is divided into two parts. The first part,

More information

Eigenvalues and eigenvectors

Eigenvalues and eigenvectors Chapter 6 Eigenvalues and eigenvectors An eigenvalue of a square matrix represents the linear operator as a scaling of the associated eigenvector, and the action of certain matrices on general vectors

More information

Matrices, Moments and Quadrature, cont d

Matrices, Moments and Quadrature, cont d Jim Lambers CME 335 Spring Quarter 2010-11 Lecture 4 Notes Matrices, Moments and Quadrature, cont d Estimation of the Regularization Parameter Consider the least squares problem of finding x such that

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.] Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Computational Methods CMSC/AMSC/MAPL 460. EigenValue decomposition Singular Value Decomposition. Ramani Duraiswami, Dept. of Computer Science

Computational Methods CMSC/AMSC/MAPL 460. EigenValue decomposition Singular Value Decomposition. Ramani Duraiswami, Dept. of Computer Science Computational Methods CMSC/AMSC/MAPL 460 EigenValue decomposition Singular Value Decomposition Ramani Duraiswami, Dept. of Computer Science Hermitian Matrices A square matrix for which A = A H is said

More information

Lecture 9: Low Rank Approximation

Lecture 9: Low Rank Approximation CSE 521: Design and Analysis of Algorithms I Fall 2018 Lecture 9: Low Rank Approximation Lecturer: Shayan Oveis Gharan February 8th Scribe: Jun Qi Disclaimer: These notes have not been subjected to the

More information

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2 Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch

More information

Statistical Geometry Processing Winter Semester 2011/2012

Statistical Geometry Processing Winter Semester 2011/2012 Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian

More information

Lecture Note 12: The Eigenvalue Problem

Lecture Note 12: The Eigenvalue Problem MATH 5330: Computational Methods of Linear Algebra Lecture Note 12: The Eigenvalue Problem 1 Theoretical Background Xianyi Zeng Department of Mathematical Sciences, UTEP The eigenvalue problem is a classical

More information

Quantum Computing Lecture 2. Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces

More information

Math 312 Final Exam Jerry L. Kazdan May 5, :00 2:00

Math 312 Final Exam Jerry L. Kazdan May 5, :00 2:00 Math 32 Final Exam Jerry L. Kazdan May, 204 2:00 2:00 Directions This exam has three parts. Part A has shorter questions, (6 points each), Part B has 6 True/False questions ( points each), and Part C has

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Chapter Two Elements of Linear Algebra

Chapter Two Elements of Linear Algebra Chapter Two Elements of Linear Algebra Previously, in chapter one, we have considered single first order differential equations involving a single unknown function. In the next chapter we will begin to

More information

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Chapter 14 SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Today we continue the topic of low-dimensional approximation to datasets and matrices. Last time we saw the singular

More information

arxiv: v5 [math.na] 16 Nov 2017

arxiv: v5 [math.na] 16 Nov 2017 RANDOM PERTURBATION OF LOW RANK MATRICES: IMPROVING CLASSICAL BOUNDS arxiv:3.657v5 [math.na] 6 Nov 07 SEAN O ROURKE, VAN VU, AND KE WANG Abstract. Matrix perturbation inequalities, such as Weyl s theorem

More information

Must-read Material : Multimedia Databases and Data Mining. Indexing - Detailed outline. Outline. Faloutsos

Must-read Material : Multimedia Databases and Data Mining. Indexing - Detailed outline. Outline. Faloutsos Must-read Material 15-826: Multimedia Databases and Data Mining Tamara G. Kolda and Brett W. Bader. Tensor decompositions and applications. Technical Report SAND2007-6702, Sandia National Laboratories,

More information

Low Rank Approximation Lecture 7. Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL

Low Rank Approximation Lecture 7. Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL Low Rank Approximation Lecture 7 Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL daniel.kressner@epfl.ch 1 Alternating least-squares / linear scheme General setting:

More information

MATH 581D FINAL EXAM Autumn December 12, 2016

MATH 581D FINAL EXAM Autumn December 12, 2016 MATH 58D FINAL EXAM Autumn 206 December 2, 206 NAME: SIGNATURE: Instructions: there are 6 problems on the final. Aim for solving 4 problems, but do as much as you can. Partial credit will be given on all

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Math 205, Summer I, Week 4b:

Math 205, Summer I, Week 4b: Math 205, Summer I, 2016 Week 4b: Chapter 5, Sections 6, 7 and 8 (5.5 is NOT on the syllabus) 5.6 Eigenvalues and Eigenvectors 5.7 Eigenspaces, nondefective matrices 5.8 Diagonalization [*** See next slide

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

Fundamentals of Multilinear Subspace Learning

Fundamentals of Multilinear Subspace Learning Chapter 3 Fundamentals of Multilinear Subspace Learning The previous chapter covered background materials on linear subspace learning. From this chapter on, we shall proceed to multiple dimensions with

More information

MAT 1302B Mathematical Methods II

MAT 1302B Mathematical Methods II MAT 1302B Mathematical Methods II Alistair Savage Mathematics and Statistics University of Ottawa Winter 2015 Lecture 19 Alistair Savage (uottawa) MAT 1302B Mathematical Methods II Winter 2015 Lecture

More information