Latent Semantic Indexing

Size: px
Start display at page:

Download "Latent Semantic Indexing"

Transcription

1 Latent Semantic Indexing 1

2 Latent Semantic Indexing Term-document matrices are very large, though most cells are zeros But the number of topics that people talk about is small (in some sense) Clothes, movies, politics, Each topic can be represented as a cluster of (semantically) related terms, e.g.: clothes=golf, jacket, shoe.. Can we represent the term-document space by a lower dimensional latent space (latent space=set of topics)? 2

3 Searching with latent topics Given a collection of documents, LSI learns clusters of frequently co-occurring terms (ex: information retrieval, ranking and web)! If you query with ranking, information retrieval LSI automatically extends the search to documents including also (and even ONLY) web 3

4 4 Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee Selection based on Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee Car GTI Polo Document base (20) Motor Bike Oil Tourer Bed lace legal button soft cat line yellow wind full sail harbour beach report June Speed Fish Pond gold Koi PC Dell RAM Floppy Core Apple Pip Tree Pea Pod Fresh Green French Lupin Seed May April Office Pen Desk VDU Friend Pal Help Can Paper Paste Pencil Roof Card Stamp Glue Happy Send Toil Work Time Cost With standard VSM 4 documents are selected

5 20 documents Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee Car GTI Polo Fish Pond gold Koi PC Dell RAM Floppy Core Apple Pip Tree Pea Pod Fresh Green French Lupin Seed May April Motor Bike Oil Tourer Bed lace legal button soft cat line yellow wind full sail harbour beach report June Speed Office Pen Desk VDU Selection based on Friend Pal Help Can Paper Paste Pencil Roof Card Stamp Glue Happy Send Toil Work Time Cost Rank of selected documents Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee The most relevant words associated with golf in these docs are: Car, and tf.idf Car 2 *(20/3) = 13 2 *(20/3) = 13 3 *(20/16) = 4

6 20 docs Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee Car GTI Polo Fish Pond gold Koi PC Dell RAM Floppy Core Apple Pip Tree Pea Pod Fresh Green French Lupin Seed May April Motor Bike Oil Tourer Selezione basata su Car GTI Bed lace legal button soft cat line yellow Car Clarkson Badge wind full sail harbour beach report June Speed Polo Red Office Pen Desk VDU Friend Pal Help Can Tiger Woods Belfry Tee Paper Paste Pencil Roof Card Stamp Glue Happy Send If we consider the co-occurring terms with higher tf*idf, car e topgear turn out to be related to more than petrol Toil Work Time Cost Rank of selected docs Car 2 *(20/3) = 13 2 *(20/3) = 13 3 *(20/16) = 4

7 20 docs Car GTI Motor Bike Oil Tourer Car GTI Car Clarkson Badge Bed lace legal button Polo Red soft cat line yellow Car Clarkson Badge Tiger Woods Belfry Tee wind full sail harbour beach Car GTI Polo report June Speed Selection based on Polo Red Fish Pond gold Koi Office Pen Desk VDU PC Dell RAM Floppy Core Apple Pip Tree Tiger Woods Belfry Tee Pea Pod Fresh Green French Lupin Seed May April Selection based on the semantic domain of Car GTI Car Clarkson Badge Polo Red We now search with all the Friend words Paper in the Card semantic Toil domain of. Pal Stamp Help Paste Glue Work The Pencil list Happy of retrieved Time docs now Can Roof Send Cost is based on and the other related words Tiger Woods Belfry Tee Car Wheel GTI Polo rank df selected docs Car 2 *(20/3) = 13 2 *(20/3) = 13 3 *(20/16) = 4

8 20 docs Car GTI Motor Bike Oil Tourer Selezione basata su Car GTI Car Clarkson Badge Polo Red Car Clarkson Badge Tiger Woods Belfry Tee Car GTI Polo Polo Red Tiger Woods Belfry Tee Selection based on semantic domain Car GTI Car Clarkson Badge Polo Red Rank Fish Pond gold Koi PC Dell RAM Floppy Core Apple Pip Tree Tiger Woods Belfry Tee Pea Pod Fresh Green French The co-occurrence based Bed soft wind report Office Friend Paper Card lace full Pen Pal Stamp legal ranking cat sail improves the Desk performance. Help Paste Glue line harbour June Pencil Happy Note button that yellow one beach of the Speed most VDU relevant Can Roof doc Send does NOT include the word, and a doc with a spurious sense disappears Lupin Seed May April Toil Work Time Cost Car Wheel GTI Polo rank of selected docs Car 2 *(20/3) = 13 2 *(20/3) = 13 3 *(20/16) = 4

9 Ranking with latent Semantic Indexing Previous example just gives the intuition Latent Semantic Indexing is an algebraic method to identify clusters of co-occurring terms, called latent topics, and to compute query-doc similarity in a latent space, in which every coordinate is a latent topic. A latent quantity is one which cannot directly observed, what is observed is a measurement which may include some amount of random errors (topics are latent in this sense: we observe them, but they are an approximation of true semantic topics) Since it is an algebraic method, needs some linear algebra background 9

10 The LSI method: how to detect topics Linear Algebra Background 10

11 Eigenvalues & Eigenvectors Eigenvectors (for a square m m matrix S) Av = λv Example (right) eigenvector eigenvalue Av = λv Def: A vector v R n, v 0, is an eigenvector of a matrix mxm A with corresponding eigenvalue λ, if: Av = λv 11

12 Algebraic method How many eigenvalues are there at most? Av = λv equation has a non-zero solution if A A A Where I is the identity matrix this is a m-th order equation in λ which can have at most m distinct solutions (roots of the characteristic polynomial) can be complex even though A is real. 12

13 Example of eigenvector/eigenvalues " A = $ 1 1 # 3 5 " $ # % ', v = " 1 $ & # 3 Av = λv 1 1 %" 1 % ' $ ' 3 5 &# 3 = 4 " 1 $ & # 3 " 1+ 3( 1) % $ # 3+ 5( 3) ' = " 4 % $ ' & # 12 & " 4 % $ ' # 12 = " 4 % $ ' & # 12 & % ', λ = 4 & % ' &

14 Example of computation remember " A = $ 1 1 # 3 5 % ' & det M = M = det(a λi) = 0 " 1 1 % $ ' # 3 5 λ " % and 4 are the $ eigenvalues ' & # 0 1 = 0 of S & (A λi)v = 0 " 1 1 % $ ' # 3 5 " $ λ 0 % " ' & # 0 λ = %" α % λ $ ' $ & # & β ' = " 0 % 2 = 2 $ ' " # & # 0 & β % " v = " 1 λ 1 % $ ' # 3 5 λ = % " α % ' $ $ ' # 3 1 $ &# β ' = " 0 % # β & $ ' & # 0 & & ( * 3α β = 0 ) β = 3α +* 3α + β = 0 a c b d = ad bc (λ 1)(λ 5) + 3 = 0 λ 2 6λ = 0 λ 2 6λ + 8 = 0 (λ 4)(λ 2) = 0 λ 1 = 4,λ 2 = 2 λ Characteristic polynomial = (A 1 λi)v = 0 " &% " α %# $ ' # $ &# β ' = " 0 % v = $ $ ' $ & # 0 & " 1 1 %"% 3α α % $ ' # 3 3 $ &# β ' = " 0 % $ ' & # 0 & ( * α β = 0 ) α = β +* 3α + 3β = 0 '!! " Note that we compute only the DIRECTION of eigenvectors 4

15 Geometric interpretation A matrix-vector multiplication Ax is a linear transformation over the vector 15

16 Matrix vector multiplication Matrix multiplication by a vector = a linear transformation of the initial vector, that implies rotation and translation 16

17 Geometric interpretation 17

18 Geometric interpretation Ax=1.34x=λx An eigenvector is a special vector that is transformed into its scalar multiple under a given matrix (no rotation!) 18

19 Geometric interpretation Here we found another eigenvector for the matrix A Note that for any single eigenvalue you have infinite eigenvectors, but they have the same direction 19

20 Matrix-vector multiplication Eigenvectors of different eigenvalues are linearly independent (i.e. α 1.. α n è α 1 v α n v n 0) For square normal matrixes eigenvectors of different eigenvalues define an orthonormal space and they are othogonal. A square matrix is NORMAL iff it commutes with its transpose, i.e. AA T =A T A Example: è AA T= = A T A 20

21 Difference between orthonormal and orthogonal? Orthogonal mean that the dot product is null (the cosin of the angle is zero). Orthonormal mean that the dot product is null and the norm of the vectos is equal to 1. What we are actually saying is that eigenvectors define a set of DIRECTIONS wich are orthogonal. If two or more vectors are orthonormal they are also orthogonal but the inverse is not true. 21

22 Example: projecting a vector on 2 orthonormal spaces (or bases ) e 1,e 1, e 2,e 2 are unary vectors and v 1,v 1, v 2,v 2 are the coordinates of v along the directions of e 1,e 1, e 2,e 2 22

23 The effect of a matrix-vector multiplication is governed by eigenvectors and eigenvalues A matrix-vector multiplication such as Ax (A normal matrix, x a generic vector as in the previous slide) can be rewritten in terms of the eigenvalues/vectors of A. Example: 2 X= 4 Ax = A(2v 1 + 4v 2 + 6v 3 ) 6 Ax = 2Av 1 + 4Av 2 + 6Av 3 = 2λ 1 v 1 + 4λ 2 v 2 + 6λ 3 v 3 Where v 1,v 2 v 3 are the (orthogonal) eigenvectors of A Even though x is an arbitrary vector, the action of A on x (transformation) is determined by the eigenvalues/vectors. Why? 23

24 Geometric explanation x is a generic vector with coordinates x i ; λ i,v i are the eigenvalues and eigenvectors of A Ax = x 1 λ 1 v 1 + x 2 λ 2 v 2 + x 3 λ 3 v 3 Multiplying a matrix and a vector has two effects over the vector: rotation (the coordinates of the vector change) and scaling (the length changes). The max compression and rotation depends on the matrix s eigenvalues λi (s1,s2 and s3 in the picture)

25 Geometric explanation In the distorsion, the max effect is played by the biggest eigenvalues (s1 and s2 in the picture ) The eigenvalues describe the distorsion operated by the matrix on the original vector

26 Summary so far A matrix A has eigenvectors v and eigenvalues λ, defined by Av=λv Eigenvalues can be computed as: We can compute only the the direction of eigenvectors, since for any eigenvalue there are infinite eigenvectors lying on the same direction If A is normal (i.e. if AA T =A T A) then the eigenvector form an othonormal basis The product of A by ANY vector x is a linear transformation of x where the rotation is determined by eigenvectors and the translation is determined by the eigenvalues. The biggest role in this transformation is played by the biggest eigenvalues. 26

27 Eigen/diagonal Decomposition Let A be a square matrix with m orthogonal eigenvectors (hence, A is normal) Theorem: Exists an eigen decomposition A=UΛU -1 Λ is a diagonal matrix (all zero except the diagonal cells) Columns of U are eigenvectors of A Diagonal elements of Λ are eigenvalues of A 27

28 Diagonal decomposition: why/how = n v v U... 1 Let U have the eigenvectors as columns: AU = A v 1... v n! " # # # $ % & & & = λ 1 v 1... λ n v n! " # # # $ % & & & = v 1... v n! " # # # $ % & & & λ 1... λ n! " # # # # $ % & & & & Then, AU can be written And A=UΛU 1. Thus AU=UΛ, or U 1 AU=Λ 28

29 Example AU=UΛ From this we compute λ 1 =1, λ 2 =3 from which we get v 11 = 2v 12 From which we get v 21 =0 and v 22 any real 29

30 Diagonal decomposition example 2! Recall A = # 2 1 " $ &; λ 1 =1,λ 2 = 3. % 1 1 The eigenvectors and form Inverting, we have U 1 Then, A=UΛU 1 = = 1/ 2 1/ 2 1/ 2 1/ 2 U = 1 1 Recall UU 1 = / 2 1/ / 2 1/

31 So what? What do these matrices have to do with Information Retrieval and document ranking? Recall M N term-document matrices But everything so far needs square normal matrices so you need to be patient and learn one last notion 31

32 Singular Value Decomposition for non-square matrixes For a non-square M N matrix A of rank r there exists a factorization (Singular Value Decomposition = SVD) as follows: A = UΣV T M M M N V is N N The columns of U are orthogonal eigenvectors of AA T (left singular vectors). The columns of V are orthogonal eigenvectors of A T A (right singular eigenvector). NOTE THAT AA T and A T A are square symmetric (and hence NORMAL) Eigenvalues λ 1 λ r of AA T = eigenvalues of A T A and: Σ = ( σ... ) diag 1 σ r Singular values. σ i = λ i 32

33 An example 33

34 Singular Value Decomposition Illustration of SVD dimensions and sparseness M All zeros! So these become zeros, too N MxM MxN NxN So these become zeros, too

35 Back to matrix-vector multiplication Remember what we said? In a matrix vector multiplication the biggest role is played by the biggest eigenvalues The diagonal matrix Σ has the eigenvalues of A T A (called the singular values of A) in decreasing order along the diagonal We can therefore apply an approximation by setting σ i =0 if σ i θ and only consider the first k singular values 35

36 Reduced SVD If we retain only k singular values, and set the rest to 0, then we don t need the matrix parts in red Then Σ is k k, U is M k, V T is k N, and A k is M N This is referred to as the reduced SVD, or rank k approximation Now all the red and yellow parts are zeros!! k 36

37 Let s recap M<N M>N Since the yellow part is zero, an exact representation of A is: A = σ 1 u 1 v 1 T +σ 2 u 2 v 2 T +...+σ r u r v r T r = min(m,n) But for some k<r, a good approximation is: A k = σ 1 u 1 v 1 T +σ 2 u 2 v 2 T σ k u k v k T 37

38 Example of rank approximation 1 = x x 1 A A*= A 38

39 Approximation error How good (bad) is this approximation? It s the best possible, measured by the Frobenius norm of the error: A X A = σ :min F k F k + 1 X rank ( X ) = k = A σ i = λ i where the σ i are ordered such that σ i σ i+1. Suggests why Frobenius error drops as k increases. 39

40 Images gives a better intuition (image = matrix of pixels) 40

41 K=10 41

42 K=20 42

43 K=30 43

44 K=100 44

45 K=322 (the original image) 45

46 We save space!! But this is only one issue 46

47 So, finally, back to IR Our initial problem was: the term-document (MxN) matrix A is highly sparse (has many zeros) However, since groups of terms tend to co-occur together, can we identify the semantic space of these clusters of terms, and apply the vector space model in the semantic space defined by such clusters? What we learned so far: Matrix A can be decomposed and rank k approximated using SVD Does this help solving our initial problems? 47

48 A is our term document matrix Latent Semantic Indexing via the SVD If A is a term/document matrix, then AA T and A T A are the matrixes of term and document co-occurrences, repectively 48

49 Meaning of A T A and AA T Word i in doc k Word j in doc k L = A A T = L ij = A ik A T kj N L ij = A ik A T kj = A ik A jk k=1 N k=1 L ij is the number of documents in which wi and wj A T co-occurr Similarly, L T ij is the number of co-occurring words in docs di and dj (or viceversa if A is a document-term matrix rather than term-document) A

50 Example

51 Term-document matrix A

52 Term co-occurrences example L trees,graph = ( ) ( ) T =2

53 So the matrix L=AA T is the matrix of term co-occurrences Remember: eigenvectors of a matrix define an orthonormal space Remember: bigger eigenvalues define the main directions of this space But: Matrix L and L T are SIMILARITY matrixes (respectively, of terms and of documents). They define a SIMILARITY SPACE (the orthonormal space of their eigenvectors) If the matrix elements are word co-occurrences, bigger eigenvalues are associated to bigger groups of similar words Similarly, bigger eigenvalues of L T =A T A are associated with bigger groups of similar documents (those in which co-occur the same terms)

54 LSI: the intuition The blue segments give the intuition of eigenvalues of L T =A T A Bigger eigenvalues are those for which the projection of all vectors on the direction of correspondent eigenvectors is higher t1 Projecting A in the term space: green, yellow and red vectors are documents. If they form small angles, they have common words (remember cosin-sim) The black vector are the unary eigenvector of A: they represent the main directions of the document vectors t2 t3 54

55 LSI intuition!" d j = (d 1 j,d 2 j,d 3 j )!" L T d j = d 1 j λ 1 v 1 + d 2 j λ 2 v 2 + d 3 j λ 3 v 3 t1 If we multiply all document vectors by L T = A T A, their distorsion is mostly determined by the highest eigenvalues t2 We now project our document vectors on the reference orthonormal system represented by the 3 black vectors t3 55

56 LSI intuition..if we remove the dimension(s) with lower eigenvalues (i.e. if we rank-reduce Σ) we don t loose much information!" d j = (d 1 j,d 2 j,d 3 j )!" L T d j d 1 j λ 1 v 1 + d 2 j λ 2 v 2 Remember that the two new axis represent a combination of co-occurring words e.g. a latent semantic space 56

57 Example t1 t2 t3 t4 d1 d2 d3 d We project terms and docs on two dimensions, v1 and v2 (the principal eigenvectors) x x Note that the direction of each eigenvector is determined by Approximated We have two latent semantic new the term-doc direction of just two terms: (t1,t2) or (t3,t4) matrix coordinates: s1:(t1,t2) and s2:(t3,t4) Even if t2 does not occur in d2, now if we query with t2 the system will return also d2!! 57

58 Co-occurrence space In principle, the space of document or term co-occurrences is much (much!) higher than the original space of terms!! But with SVD we consider only the most relevant ones, trough rank reduction A =U V T U k Σ k V k T = A k 58

59 Summary so-far We compute the SVD rank-k approximation for the term-document matrix A This approximation is based on considering only the principal eigenvalues of the term cooccurrence and document similarity matrixes (L=AA T and LT=A T A) The eigenvectors of the eigenvalues of L=AA T and LT=A T A represent the main directions identified by term vectors and document vectors, respectively. 59

60 LSI: what are the steps From term-doc matrix A, compute the approximation A k. with SVD Project docs, terms and queries in a space of k<<r dimensions (the k survived eigenvectors) and compute cos-similarity as usual These dimensions are not the original axes, but those defined by the orthonormal space of the reduced matrix Ak Aq! A k q! = σ 1 q 1 v! 1 +σ 2 q 2 v! σ k q k v! k Where σ i q i (i=1,2..k<<r) are the new coordinates of q in the orthonormal space of Ak 60

61 Projecting terms documents and queries in the LS space A If A=UΣV T we also have that: V = A T UΣ -1 t = t T ΣV T d = d T UΣ -1 q = q T UΣ -1 After rank k- approximation : A Ak=U k Σ k V k T d k d T U k Σ k -1 q k q T U k Σ k -1 sim(q, d) = sim(q T U k Σ k -1, d T U k Σ k -1 ) 61

62 Consider a term-doc matrix MxN (M=11, N=3) and a query q query A

63 1. Compute SVD: A= UΣVT

64 2. Obtain a low rank approximation (k=2) A k = U k Σ k V T k

65 3a. Compute doc/query similarity For N documents, A k has N columns, each representing the coordinates of a document d i projected in the k LSI dimensions A query is considered like a document, and is projected in the LSI space

66 3c. Compute the query vector q k = q T U k Σ k -1 q is projected in the 2-dimension LSI space!

67 Documents and queries projected in the LSI space

68 q/d similarity

69 An overview of a semantic network of terms based on the top 100 most significant latent semantic dimensions (Zhu&Chen) 69

70 Conclusion LSI performs a low-rank approximation of document-term matrix (typical rank ) General idea Map documents (and terms) to a lowdimensional representation. Design a mapping such that the low-dimensional space reflects semantic associations between words (latent semantic space). Compute document similarity based on the cossim in this latent semantic space

71 Another LSI Example

72

73 t1 t2 t3 t4 d1 d2 d3 A T AA T t1 t2 t3 Term co-occurrences

74 A k =U k Σ k V kt A Now it is like if t3 belongs to d1!

Information Retrieval

Information Retrieval Introduction to Information CS276: Information and Web Search Christopher Manning and Pandu Nayak Lecture 13: Latent Semantic Indexing Ch. 18 Today s topic Latent Semantic Indexing Term-document matrices

More information

Latent Semantic Models. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze

Latent Semantic Models. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze Latent Semantic Models Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze 1 Vector Space Model: Pros Automatic selection of index terms Partial matching of queries

More information

Linear Algebra Background

Linear Algebra Background CS76A Text Retrieval and Mining Lecture 5 Recap: Clustering Hierarchical clustering Agglomerative clustering techniques Evaluation Term vs. document space clustering Multi-lingual docs Feature selection

More information

Linear Algebra - Part II

Linear Algebra - Part II Linear Algebra - Part II Projection, Eigendecomposition, SVD (Adapted from Sargur Srihari s slides) Brief Review from Part 1 Symmetric Matrix: A = A T Orthogonal Matrix: A T A = AA T = I and A 1 = A T

More information

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2016 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,

More information

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2014 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,

More information

CS 143 Linear Algebra Review

CS 143 Linear Algebra Review CS 143 Linear Algebra Review Stefan Roth September 29, 2003 Introductory Remarks This review does not aim at mathematical rigor very much, but instead at ease of understanding and conciseness. Please see

More information

Linear Algebra for Machine Learning. Sargur N. Srihari

Linear Algebra for Machine Learning. Sargur N. Srihari Linear Algebra for Machine Learning Sargur N. srihari@cedar.buffalo.edu 1 Overview Linear Algebra is based on continuous math rather than discrete math Computer scientists have little experience with it

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

CSE 494/598 Lecture-4: Correlation Analysis. **Content adapted from last year s slides

CSE 494/598 Lecture-4: Correlation Analysis. **Content adapted from last year s slides CSE 494/598 Lecture-4: Correlation Analysis LYDIA MANIKONDA HT TP://WWW.PUBLIC.ASU.EDU/~LMANIKON / **Content adapted from last year s slides Announcements Project-1 Due: February 12 th 2016 Analysis report:

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

Problems. Looks for literal term matches. Problems:

Problems. Looks for literal term matches. Problems: Problems Looks for literal term matches erms in queries (esp short ones) don t always capture user s information need well Problems: Synonymy: other words with the same meaning Car and automobile 电脑 vs.

More information

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD DATA MINING LECTURE 8 Dimensionality Reduction PCA -- SVD The curse of dimensionality Real data usually have thousands, or millions of dimensions E.g., web documents, where the dimensionality is the vocabulary

More information

UNIT 6: The singular value decomposition.

UNIT 6: The singular value decomposition. UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T

More information

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012 Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component

More information

Latent Semantic Analysis. Hongning Wang

Latent Semantic Analysis. Hongning Wang Latent Semantic Analysis Hongning Wang CS@UVa VS model in practice Document and query are represented by term vectors Terms are not necessarily orthogonal to each other Synonymy: car v.s. automobile Polysemy:

More information

Singular Value Decomposition

Singular Value Decomposition Singular Value Decomposition Motivatation The diagonalization theorem play a part in many interesting applications. Unfortunately not all matrices can be factored as A = PDP However a factorization A =

More information

Let A an n n real nonsymmetric matrix. The eigenvalue problem: λ 1 = 1 with eigenvector u 1 = ( ) λ 2 = 2 with eigenvector u 2 = ( 1

Let A an n n real nonsymmetric matrix. The eigenvalue problem: λ 1 = 1 with eigenvector u 1 = ( ) λ 2 = 2 with eigenvector u 2 = ( 1 Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Au = λu Example: ( ) 2 0 A = 2 1 λ 1 = 1 with eigenvector

More information

Latent Semantic Analysis. Hongning Wang

Latent Semantic Analysis. Hongning Wang Latent Semantic Analysis Hongning Wang CS@UVa Recap: vector space model Represent both doc and query by concept vectors Each concept defines one dimension K concepts define a high-dimensional space Element

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Dot Products K. Behrend April 3, 008 Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Contents The dot product 3. Length of a vector........................

More information

Linear Algebra Methods for Data Mining

Linear Algebra Methods for Data Mining Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 1. Basic Linear Algebra Linear Algebra Methods for Data Mining, Spring 2007, University of Helsinki Example

More information

Maths for Signals and Systems Linear Algebra in Engineering

Maths for Signals and Systems Linear Algebra in Engineering Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE

More information

Review of Linear Algebra

Review of Linear Algebra Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=

More information

MATH 583A REVIEW SESSION #1

MATH 583A REVIEW SESSION #1 MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),

More information

Linear Algebra in Actuarial Science: Slides to the lecture

Linear Algebra in Actuarial Science: Slides to the lecture Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations

More information

EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4]

EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: Au = λu λ C : eigenvalue u C n : eigenvector Example:

More information

Singular Value Decompsition

Singular Value Decompsition Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost

More information

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP) MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m

More information

Matrices A brief introduction

Matrices A brief introduction Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino Semester 1, 2014-15 B. Bona (DAUIN) Matrices Semester 1, 2014-15 1 / 41 Definitions Definition A matrix is a set of N real or complex

More information

Study Notes on Matrices & Determinants for GATE 2017

Study Notes on Matrices & Determinants for GATE 2017 Study Notes on Matrices & Determinants for GATE 2017 Matrices and Determinates are undoubtedly one of the most scoring and high yielding topics in GATE. At least 3-4 questions are always anticipated from

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Functional Analysis Review

Functional Analysis Review Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Introduction to Matrix Algebra August 18, 2010 1 Vectors 1.1 Notations A p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the line. When p

More information

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science Computational Methods CMSC/AMSC/MAPL 460 Eigenvalues and Eigenvectors Ramani Duraiswami, Dept. of Computer Science Eigen Values of a Matrix Recap: A N N matrix A has an eigenvector x (non-zero) with corresponding

More information

Image Registration Lecture 2: Vectors and Matrices

Image Registration Lecture 2: Vectors and Matrices Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx

More information

Manning & Schuetze, FSNLP (c) 1999,2000

Manning & Schuetze, FSNLP (c) 1999,2000 558 15 Topics in Information Retrieval (15.10) y 4 3 2 1 0 0 1 2 3 4 5 6 7 8 Figure 15.7 An example of linear regression. The line y = 0.25x + 1 is the best least-squares fit for the four points (1,1),

More information

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice 3. Eigenvalues and Eigenvectors, Spectral Representation 3.. Eigenvalues and Eigenvectors A vector ' is eigenvector of a matrix K, if K' is parallel to ' and ' 6, i.e., K' k' k is the eigenvalue. If is

More information

Basic Calculus Review

Basic Calculus Review Basic Calculus Review Lorenzo Rosasco ISML Mod. 2 - Machine Learning Vector Spaces Functionals and Operators (Matrices) Vector Space A vector space is a set V with binary operations +: V V V and : R V

More information

. = V c = V [x]v (5.1) c 1. c k

. = V c = V [x]v (5.1) c 1. c k Chapter 5 Linear Algebra It can be argued that all of linear algebra can be understood using the four fundamental subspaces associated with a matrix Because they form the foundation on which we later work,

More information

MATH 221, Spring Homework 10 Solutions

MATH 221, Spring Homework 10 Solutions MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the

More information

Matrices A brief introduction

Matrices A brief introduction Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino Semester 1, 2014-15 B. Bona (DAUIN) Matrices Semester 1, 2014-15 1 / 44 Definitions Definition A matrix is a set of N real or complex

More information

Introduction to Information Retrieval

Introduction to Information Retrieval Introduction to Information Retrieval http://informationretrieval.org IIR 18: Latent Semantic Indexing Hinrich Schütze Center for Information and Language Processing, University of Munich 2013-07-10 1/43

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

Matrices and Linear Algebra

Matrices and Linear Algebra Contents Quantitative methods for Economics and Business University of Ferrara Academic year 2017-2018 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Matrix Vector Products

Matrix Vector Products We covered these notes in the tutorial sessions I strongly recommend that you further read the presented materials in classical books on linear algebra Please make sure that you understand the proofs and

More information

Quantum Computing Lecture 2. Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

CS 340 Lec. 6: Linear Dimensionality Reduction

CS 340 Lec. 6: Linear Dimensionality Reduction CS 340 Lec. 6: Linear Dimensionality Reduction AD January 2011 AD () January 2011 1 / 46 Linear Dimensionality Reduction Introduction & Motivation Brief Review of Linear Algebra Principal Component Analysis

More information

Linear Least Squares. Using SVD Decomposition.

Linear Least Squares. Using SVD Decomposition. Linear Least Squares. Using SVD Decomposition. Dmitriy Leykekhman Spring 2011 Goals SVD-decomposition. Solving LLS with SVD-decomposition. D. Leykekhman Linear Least Squares 1 SVD Decomposition. For any

More information

Singular Value Decomposition (SVD)

Singular Value Decomposition (SVD) School of Computing National University of Singapore CS CS524 Theoretical Foundations of Multimedia More Linear Algebra Singular Value Decomposition (SVD) The highpoint of linear algebra Gilbert Strang

More information

Foundations of Computer Vision

Foundations of Computer Vision Foundations of Computer Vision Wesley. E. Snyder North Carolina State University Hairong Qi University of Tennessee, Knoxville Last Edited February 8, 2017 1 3.2. A BRIEF REVIEW OF LINEAR ALGEBRA Apply

More information

Large Scale Data Analysis Using Deep Learning

Large Scale Data Analysis Using Deep Learning Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset

More information

Linear Algebra Fundamentals

Linear Algebra Fundamentals Linear Algebra Fundamentals It can be argued that all of linear algebra can be understood using the four fundamental subspaces associated with a matrix. Because they form the foundation on which we later

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will

More information

235 Final exam review questions

235 Final exam review questions 5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an

More information

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018 Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry

More information

Symmetric and anti symmetric matrices

Symmetric and anti symmetric matrices Symmetric and anti symmetric matrices In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, matrix A is symmetric if. A = A Because equal matrices have equal

More information

Linear Algebra Primer

Linear Algebra Primer Linear Algebra Primer David Doria daviddoria@gmail.com Wednesday 3 rd December, 2008 Contents Why is it called Linear Algebra? 4 2 What is a Matrix? 4 2. Input and Output.....................................

More information

vector space retrieval many slides courtesy James Amherst

vector space retrieval many slides courtesy James Amherst vector space retrieval many slides courtesy James Allan@umass Amherst 1 what is a retrieval model? Model is an idealization or abstraction of an actual process Mathematical models are used to study the

More information

Manning & Schuetze, FSNLP, (c)

Manning & Schuetze, FSNLP, (c) page 554 554 15 Topics in Information Retrieval co-occurrence Latent Semantic Indexing Term 1 Term 2 Term 3 Term 4 Query user interface Document 1 user interface HCI interaction Document 2 HCI interaction

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2016 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent. Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u

More information

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is, 65 Diagonalizable Matrices It is useful to introduce few more concepts, that are common in the literature Definition 65 The characteristic polynomial of an n n matrix A is the function p(λ) det(a λi) Example

More information

TBP MATH33A Review Sheet. November 24, 2018

TBP MATH33A Review Sheet. November 24, 2018 TBP MATH33A Review Sheet November 24, 2018 General Transformation Matrices: Function Scaling by k Orthogonal projection onto line L Implementation If we want to scale I 2 by k, we use the following: [

More information

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015 Solutions to Final Practice Problems Written by Victoria Kala vtkala@math.ucsb.edu Last updated /5/05 Answers This page contains answers only. See the following pages for detailed solutions. (. (a x. See

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite

More information

CSE 494/598 Lecture-6: Latent Semantic Indexing. **Content adapted from last year s slides

CSE 494/598 Lecture-6: Latent Semantic Indexing. **Content adapted from last year s slides CSE 494/598 Lecture-6: Latent Semantic Indexing LYDIA MANIKONDA HT TP://WWW.PUBLIC.ASU.EDU/~LMANIKON / **Content adapted from last year s slides Announcements Homework-1 and Quiz-1 Project part-2 released

More information

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix DIAGONALIZATION Definition We say that a matrix A of size n n is diagonalizable if there is a basis of R n consisting of eigenvectors of A ie if there are n linearly independent vectors v v n such that

More information

Principal Component Analysis

Principal Component Analysis Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Embeddings Learned By Matrix Factorization

Embeddings Learned By Matrix Factorization Embeddings Learned By Matrix Factorization Benjamin Roth; Folien von Hinrich Schütze Center for Information and Language Processing, LMU Munich Overview WordSpace limitations LinAlgebra review Input matrix

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

Applied Linear Algebra in Geoscience Using MATLAB

Applied Linear Algebra in Geoscience Using MATLAB Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in

More information

Notes on Linear Algebra

Notes on Linear Algebra 1 Notes on Linear Algebra Jean Walrand August 2005 I INTRODUCTION Linear Algebra is the theory of linear transformations Applications abound in estimation control and Markov chains You should be familiar

More information

PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211

PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211 PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211 IIR 18: Latent Semantic Indexing Handout version Petr Sojka, Hinrich Schütze et al. Faculty of Informatics, Masaryk University,

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the

More information

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes Repeated Eigenvalues and Symmetric Matrices. Introduction In this Section we further develop the theory of eigenvalues and eigenvectors in two distinct directions. Firstly we look at matrices where one

More information

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued)

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued) 1 A linear system of equations of the form Sections 75, 78 & 81 a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2 a m1 x 1 + a m2 x 2 + + a mn x n = b m can be written in matrix

More information

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p Math Bootcamp 2012 1 Review of matrix algebra 1.1 Vectors and rules of operations An p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

EE731 Lecture Notes: Matrix Computations for Signal Processing

EE731 Lecture Notes: Matrix Computations for Signal Processing EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University October 17, 005 Lecture 3 3 he Singular Value Decomposition

More information

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.

More information

Linear Algebra and Eigenproblems

Linear Algebra and Eigenproblems Appendix A A Linear Algebra and Eigenproblems A working knowledge of linear algebra is key to understanding many of the issues raised in this work. In particular, many of the discussions of the details

More information

.. CSC 566 Advanced Data Mining Alexander Dekhtyar..

.. CSC 566 Advanced Data Mining Alexander Dekhtyar.. .. CSC 566 Advanced Data Mining Alexander Dekhtyar.. Information Retrieval Latent Semantic Indexing Preliminaries Vector Space Representation of Documents: TF-IDF Documents. A single text document is a

More information

Background Mathematics (2/2) 1. David Barber

Background Mathematics (2/2) 1. David Barber Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and

More information

3 (Maths) Linear Algebra

3 (Maths) Linear Algebra 3 (Maths) Linear Algebra References: Simon and Blume, chapters 6 to 11, 16 and 23; Pemberton and Rau, chapters 11 to 13 and 25; Sundaram, sections 1.3 and 1.5. The methods and concepts of linear algebra

More information

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.

More information

Elementary maths for GMT

Elementary maths for GMT Elementary maths for GMT Linear Algebra Part 2: Matrices, Elimination and Determinant m n matrices The system of m linear equations in n variables x 1, x 2,, x n a 11 x 1 + a 12 x 2 + + a 1n x n = b 1

More information