Latent Semantic Indexing
|
|
- Godfrey Booker
- 6 years ago
- Views:
Transcription
1 Latent Semantic Indexing 1
2 Latent Semantic Indexing Term-document matrices are very large, though most cells are zeros But the number of topics that people talk about is small (in some sense) Clothes, movies, politics, Each topic can be represented as a cluster of (semantically) related terms, e.g.: clothes=golf, jacket, shoe.. Can we represent the term-document space by a lower dimensional latent space (latent space=set of topics)? 2
3 Searching with latent topics Given a collection of documents, LSI learns clusters of frequently co-occurring terms (ex: information retrieval, ranking and web)! If you query with ranking, information retrieval LSI automatically extends the search to documents including also (and even ONLY) web 3
4 4 Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee Selection based on Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee Car GTI Polo Document base (20) Motor Bike Oil Tourer Bed lace legal button soft cat line yellow wind full sail harbour beach report June Speed Fish Pond gold Koi PC Dell RAM Floppy Core Apple Pip Tree Pea Pod Fresh Green French Lupin Seed May April Office Pen Desk VDU Friend Pal Help Can Paper Paste Pencil Roof Card Stamp Glue Happy Send Toil Work Time Cost With standard VSM 4 documents are selected
5 20 documents Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee Car GTI Polo Fish Pond gold Koi PC Dell RAM Floppy Core Apple Pip Tree Pea Pod Fresh Green French Lupin Seed May April Motor Bike Oil Tourer Bed lace legal button soft cat line yellow wind full sail harbour beach report June Speed Office Pen Desk VDU Selection based on Friend Pal Help Can Paper Paste Pencil Roof Card Stamp Glue Happy Send Toil Work Time Cost Rank of selected documents Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee The most relevant words associated with golf in these docs are: Car, and tf.idf Car 2 *(20/3) = 13 2 *(20/3) = 13 3 *(20/16) = 4
6 20 docs Car GTI Car Clarkson Badge Polo Red Tiger Woods Belfry Tee Car GTI Polo Fish Pond gold Koi PC Dell RAM Floppy Core Apple Pip Tree Pea Pod Fresh Green French Lupin Seed May April Motor Bike Oil Tourer Selezione basata su Car GTI Bed lace legal button soft cat line yellow Car Clarkson Badge wind full sail harbour beach report June Speed Polo Red Office Pen Desk VDU Friend Pal Help Can Tiger Woods Belfry Tee Paper Paste Pencil Roof Card Stamp Glue Happy Send If we consider the co-occurring terms with higher tf*idf, car e topgear turn out to be related to more than petrol Toil Work Time Cost Rank of selected docs Car 2 *(20/3) = 13 2 *(20/3) = 13 3 *(20/16) = 4
7 20 docs Car GTI Motor Bike Oil Tourer Car GTI Car Clarkson Badge Bed lace legal button Polo Red soft cat line yellow Car Clarkson Badge Tiger Woods Belfry Tee wind full sail harbour beach Car GTI Polo report June Speed Selection based on Polo Red Fish Pond gold Koi Office Pen Desk VDU PC Dell RAM Floppy Core Apple Pip Tree Tiger Woods Belfry Tee Pea Pod Fresh Green French Lupin Seed May April Selection based on the semantic domain of Car GTI Car Clarkson Badge Polo Red We now search with all the Friend words Paper in the Card semantic Toil domain of. Pal Stamp Help Paste Glue Work The Pencil list Happy of retrieved Time docs now Can Roof Send Cost is based on and the other related words Tiger Woods Belfry Tee Car Wheel GTI Polo rank df selected docs Car 2 *(20/3) = 13 2 *(20/3) = 13 3 *(20/16) = 4
8 20 docs Car GTI Motor Bike Oil Tourer Selezione basata su Car GTI Car Clarkson Badge Polo Red Car Clarkson Badge Tiger Woods Belfry Tee Car GTI Polo Polo Red Tiger Woods Belfry Tee Selection based on semantic domain Car GTI Car Clarkson Badge Polo Red Rank Fish Pond gold Koi PC Dell RAM Floppy Core Apple Pip Tree Tiger Woods Belfry Tee Pea Pod Fresh Green French The co-occurrence based Bed soft wind report Office Friend Paper Card lace full Pen Pal Stamp legal ranking cat sail improves the Desk performance. Help Paste Glue line harbour June Pencil Happy Note button that yellow one beach of the Speed most VDU relevant Can Roof doc Send does NOT include the word, and a doc with a spurious sense disappears Lupin Seed May April Toil Work Time Cost Car Wheel GTI Polo rank of selected docs Car 2 *(20/3) = 13 2 *(20/3) = 13 3 *(20/16) = 4
9 Ranking with latent Semantic Indexing Previous example just gives the intuition Latent Semantic Indexing is an algebraic method to identify clusters of co-occurring terms, called latent topics, and to compute query-doc similarity in a latent space, in which every coordinate is a latent topic. A latent quantity is one which cannot directly observed, what is observed is a measurement which may include some amount of random errors (topics are latent in this sense: we observe them, but they are an approximation of true semantic topics) Since it is an algebraic method, needs some linear algebra background 9
10 The LSI method: how to detect topics Linear Algebra Background 10
11 Eigenvalues & Eigenvectors Eigenvectors (for a square m m matrix S) Av = λv Example (right) eigenvector eigenvalue Av = λv Def: A vector v R n, v 0, is an eigenvector of a matrix mxm A with corresponding eigenvalue λ, if: Av = λv 11
12 Algebraic method How many eigenvalues are there at most? Av = λv equation has a non-zero solution if A A A Where I is the identity matrix this is a m-th order equation in λ which can have at most m distinct solutions (roots of the characteristic polynomial) can be complex even though A is real. 12
13 Example of eigenvector/eigenvalues " A = $ 1 1 # 3 5 " $ # % ', v = " 1 $ & # 3 Av = λv 1 1 %" 1 % ' $ ' 3 5 &# 3 = 4 " 1 $ & # 3 " 1+ 3( 1) % $ # 3+ 5( 3) ' = " 4 % $ ' & # 12 & " 4 % $ ' # 12 = " 4 % $ ' & # 12 & % ', λ = 4 & % ' &
14 Example of computation remember " A = $ 1 1 # 3 5 % ' & det M = M = det(a λi) = 0 " 1 1 % $ ' # 3 5 λ " % and 4 are the $ eigenvalues ' & # 0 1 = 0 of S & (A λi)v = 0 " 1 1 % $ ' # 3 5 " $ λ 0 % " ' & # 0 λ = %" α % λ $ ' $ & # & β ' = " 0 % 2 = 2 $ ' " # & # 0 & β % " v = " 1 λ 1 % $ ' # 3 5 λ = % " α % ' $ $ ' # 3 1 $ &# β ' = " 0 % # β & $ ' & # 0 & & ( * 3α β = 0 ) β = 3α +* 3α + β = 0 a c b d = ad bc (λ 1)(λ 5) + 3 = 0 λ 2 6λ = 0 λ 2 6λ + 8 = 0 (λ 4)(λ 2) = 0 λ 1 = 4,λ 2 = 2 λ Characteristic polynomial = (A 1 λi)v = 0 " &% " α %# $ ' # $ &# β ' = " 0 % v = $ $ ' $ & # 0 & " 1 1 %"% 3α α % $ ' # 3 3 $ &# β ' = " 0 % $ ' & # 0 & ( * α β = 0 ) α = β +* 3α + 3β = 0 '!! " Note that we compute only the DIRECTION of eigenvectors 4
15 Geometric interpretation A matrix-vector multiplication Ax is a linear transformation over the vector 15
16 Matrix vector multiplication Matrix multiplication by a vector = a linear transformation of the initial vector, that implies rotation and translation 16
17 Geometric interpretation 17
18 Geometric interpretation Ax=1.34x=λx An eigenvector is a special vector that is transformed into its scalar multiple under a given matrix (no rotation!) 18
19 Geometric interpretation Here we found another eigenvector for the matrix A Note that for any single eigenvalue you have infinite eigenvectors, but they have the same direction 19
20 Matrix-vector multiplication Eigenvectors of different eigenvalues are linearly independent (i.e. α 1.. α n è α 1 v α n v n 0) For square normal matrixes eigenvectors of different eigenvalues define an orthonormal space and they are othogonal. A square matrix is NORMAL iff it commutes with its transpose, i.e. AA T =A T A Example: è AA T= = A T A 20
21 Difference between orthonormal and orthogonal? Orthogonal mean that the dot product is null (the cosin of the angle is zero). Orthonormal mean that the dot product is null and the norm of the vectos is equal to 1. What we are actually saying is that eigenvectors define a set of DIRECTIONS wich are orthogonal. If two or more vectors are orthonormal they are also orthogonal but the inverse is not true. 21
22 Example: projecting a vector on 2 orthonormal spaces (or bases ) e 1,e 1, e 2,e 2 are unary vectors and v 1,v 1, v 2,v 2 are the coordinates of v along the directions of e 1,e 1, e 2,e 2 22
23 The effect of a matrix-vector multiplication is governed by eigenvectors and eigenvalues A matrix-vector multiplication such as Ax (A normal matrix, x a generic vector as in the previous slide) can be rewritten in terms of the eigenvalues/vectors of A. Example: 2 X= 4 Ax = A(2v 1 + 4v 2 + 6v 3 ) 6 Ax = 2Av 1 + 4Av 2 + 6Av 3 = 2λ 1 v 1 + 4λ 2 v 2 + 6λ 3 v 3 Where v 1,v 2 v 3 are the (orthogonal) eigenvectors of A Even though x is an arbitrary vector, the action of A on x (transformation) is determined by the eigenvalues/vectors. Why? 23
24 Geometric explanation x is a generic vector with coordinates x i ; λ i,v i are the eigenvalues and eigenvectors of A Ax = x 1 λ 1 v 1 + x 2 λ 2 v 2 + x 3 λ 3 v 3 Multiplying a matrix and a vector has two effects over the vector: rotation (the coordinates of the vector change) and scaling (the length changes). The max compression and rotation depends on the matrix s eigenvalues λi (s1,s2 and s3 in the picture)
25 Geometric explanation In the distorsion, the max effect is played by the biggest eigenvalues (s1 and s2 in the picture ) The eigenvalues describe the distorsion operated by the matrix on the original vector
26 Summary so far A matrix A has eigenvectors v and eigenvalues λ, defined by Av=λv Eigenvalues can be computed as: We can compute only the the direction of eigenvectors, since for any eigenvalue there are infinite eigenvectors lying on the same direction If A is normal (i.e. if AA T =A T A) then the eigenvector form an othonormal basis The product of A by ANY vector x is a linear transformation of x where the rotation is determined by eigenvectors and the translation is determined by the eigenvalues. The biggest role in this transformation is played by the biggest eigenvalues. 26
27 Eigen/diagonal Decomposition Let A be a square matrix with m orthogonal eigenvectors (hence, A is normal) Theorem: Exists an eigen decomposition A=UΛU -1 Λ is a diagonal matrix (all zero except the diagonal cells) Columns of U are eigenvectors of A Diagonal elements of Λ are eigenvalues of A 27
28 Diagonal decomposition: why/how = n v v U... 1 Let U have the eigenvectors as columns: AU = A v 1... v n! " # # # $ % & & & = λ 1 v 1... λ n v n! " # # # $ % & & & = v 1... v n! " # # # $ % & & & λ 1... λ n! " # # # # $ % & & & & Then, AU can be written And A=UΛU 1. Thus AU=UΛ, or U 1 AU=Λ 28
29 Example AU=UΛ From this we compute λ 1 =1, λ 2 =3 from which we get v 11 = 2v 12 From which we get v 21 =0 and v 22 any real 29
30 Diagonal decomposition example 2! Recall A = # 2 1 " $ &; λ 1 =1,λ 2 = 3. % 1 1 The eigenvectors and form Inverting, we have U 1 Then, A=UΛU 1 = = 1/ 2 1/ 2 1/ 2 1/ 2 U = 1 1 Recall UU 1 = / 2 1/ / 2 1/
31 So what? What do these matrices have to do with Information Retrieval and document ranking? Recall M N term-document matrices But everything so far needs square normal matrices so you need to be patient and learn one last notion 31
32 Singular Value Decomposition for non-square matrixes For a non-square M N matrix A of rank r there exists a factorization (Singular Value Decomposition = SVD) as follows: A = UΣV T M M M N V is N N The columns of U are orthogonal eigenvectors of AA T (left singular vectors). The columns of V are orthogonal eigenvectors of A T A (right singular eigenvector). NOTE THAT AA T and A T A are square symmetric (and hence NORMAL) Eigenvalues λ 1 λ r of AA T = eigenvalues of A T A and: Σ = ( σ... ) diag 1 σ r Singular values. σ i = λ i 32
33 An example 33
34 Singular Value Decomposition Illustration of SVD dimensions and sparseness M All zeros! So these become zeros, too N MxM MxN NxN So these become zeros, too
35 Back to matrix-vector multiplication Remember what we said? In a matrix vector multiplication the biggest role is played by the biggest eigenvalues The diagonal matrix Σ has the eigenvalues of A T A (called the singular values of A) in decreasing order along the diagonal We can therefore apply an approximation by setting σ i =0 if σ i θ and only consider the first k singular values 35
36 Reduced SVD If we retain only k singular values, and set the rest to 0, then we don t need the matrix parts in red Then Σ is k k, U is M k, V T is k N, and A k is M N This is referred to as the reduced SVD, or rank k approximation Now all the red and yellow parts are zeros!! k 36
37 Let s recap M<N M>N Since the yellow part is zero, an exact representation of A is: A = σ 1 u 1 v 1 T +σ 2 u 2 v 2 T +...+σ r u r v r T r = min(m,n) But for some k<r, a good approximation is: A k = σ 1 u 1 v 1 T +σ 2 u 2 v 2 T σ k u k v k T 37
38 Example of rank approximation 1 = x x 1 A A*= A 38
39 Approximation error How good (bad) is this approximation? It s the best possible, measured by the Frobenius norm of the error: A X A = σ :min F k F k + 1 X rank ( X ) = k = A σ i = λ i where the σ i are ordered such that σ i σ i+1. Suggests why Frobenius error drops as k increases. 39
40 Images gives a better intuition (image = matrix of pixels) 40
41 K=10 41
42 K=20 42
43 K=30 43
44 K=100 44
45 K=322 (the original image) 45
46 We save space!! But this is only one issue 46
47 So, finally, back to IR Our initial problem was: the term-document (MxN) matrix A is highly sparse (has many zeros) However, since groups of terms tend to co-occur together, can we identify the semantic space of these clusters of terms, and apply the vector space model in the semantic space defined by such clusters? What we learned so far: Matrix A can be decomposed and rank k approximated using SVD Does this help solving our initial problems? 47
48 A is our term document matrix Latent Semantic Indexing via the SVD If A is a term/document matrix, then AA T and A T A are the matrixes of term and document co-occurrences, repectively 48
49 Meaning of A T A and AA T Word i in doc k Word j in doc k L = A A T = L ij = A ik A T kj N L ij = A ik A T kj = A ik A jk k=1 N k=1 L ij is the number of documents in which wi and wj A T co-occurr Similarly, L T ij is the number of co-occurring words in docs di and dj (or viceversa if A is a document-term matrix rather than term-document) A
50 Example
51 Term-document matrix A
52 Term co-occurrences example L trees,graph = ( ) ( ) T =2
53 So the matrix L=AA T is the matrix of term co-occurrences Remember: eigenvectors of a matrix define an orthonormal space Remember: bigger eigenvalues define the main directions of this space But: Matrix L and L T are SIMILARITY matrixes (respectively, of terms and of documents). They define a SIMILARITY SPACE (the orthonormal space of their eigenvectors) If the matrix elements are word co-occurrences, bigger eigenvalues are associated to bigger groups of similar words Similarly, bigger eigenvalues of L T =A T A are associated with bigger groups of similar documents (those in which co-occur the same terms)
54 LSI: the intuition The blue segments give the intuition of eigenvalues of L T =A T A Bigger eigenvalues are those for which the projection of all vectors on the direction of correspondent eigenvectors is higher t1 Projecting A in the term space: green, yellow and red vectors are documents. If they form small angles, they have common words (remember cosin-sim) The black vector are the unary eigenvector of A: they represent the main directions of the document vectors t2 t3 54
55 LSI intuition!" d j = (d 1 j,d 2 j,d 3 j )!" L T d j = d 1 j λ 1 v 1 + d 2 j λ 2 v 2 + d 3 j λ 3 v 3 t1 If we multiply all document vectors by L T = A T A, their distorsion is mostly determined by the highest eigenvalues t2 We now project our document vectors on the reference orthonormal system represented by the 3 black vectors t3 55
56 LSI intuition..if we remove the dimension(s) with lower eigenvalues (i.e. if we rank-reduce Σ) we don t loose much information!" d j = (d 1 j,d 2 j,d 3 j )!" L T d j d 1 j λ 1 v 1 + d 2 j λ 2 v 2 Remember that the two new axis represent a combination of co-occurring words e.g. a latent semantic space 56
57 Example t1 t2 t3 t4 d1 d2 d3 d We project terms and docs on two dimensions, v1 and v2 (the principal eigenvectors) x x Note that the direction of each eigenvector is determined by Approximated We have two latent semantic new the term-doc direction of just two terms: (t1,t2) or (t3,t4) matrix coordinates: s1:(t1,t2) and s2:(t3,t4) Even if t2 does not occur in d2, now if we query with t2 the system will return also d2!! 57
58 Co-occurrence space In principle, the space of document or term co-occurrences is much (much!) higher than the original space of terms!! But with SVD we consider only the most relevant ones, trough rank reduction A =U V T U k Σ k V k T = A k 58
59 Summary so-far We compute the SVD rank-k approximation for the term-document matrix A This approximation is based on considering only the principal eigenvalues of the term cooccurrence and document similarity matrixes (L=AA T and LT=A T A) The eigenvectors of the eigenvalues of L=AA T and LT=A T A represent the main directions identified by term vectors and document vectors, respectively. 59
60 LSI: what are the steps From term-doc matrix A, compute the approximation A k. with SVD Project docs, terms and queries in a space of k<<r dimensions (the k survived eigenvectors) and compute cos-similarity as usual These dimensions are not the original axes, but those defined by the orthonormal space of the reduced matrix Ak Aq! A k q! = σ 1 q 1 v! 1 +σ 2 q 2 v! σ k q k v! k Where σ i q i (i=1,2..k<<r) are the new coordinates of q in the orthonormal space of Ak 60
61 Projecting terms documents and queries in the LS space A If A=UΣV T we also have that: V = A T UΣ -1 t = t T ΣV T d = d T UΣ -1 q = q T UΣ -1 After rank k- approximation : A Ak=U k Σ k V k T d k d T U k Σ k -1 q k q T U k Σ k -1 sim(q, d) = sim(q T U k Σ k -1, d T U k Σ k -1 ) 61
62 Consider a term-doc matrix MxN (M=11, N=3) and a query q query A
63 1. Compute SVD: A= UΣVT
64 2. Obtain a low rank approximation (k=2) A k = U k Σ k V T k
65 3a. Compute doc/query similarity For N documents, A k has N columns, each representing the coordinates of a document d i projected in the k LSI dimensions A query is considered like a document, and is projected in the LSI space
66 3c. Compute the query vector q k = q T U k Σ k -1 q is projected in the 2-dimension LSI space!
67 Documents and queries projected in the LSI space
68 q/d similarity
69 An overview of a semantic network of terms based on the top 100 most significant latent semantic dimensions (Zhu&Chen) 69
70 Conclusion LSI performs a low-rank approximation of document-term matrix (typical rank ) General idea Map documents (and terms) to a lowdimensional representation. Design a mapping such that the low-dimensional space reflects semantic associations between words (latent semantic space). Compute document similarity based on the cossim in this latent semantic space
71 Another LSI Example
72
73 t1 t2 t3 t4 d1 d2 d3 A T AA T t1 t2 t3 Term co-occurrences
74 A k =U k Σ k V kt A Now it is like if t3 belongs to d1!
Information Retrieval
Introduction to Information CS276: Information and Web Search Christopher Manning and Pandu Nayak Lecture 13: Latent Semantic Indexing Ch. 18 Today s topic Latent Semantic Indexing Term-document matrices
More informationLatent Semantic Models. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze
Latent Semantic Models Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze 1 Vector Space Model: Pros Automatic selection of index terms Partial matching of queries
More informationLinear Algebra Background
CS76A Text Retrieval and Mining Lecture 5 Recap: Clustering Hierarchical clustering Agglomerative clustering techniques Evaluation Term vs. document space clustering Multi-lingual docs Feature selection
More informationLinear Algebra - Part II
Linear Algebra - Part II Projection, Eigendecomposition, SVD (Adapted from Sargur Srihari s slides) Brief Review from Part 1 Symmetric Matrix: A = A T Orthogonal Matrix: A T A = AA T = I and A 1 = A T
More informationLatent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology
Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2016 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,
More informationLatent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology
Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2014 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,
More informationCS 143 Linear Algebra Review
CS 143 Linear Algebra Review Stefan Roth September 29, 2003 Introductory Remarks This review does not aim at mathematical rigor very much, but instead at ease of understanding and conciseness. Please see
More informationLinear Algebra for Machine Learning. Sargur N. Srihari
Linear Algebra for Machine Learning Sargur N. srihari@cedar.buffalo.edu 1 Overview Linear Algebra is based on continuous math rather than discrete math Computer scientists have little experience with it
More informationLinear Algebra Review. Vectors
Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors
More informationLinear Algebra & Geometry why is linear algebra useful in computer vision?
Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia
More informationCSE 494/598 Lecture-4: Correlation Analysis. **Content adapted from last year s slides
CSE 494/598 Lecture-4: Correlation Analysis LYDIA MANIKONDA HT TP://WWW.PUBLIC.ASU.EDU/~LMANIKON / **Content adapted from last year s slides Announcements Project-1 Due: February 12 th 2016 Analysis report:
More informationMatrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =
30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can
More informationProblems. Looks for literal term matches. Problems:
Problems Looks for literal term matches erms in queries (esp short ones) don t always capture user s information need well Problems: Synonymy: other words with the same meaning Car and automobile 电脑 vs.
More informationDATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD
DATA MINING LECTURE 8 Dimensionality Reduction PCA -- SVD The curse of dimensionality Real data usually have thousands, or millions of dimensions E.g., web documents, where the dimensionality is the vocabulary
More informationUNIT 6: The singular value decomposition.
UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T
More informationMachine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012
Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component
More informationLatent Semantic Analysis. Hongning Wang
Latent Semantic Analysis Hongning Wang CS@UVa VS model in practice Document and query are represented by term vectors Terms are not necessarily orthogonal to each other Synonymy: car v.s. automobile Polysemy:
More informationSingular Value Decomposition
Singular Value Decomposition Motivatation The diagonalization theorem play a part in many interesting applications. Unfortunately not all matrices can be factored as A = PDP However a factorization A =
More informationLet A an n n real nonsymmetric matrix. The eigenvalue problem: λ 1 = 1 with eigenvector u 1 = ( ) λ 2 = 2 with eigenvector u 2 = ( 1
Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Au = λu Example: ( ) 2 0 A = 2 1 λ 1 = 1 with eigenvector
More informationLatent Semantic Analysis. Hongning Wang
Latent Semantic Analysis Hongning Wang CS@UVa Recap: vector space model Represent both doc and query by concept vectors Each concept defines one dimension K concepts define a high-dimensional space Element
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory
More informationDot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.
Dot Products K. Behrend April 3, 008 Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Contents The dot product 3. Length of a vector........................
More informationLinear Algebra Methods for Data Mining
Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 1. Basic Linear Algebra Linear Algebra Methods for Data Mining, Spring 2007, University of Helsinki Example
More informationMaths for Signals and Systems Linear Algebra in Engineering
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE
More informationReview of Linear Algebra
Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=
More informationMATH 583A REVIEW SESSION #1
MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),
More informationLinear Algebra in Actuarial Science: Slides to the lecture
Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations
More informationEIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4]
EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: Au = λu λ C : eigenvalue u C n : eigenvector Example:
More informationSingular Value Decompsition
Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost
More informationMATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)
MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m
More informationMatrices A brief introduction
Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino Semester 1, 2014-15 B. Bona (DAUIN) Matrices Semester 1, 2014-15 1 / 41 Definitions Definition A matrix is a set of N real or complex
More informationStudy Notes on Matrices & Determinants for GATE 2017
Study Notes on Matrices & Determinants for GATE 2017 Matrices and Determinates are undoubtedly one of the most scoring and high yielding topics in GATE. At least 3-4 questions are always anticipated from
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationFunctional Analysis Review
Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all
More informationIntroduction to Matrix Algebra
Introduction to Matrix Algebra August 18, 2010 1 Vectors 1.1 Notations A p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the line. When p
More informationComputational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science
Computational Methods CMSC/AMSC/MAPL 460 Eigenvalues and Eigenvectors Ramani Duraiswami, Dept. of Computer Science Eigen Values of a Matrix Recap: A N N matrix A has an eigenvector x (non-zero) with corresponding
More informationImage Registration Lecture 2: Vectors and Matrices
Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this
More informationLinear Algebra Review. Fei-Fei Li
Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More informationDimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas
Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx
More informationManning & Schuetze, FSNLP (c) 1999,2000
558 15 Topics in Information Retrieval (15.10) y 4 3 2 1 0 0 1 2 3 4 5 6 7 8 Figure 15.7 An example of linear regression. The line y = 0.25x + 1 is the best least-squares fit for the four points (1,1),
More informationlinearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice
3. Eigenvalues and Eigenvectors, Spectral Representation 3.. Eigenvalues and Eigenvectors A vector ' is eigenvector of a matrix K, if K' is parallel to ' and ' 6, i.e., K' k' k is the eigenvalue. If is
More informationBasic Calculus Review
Basic Calculus Review Lorenzo Rosasco ISML Mod. 2 - Machine Learning Vector Spaces Functionals and Operators (Matrices) Vector Space A vector space is a set V with binary operations +: V V V and : R V
More information. = V c = V [x]v (5.1) c 1. c k
Chapter 5 Linear Algebra It can be argued that all of linear algebra can be understood using the four fundamental subspaces associated with a matrix Because they form the foundation on which we later work,
More informationMATH 221, Spring Homework 10 Solutions
MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the
More informationMatrices A brief introduction
Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino Semester 1, 2014-15 B. Bona (DAUIN) Matrices Semester 1, 2014-15 1 / 44 Definitions Definition A matrix is a set of N real or complex
More informationIntroduction to Information Retrieval
Introduction to Information Retrieval http://informationretrieval.org IIR 18: Latent Semantic Indexing Hinrich Schütze Center for Information and Language Processing, University of Munich 2013-07-10 1/43
More informationLinear Algebra & Geometry why is linear algebra useful in computer vision?
Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia
More informationMatrices and Linear Algebra
Contents Quantitative methods for Economics and Business University of Ferrara Academic year 2017-2018 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationMatrix Vector Products
We covered these notes in the tutorial sessions I strongly recommend that you further read the presented materials in classical books on linear algebra Please make sure that you understand the proofs and
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationConceptual Questions for Review
Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.
More informationCS 340 Lec. 6: Linear Dimensionality Reduction
CS 340 Lec. 6: Linear Dimensionality Reduction AD January 2011 AD () January 2011 1 / 46 Linear Dimensionality Reduction Introduction & Motivation Brief Review of Linear Algebra Principal Component Analysis
More informationLinear Least Squares. Using SVD Decomposition.
Linear Least Squares. Using SVD Decomposition. Dmitriy Leykekhman Spring 2011 Goals SVD-decomposition. Solving LLS with SVD-decomposition. D. Leykekhman Linear Least Squares 1 SVD Decomposition. For any
More informationSingular Value Decomposition (SVD)
School of Computing National University of Singapore CS CS524 Theoretical Foundations of Multimedia More Linear Algebra Singular Value Decomposition (SVD) The highpoint of linear algebra Gilbert Strang
More informationFoundations of Computer Vision
Foundations of Computer Vision Wesley. E. Snyder North Carolina State University Hairong Qi University of Tennessee, Knoxville Last Edited February 8, 2017 1 3.2. A BRIEF REVIEW OF LINEAR ALGEBRA Apply
More informationLarge Scale Data Analysis Using Deep Learning
Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset
More informationLinear Algebra Fundamentals
Linear Algebra Fundamentals It can be argued that all of linear algebra can be understood using the four fundamental subspaces associated with a matrix. Because they form the foundation on which we later
More informationCS 246 Review of Linear Algebra 01/17/19
1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector
More informationThe Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)
Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will
More information235 Final exam review questions
5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an
More informationMATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018
Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry
More informationSymmetric and anti symmetric matrices
Symmetric and anti symmetric matrices In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, matrix A is symmetric if. A = A Because equal matrices have equal
More informationLinear Algebra Primer
Linear Algebra Primer David Doria daviddoria@gmail.com Wednesday 3 rd December, 2008 Contents Why is it called Linear Algebra? 4 2 What is a Matrix? 4 2. Input and Output.....................................
More informationvector space retrieval many slides courtesy James Amherst
vector space retrieval many slides courtesy James Allan@umass Amherst 1 what is a retrieval model? Model is an idealization or abstraction of an actual process Mathematical models are used to study the
More informationManning & Schuetze, FSNLP, (c)
page 554 554 15 Topics in Information Retrieval co-occurrence Latent Semantic Indexing Term 1 Term 2 Term 3 Term 4 Query user interface Document 1 user interface HCI interaction Document 2 HCI interaction
More informationLinear Algebra review Powers of a diagonalizable matrix Spectral decomposition
Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2016 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing
More informationj=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.
Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u
More informationA = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,
65 Diagonalizable Matrices It is useful to introduce few more concepts, that are common in the literature Definition 65 The characteristic polynomial of an n n matrix A is the function p(λ) det(a λi) Example
More informationTBP MATH33A Review Sheet. November 24, 2018
TBP MATH33A Review Sheet November 24, 2018 General Transformation Matrices: Function Scaling by k Orthogonal projection onto line L Implementation If we want to scale I 2 by k, we use the following: [
More informationSolutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015
Solutions to Final Practice Problems Written by Victoria Kala vtkala@math.ucsb.edu Last updated /5/05 Answers This page contains answers only. See the following pages for detailed solutions. (. (a x. See
More informationLinear Algebra Review
Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite
More informationCSE 494/598 Lecture-6: Latent Semantic Indexing. **Content adapted from last year s slides
CSE 494/598 Lecture-6: Latent Semantic Indexing LYDIA MANIKONDA HT TP://WWW.PUBLIC.ASU.EDU/~LMANIKON / **Content adapted from last year s slides Announcements Homework-1 and Quiz-1 Project part-2 released
More informationDIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix
DIAGONALIZATION Definition We say that a matrix A of size n n is diagonalizable if there is a basis of R n consisting of eigenvectors of A ie if there are n linearly independent vectors v v n such that
More informationPrincipal Component Analysis
Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used
More informationCOMP 558 lecture 18 Nov. 15, 2010
Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to
More informationEmbeddings Learned By Matrix Factorization
Embeddings Learned By Matrix Factorization Benjamin Roth; Folien von Hinrich Schütze Center for Information and Language Processing, LMU Munich Overview WordSpace limitations LinAlgebra review Input matrix
More informationProbabilistic Latent Semantic Analysis
Probabilistic Latent Semantic Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationApplied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in
More informationNotes on Linear Algebra
1 Notes on Linear Algebra Jean Walrand August 2005 I INTRODUCTION Linear Algebra is the theory of linear transformations Applications abound in estimation control and Markov chains You should be familiar
More informationPV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211
PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211 IIR 18: Latent Semantic Indexing Handout version Petr Sojka, Hinrich Schütze et al. Faculty of Informatics, Masaryk University,
More informationLecture: Face Recognition and Feature Reduction
Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the
More information22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes
Repeated Eigenvalues and Symmetric Matrices. Introduction In this Section we further develop the theory of eigenvalues and eigenvectors in two distinct directions. Firstly we look at matrices where one
More information1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued)
1 A linear system of equations of the form Sections 75, 78 & 81 a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2 a m1 x 1 + a m2 x 2 + + a mn x n = b m can be written in matrix
More informationMath Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p
Math Bootcamp 2012 1 Review of matrix algebra 1.1 Vectors and rules of operations An p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the
More informationLinear Algebra Review. Fei-Fei Li
Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector
More informationEE731 Lecture Notes: Matrix Computations for Signal Processing
EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University October 17, 005 Lecture 3 3 he Singular Value Decomposition
More informationA matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and
Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.
More informationLinear Algebra and Eigenproblems
Appendix A A Linear Algebra and Eigenproblems A working knowledge of linear algebra is key to understanding many of the issues raised in this work. In particular, many of the discussions of the details
More information.. CSC 566 Advanced Data Mining Alexander Dekhtyar..
.. CSC 566 Advanced Data Mining Alexander Dekhtyar.. Information Retrieval Latent Semantic Indexing Preliminaries Vector Space Representation of Documents: TF-IDF Documents. A single text document is a
More informationBackground Mathematics (2/2) 1. David Barber
Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and
More information3 (Maths) Linear Algebra
3 (Maths) Linear Algebra References: Simon and Blume, chapters 6 to 11, 16 and 23; Pemberton and Rau, chapters 11 to 13 and 25; Sundaram, sections 1.3 and 1.5. The methods and concepts of linear algebra
More informationA matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and
Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.
More informationElementary maths for GMT
Elementary maths for GMT Linear Algebra Part 2: Matrices, Elimination and Determinant m n matrices The system of m linear equations in n variables x 1, x 2,, x n a 11 x 1 + a 12 x 2 + + a 1n x n = b 1
More information