Fast LSI-based techniques for query expansion in text retrieval systems

Size: px
Start display at page:

Download "Fast LSI-based techniques for query expansion in text retrieval systems"

Transcription

1 Fast LSI-based techniques for query expansion in text retrieval systems L. Laura U. Nanni F. Sarracco Department of Computer and System Science University of Rome La Sapienza 2nd Workshop on Text-based Information Retrieval Koblenz, 11 Sept. 2005

2 Outline Preliminaries 1 Preliminaries Classical Text Matching Query expansion by Thesauri Spectral Techniques 2 LS-Thesaurus LS-Filter 3 4

3 Term-Documents matrix Classical Text Matching Query expansion by Thesauri Spectral Techniques Collection D = {d 1,..., d n } of text documents. T = {t 1,..., t m }: set of distinct index terms in D: A = t 1 t 2. d 1 d 2 d n a 1,1 a 1,2 a 1,n a 2,1 a 2,2 a 2,n... t m a m,1 a m,2 a m,n a i,j is a function of the weight of term t i in document d j

4 Term-Documents matrix Classical Text Matching Query expansion by Thesauri Spectral Techniques Collection D = {d 1,..., d n } of text documents. T = {t 1,..., t m }: set of distinct index terms in D: A = t 1 t 2. d 1 d 2 d n a 1,1 a 1,2 a 1,n a 2,1 a 2,2 a 2,n... t m a m,1 a m,2 a m,n a i,j is a function of the weight of term t i in document d j

5 Term-Documents matrix Classical Text Matching Query expansion by Thesauri Spectral Techniques Collection D = {d 1,..., d n } of text documents. T = {t 1,..., t m }: set of distinct index terms in D: A = t 1 t 2. d 1 d 2 d n a 1,1 a 1,2 a 1,n a 2,1 a 2,2 a 2,n... t m a m,1 a m,2 a m,n a i,j is a function of the weight of term t i in document d j

6 Term-Documents matrix Classical Text Matching Query expansion by Thesauri Spectral Techniques Collection D = {d 1,..., d n } of text documents. T = {t 1,..., t m }: set of distinct index terms in D: A = t 1 t 2. d 1 d 2 d n a 1,1 a 1,2 a 1,n a 2,1 a 2,2 a 2,n... t m a m,1 a m,2 a m,n a i,j is a function of the weight of term t i in document d j

7 TF-IDF term-weighting schemes Classical Text Matching Query expansion by Thesauri Spectral Techniques The value of a i,j is a function of two factors A LOCAL FACTOR L(i, j) measuring the relevance of term t i in document d j. We used: freq(i, j) L(i, j) = max i [1,m] freq(i, j) A GLOBAL FACTOR G(i) to de-amplify the relative weight of terms which are very frequently used in the collection. We used: G(i) = log n n i n i is the number of documents containing term t i.

8 TF-IDF term-weighting schemes Classical Text Matching Query expansion by Thesauri Spectral Techniques The value of a i,j is a function of two factors A LOCAL FACTOR L(i, j) measuring the relevance of term t i in document d j. We used: freq(i, j) L(i, j) = max i [1,m] freq(i, j) A GLOBAL FACTOR G(i) to de-amplify the relative weight of terms which are very frequently used in the collection. We used: G(i) = log n n i n i is the number of documents containing term t i.

9 TF-IDF term-weighting schemes Classical Text Matching Query expansion by Thesauri Spectral Techniques The value of a i,j is a function of two factors A LOCAL FACTOR L(i, j) measuring the relevance of term t i in document d j. We used: freq(i, j) L(i, j) = max i [1,m] freq(i, j) A GLOBAL FACTOR G(i) to de-amplify the relative weight of terms which are very frequently used in the collection. We used: G(i) = log n n i n i is the number of documents containing term t i.

10 Text Matching Algorithm Classical Text Matching Query expansion by Thesauri Spectral Techniques Input: a query vector q = {q 1,..., q m } Output: the rank vector r = q A A and q are sparse q A can be computed very efficiently. If the size of the query is limited to k terms TM has cost O(kn). Issues Polysemy: e.g., polo Synonymy: e.g., automobile, car, machine

11 Text Matching Algorithm Classical Text Matching Query expansion by Thesauri Spectral Techniques Input: a query vector q = {q 1,..., q m } Output: the rank vector r = q A A and q are sparse q A can be computed very efficiently. If the size of the query is limited to k terms TM has cost O(kn). Issues Polysemy: e.g., polo Synonymy: e.g., automobile, car, machine

12 Text Matching Algorithm Classical Text Matching Query expansion by Thesauri Spectral Techniques Input: a query vector q = {q 1,..., q m } Output: the rank vector r = q A A and q are sparse q A can be computed very efficiently. If the size of the query is limited to k terms TM has cost O(kn). Issues Polysemy: e.g., polo Synonymy: e.g., automobile, car, machine

13 Text Matching Algorithm Classical Text Matching Query expansion by Thesauri Spectral Techniques Input: a query vector q = {q 1,..., q m } Output: the rank vector r = q A A and q are sparse q A can be computed very efficiently. If the size of the query is limited to k terms TM has cost O(kn). Issues Polysemy: e.g., polo Synonymy: e.g., automobile, car, machine

14 Text Matching Algorithm Classical Text Matching Query expansion by Thesauri Spectral Techniques Input: a query vector q = {q 1,..., q m } Output: the rank vector r = q A A and q are sparse q A can be computed very efficiently. If the size of the query is limited to k terms TM has cost O(kn). Issues Polysemy: e.g., polo Synonymy: e.g., automobile, car, machine

15 Query expansion Preliminaries Classical Text Matching Query expansion by Thesauri Spectral Techniques QUERY EXPANSION (or QUERY REWEIGHTING) The process aimed to alter the weights, and possibly the terms, of a query. Two approaches: 1 use relevance feedback 2 use some knowledge on terms relationship (thesaurus) The term-term correlation matrix A A T gives a statistic estimation of relationships among terms in the collection Query expansion: q q A A T

16 Query expansion Preliminaries Classical Text Matching Query expansion by Thesauri Spectral Techniques QUERY EXPANSION (or QUERY REWEIGHTING) The process aimed to alter the weights, and possibly the terms, of a query. Two approaches: 1 use relevance feedback 2 use some knowledge on terms relationship (thesaurus) The term-term correlation matrix A A T gives a statistic estimation of relationships among terms in the collection Query expansion: q q A A T

17 Query expansion Preliminaries Classical Text Matching Query expansion by Thesauri Spectral Techniques QUERY EXPANSION (or QUERY REWEIGHTING) The process aimed to alter the weights, and possibly the terms, of a query. Two approaches: 1 use relevance feedback 2 use some knowledge on terms relationship (thesaurus) The term-term correlation matrix A A T gives a statistic estimation of relationships among terms in the collection Query expansion: q q A A T

18 Similarity Thesaurus Preliminaries Classical Text Matching Query expansion by Thesauri Spectral Techniques Qiu and Frei presented an alternative approach to compute A Idea: compute the probability that a document is representive of a term They propose the following weighting scheme: a i,j = freq(i,j) ( ) itf (j) max j freq(i,j) r Pn l=1 (( freq(i,l) max l freq(i,l) ) itf (j))2 if freq(i, j) > 0 0 otherwise max j freq(i, j): is the maximum frequency of term t i over all the documents in the collection. itf j = log m m j, m j the number of distinct terms in the document d j ;

19 Classical Text Matching Query expansion by Thesauri Spectral Techniques Query expansion by Similarity Thesaurus Algorithm 1: Compute A with the previous weighting function; 2: Compute similarity thesaurus: S A A T ; 3: Given a query vector q: 4: s q S; 5: s ξ( s, x r ); 6: s s q, where q = m i=1 q i; 7: q q + s ; Search quality Use of S.T. improves Text-Matching but does not completely resolve polysemy and synonymy.

20 Classical Text Matching Query expansion by Thesauri Spectral Techniques Query expansion by Similarity Thesaurus Algorithm 1: Compute A with the previous weighting function; 2: Compute similarity thesaurus: S A A T ; 3: Given a query vector q: 4: s q S; 5: s ξ( s, x r ); 6: s s q, where q = m i=1 q i; 7: q q + s ; Search quality Use of S.T. improves Text-Matching but does not completely resolve polysemy and synonymy.

21 Latent Semantic Indexing Classical Text Matching Query expansion by Thesauri Spectral Techniques Idea Use the Singular Value Decomposition to produce a low rank approximation of A n r r n m m A = U r S r V T

22 Latent Semantic Indexing Classical Text Matching Query expansion by Thesauri Spectral Techniques Idea Use the Singular Value Decomposition to produce a low rank approximation of A n k k n m m A k = U k S k V k k k T

23 LSI Search Preliminaries Classical Text Matching Query expansion by Thesauri Spectral Techniques Issue The documents collection is represented in the reduced k-dimensional subspace: D = Σ 1 k Uk T A Also the query vector is projected in the k-dimensional subspace: q k = Σ 1 k Uk T q The rank of i-th document is given by r i = cos α i = q k d i q k d i Requires the computation of n cosines very slow, unusable for large collections

24 LSI Search Preliminaries Classical Text Matching Query expansion by Thesauri Spectral Techniques Issue The documents collection is represented in the reduced k-dimensional subspace: D = Σ 1 k Uk T A Also the query vector is projected in the k-dimensional subspace: q k = Σ 1 k Uk T q The rank of i-th document is given by r i = cos α i = q k d i q k d i Requires the computation of n cosines very slow, unusable for large collections

25 LSI Search Preliminaries Classical Text Matching Query expansion by Thesauri Spectral Techniques Issue The documents collection is represented in the reduced k-dimensional subspace: D = Σ 1 k Uk T A Also the query vector is projected in the k-dimensional subspace: q k = Σ 1 k Uk T q The rank of i-th document is given by r i = cos α i = q k d i q k d i Requires the computation of n cosines very slow, unusable for large collections

26 LSI Search Preliminaries Classical Text Matching Query expansion by Thesauri Spectral Techniques Issue The documents collection is represented in the reduced k-dimensional subspace: D = Σ 1 k Uk T A Also the query vector is projected in the k-dimensional subspace: q k = Σ 1 k Uk T q The rank of i-th document is given by r i = cos α i = q k d i q k d i Requires the computation of n cosines very slow, unusable for large collections

27 LS-Thesaurus Preliminaries LS-Thesaurus LS-Filter Idea Compute the similarity matrix starting from a low rank approximation of A (as in LSI). We can formally define our similarity matrix S k in the following way: S k = A k A T k = U k Σ k V T k V k Σ T k U T k = U k Σ 2 k U T k Note that U and the diagonal elements of Σ correspond respectively to the eigenvectors and the eigenvalues of matrix A A T.

28 LS-Thesaurus LS-Filter Algorithm LS-Thesaurus - Pre-Process 1: Input: a collection of documents D; 2: Output: a Similarity Thesaurus, i.e., a m m matrix S k ; 3: 4: Compute the term-document matrix A from D; 5: (U, Λ) EIGEN(A A T ); 6: U k U (1:m,1:k) ; 7: Λ k Λ (1:k,1:k) ; 8: S k U k Λ k U T k ;

29 LS-Thesaurus LS-Filter Algorithm LS-Thesaurus - Expand 1: Input: a query vector q, a similarity thesaurus S k, 2: a positive integer x r ; 3: Output: a query vector q ; 4: 5: s q S k ; 6: s ξ( s, x r ); 7: s s q, where q = m i=1 q i; 8: q q + s ;

30 LS-Filter Preliminaries LS-Thesaurus LS-Filter Assumptions In LSI algorithm documents and queries are projected (and compared) in a k-dimensional subspace The axes of this subspace represent the k most important concepts arising from the documents in the collection The user query tries to catch one (or more) of these concepts by using an appropriate set of terms Idea Let the system select the terms which best represent the required concepts

31 LS-Filter Preliminaries LS-Thesaurus LS-Filter Assumptions In LSI algorithm documents and queries are projected (and compared) in a k-dimensional subspace The axes of this subspace represent the k most important concepts arising from the documents in the collection The user query tries to catch one (or more) of these concepts by using an appropriate set of terms Idea Let the system select the terms which best represent the required concepts

32 LS-Filter - Schematic view LS-Thesaurus LS-Filter 3 4 P tc 6 P ct 8 Search engine The user inserts a query

33 LS-Filter - Schematic view LS-Thesaurus LS-Filter 3 4 P tc 6 P ct 8 Search engine The query vector is projected in the concepts space

34 LS-Filter - Schematic view LS-Thesaurus LS-Filter 3 4 P tc 6 P ct 8 Search engine Minor concepts are removed

35 LS-Filter - Schematic view LS-Thesaurus LS-Filter 3 4 P tc 6 P ct 8 Search engine Concepts vector is projected back into the terms space

36 LS-Filter - Schematic view LS-Thesaurus LS-Filter 3 4 P tc 6 P ct 8 Search engine Minor terms are removed

37 LS-Filter - Schematic view LS-Thesaurus LS-Filter 3 4 P tc 6 P ct 8 Search engine Remaining terms are added to the original query vector

38 LS-Thesaurus LS-Filter Algorithm LS-Filter - Pre-Process 1: Input: a collection of documents D; 2: Output: a pair of matrices (P, P 1 ); 3: 4: Compute the term-document matrix A from D; 5: (U, Σ, V ) SVD(A); 6: U k U (1:m,1:k) ; 7: Σ k Σ (1:k,1:k) ; 8: P Σ 1 k U T k ; 9: P 1 U k Σ k ;

39 Algorithm LS-Filter - Expand LS-Thesaurus LS-Filter 1: Input: a query vector q, matrices (P, P 1 ), 2: two positive integers x c and x t ; 3: Output: a query vector q ; 4: 5: p P q; 6: p ξ( p, x c ); 7: p P 1 p ; 8: q ξ( p, x t );

40 Algorithms Preliminaries We compared the behavior of the following approaches: TM: the simple text matching; LS-T: text matching with queries previously expanded by LS-Thesaurus algorithm; LS-F: text matching with queries previously expanded by LS-Filter algorithm; LSI: the full LSI computation.

41 Dataset Preliminaries Three books from O Reilly in html format about Perl, Unix and Java 3000 documents (html pages) 150 short queries, with human made collection of relevant documents publicly available on the web at URL:

42 Retrieval effectivess 100 LSI TM LS-F LS-T 80 % precision % recall

43 Time performances LSI TM, LS-T, LS-F 8 7 LSI - time (seconds) TM, LS-T & LS-F - time (seconds) number of documents

44 Examples of LS-Filter job control shell logout zip file job shell file background bourne util echo login zip number perl checksum list prompt control ctrl object line filename

45 What we have done Previous techniques: TM: fast but prone to synonymy and polysemy LSI: effective but slow We introduced 2 techniques to improve TM search by using query expansion We compared them with TM and LSI search: we used a non-standard mid-size dataset generally their performances are better than TM, but not homogeneous

46 What we have done Previous techniques: TM: fast but prone to synonymy and polysemy LSI: effective but slow We introduced 2 techniques to improve TM search by using query expansion We compared them with TM and LSI search: we used a non-standard mid-size dataset generally their performances are better than TM, but not homogeneous

47 What we have done Previous techniques: TM: fast but prone to synonymy and polysemy LSI: effective but slow We introduced 2 techniques to improve TM search by using query expansion We compared them with TM and LSI search: we used a non-standard mid-size dataset generally their performances are better than TM, but not homogeneous

48 What we want to do Further experiments on standard datasets Exploit user relevance feedback to discriminate relevant concepts in case of ambiguous queries Very big data structures: can we decrease spatial cost?

49 Bibliography I Preliminaries H. Bast and D. Majumdar. Why spectral retrieval works. In Proocedings of ACM SIGIR, S. Deerwester, S.T. Dumais, T.K. Landauer, G.W. Furnas, and R.A. Harshman. Indexing by latent semantic analysis. J.Soc.Inform.Sci., 41(6): , C. Papadimitriou, P. Raghavan, H. Tamaki, and S. Vempala Latent semantic indexing: A probabilistic analysis. J. Comput. Systems Sciences, 61(2): , 2000.

50 Bibliography II Preliminaries Y. Qiu and H. Frei. Improving the retrieval effectiveness by a similarity thesaurus. Technical Report 225, ETH Zurich, 1994.

51 Thank you

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2014 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,

More information

INF 141 IR METRICS LATENT SEMANTIC ANALYSIS AND INDEXING. Crista Lopes

INF 141 IR METRICS LATENT SEMANTIC ANALYSIS AND INDEXING. Crista Lopes INF 141 IR METRICS LATENT SEMANTIC ANALYSIS AND INDEXING Crista Lopes Outline Precision and Recall The problem with indexing so far Intuition for solving it Overview of the solution The Math How to measure

More information

Problems. Looks for literal term matches. Problems:

Problems. Looks for literal term matches. Problems: Problems Looks for literal term matches erms in queries (esp short ones) don t always capture user s information need well Problems: Synonymy: other words with the same meaning Car and automobile 电脑 vs.

More information

Information Retrieval

Information Retrieval Introduction to Information CS276: Information and Web Search Christopher Manning and Pandu Nayak Lecture 13: Latent Semantic Indexing Ch. 18 Today s topic Latent Semantic Indexing Term-document matrices

More information

Manning & Schuetze, FSNLP (c) 1999,2000

Manning & Schuetze, FSNLP (c) 1999,2000 558 15 Topics in Information Retrieval (15.10) y 4 3 2 1 0 0 1 2 3 4 5 6 7 8 Figure 15.7 An example of linear regression. The line y = 0.25x + 1 is the best least-squares fit for the four points (1,1),

More information

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2016 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,

More information

CS47300: Web Information Search and Management

CS47300: Web Information Search and Management CS47300: Web Information Search and Management Prof. Chris Clifton 6 September 2017 Material adapted from course created by Dr. Luo Si, now leading Alibaba research group 1 Vector Space Model Disadvantages:

More information

Latent Semantic Analysis. Hongning Wang

Latent Semantic Analysis. Hongning Wang Latent Semantic Analysis Hongning Wang CS@UVa Recap: vector space model Represent both doc and query by concept vectors Each concept defines one dimension K concepts define a high-dimensional space Element

More information

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012 Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component

More information

Latent Semantic Analysis. Hongning Wang

Latent Semantic Analysis. Hongning Wang Latent Semantic Analysis Hongning Wang CS@UVa VS model in practice Document and query are represented by term vectors Terms are not necessarily orthogonal to each other Synonymy: car v.s. automobile Polysemy:

More information

.. CSC 566 Advanced Data Mining Alexander Dekhtyar..

.. CSC 566 Advanced Data Mining Alexander Dekhtyar.. .. CSC 566 Advanced Data Mining Alexander Dekhtyar.. Information Retrieval Latent Semantic Indexing Preliminaries Vector Space Representation of Documents: TF-IDF Documents. A single text document is a

More information

Latent Semantic Models. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze

Latent Semantic Models. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze Latent Semantic Models Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze 1 Vector Space Model: Pros Automatic selection of index terms Partial matching of queries

More information

Retrieval by Content. Part 2: Text Retrieval Term Frequency and Inverse Document Frequency. Srihari: CSE 626 1

Retrieval by Content. Part 2: Text Retrieval Term Frequency and Inverse Document Frequency. Srihari: CSE 626 1 Retrieval by Content Part 2: Text Retrieval Term Frequency and Inverse Document Frequency Srihari: CSE 626 1 Text Retrieval Retrieval of text-based information is referred to as Information Retrieval (IR)

More information

Variable Latent Semantic Indexing

Variable Latent Semantic Indexing Variable Latent Semantic Indexing Prabhakar Raghavan Yahoo! Research Sunnyvale, CA November 2005 Joint work with A. Dasgupta, R. Kumar, A. Tomkins. Yahoo! Research. Outline 1 Introduction 2 Background

More information

Manning & Schuetze, FSNLP, (c)

Manning & Schuetze, FSNLP, (c) page 554 554 15 Topics in Information Retrieval co-occurrence Latent Semantic Indexing Term 1 Term 2 Term 3 Term 4 Query user interface Document 1 user interface HCI interaction Document 2 HCI interaction

More information

A Note on the Effect of Term Weighting on Selecting Intrinsic Dimensionality of Data

A Note on the Effect of Term Weighting on Selecting Intrinsic Dimensionality of Data BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 9, No 1 Sofia 2009 A Note on the Effect of Term Weighting on Selecting Intrinsic Dimensionality of Data Ch. Aswani Kumar 1,

More information

INFO 4300 / CS4300 Information Retrieval. slides adapted from Hinrich Schütze s, linked from

INFO 4300 / CS4300 Information Retrieval. slides adapted from Hinrich Schütze s, linked from INFO 4300 / CS4300 Information Retrieval slides adapted from Hinrich Schütze s, linked from http://informationretrieval.org/ IR 13: Query Expansion and Probabilistic Retrieval Paul Ginsparg Cornell University,

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

Can Vector Space Bases Model Context?

Can Vector Space Bases Model Context? Can Vector Space Bases Model Context? Massimo Melucci University of Padua Department of Information Engineering Via Gradenigo, 6/a 35031 Padova Italy melo@dei.unipd.it Abstract Current Information Retrieval

More information

Matrix decompositions and latent semantic indexing

Matrix decompositions and latent semantic indexing 18 Matrix decompositions and latent semantic indexing On page 113, we introduced the notion of a term-document matrix: an M N matrix C, each of whose rows represents a term and each of whose columns represents

More information

Why Spectral Retrieval Works

Why Spectral Retrieval Works Why Spectral Retrieval Works Holger Bast Max-Planck-Institut für Informatik Saarbrücken, Germany bast@mpi-inf.mpg.de Debapriyo Majumdar Max-Planck-Institut für Informatik Saarbrücken, Germany deb@mpi-inf.mpg.de

More information

Latent Semantic Tensor Indexing for Community-based Question Answering

Latent Semantic Tensor Indexing for Community-based Question Answering Latent Semantic Tensor Indexing for Community-based Question Answering Xipeng Qiu, Le Tian, Xuanjing Huang Fudan University, 825 Zhangheng Road, Shanghai, China xpqiu@fudan.edu.cn, tianlefdu@gmail.com,

More information

Latent semantic indexing

Latent semantic indexing Latent semantic indexing Relationship between concepts and words is many-to-many. Solve problems of synonymy and ambiguity by representing documents as vectors of ideas or concepts, not terms. For retrieval,

More information

EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4]

EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: Au = λu λ C : eigenvalue u C n : eigenvector Example:

More information

Notes on Latent Semantic Analysis

Notes on Latent Semantic Analysis Notes on Latent Semantic Analysis Costas Boulis 1 Introduction One of the most fundamental problems of information retrieval (IR) is to find all documents (and nothing but those) that are semantically

More information

CS 3750 Advanced Machine Learning. Applications of SVD and PCA (LSA and Link analysis) Cem Akkaya

CS 3750 Advanced Machine Learning. Applications of SVD and PCA (LSA and Link analysis) Cem Akkaya CS 375 Advanced Machine Learning Applications of SVD and PCA (LSA and Link analysis) Cem Akkaya Outline SVD and LSI Kleinberg s Algorithm PageRank Algorithm Vector Space Model Vector space model represents

More information

1. Ignoring case, extract all unique words from the entire set of documents.

1. Ignoring case, extract all unique words from the entire set of documents. CS 378 Introduction to Data Mining Spring 29 Lecture 2 Lecturer: Inderjit Dhillon Date: Jan. 27th, 29 Keywords: Vector space model, Latent Semantic Indexing(LSI), SVD 1 Vector Space Model The basic idea

More information

Linear Algebra Background

Linear Algebra Background CS76A Text Retrieval and Mining Lecture 5 Recap: Clustering Hierarchical clustering Agglomerative clustering techniques Evaluation Term vs. document space clustering Multi-lingual docs Feature selection

More information

Introduction to Information Retrieval

Introduction to Information Retrieval Introduction to Information Retrieval http://informationretrieval.org IIR 18: Latent Semantic Indexing Hinrich Schütze Center for Information and Language Processing, University of Munich 2013-07-10 1/43

More information

Let A an n n real nonsymmetric matrix. The eigenvalue problem: λ 1 = 1 with eigenvector u 1 = ( ) λ 2 = 2 with eigenvector u 2 = ( 1

Let A an n n real nonsymmetric matrix. The eigenvalue problem: λ 1 = 1 with eigenvector u 1 = ( ) λ 2 = 2 with eigenvector u 2 = ( 1 Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Au = λu Example: ( ) 2 0 A = 2 1 λ 1 = 1 with eigenvector

More information

Matrix Factorization & Latent Semantic Analysis Review. Yize Li, Lanbo Zhang

Matrix Factorization & Latent Semantic Analysis Review. Yize Li, Lanbo Zhang Matrix Factorization & Latent Semantic Analysis Review Yize Li, Lanbo Zhang Overview SVD in Latent Semantic Indexing Non-negative Matrix Factorization Probabilistic Latent Semantic Indexing Vector Space

More information

Fall CS646: Information Retrieval. Lecture 6 Boolean Search and Vector Space Model. Jiepu Jiang University of Massachusetts Amherst 2016/09/26

Fall CS646: Information Retrieval. Lecture 6 Boolean Search and Vector Space Model. Jiepu Jiang University of Massachusetts Amherst 2016/09/26 Fall 2016 CS646: Information Retrieval Lecture 6 Boolean Search and Vector Space Model Jiepu Jiang University of Massachusetts Amherst 2016/09/26 Outline Today Boolean Retrieval Vector Space Model Latent

More information

Natural Language Processing. Topics in Information Retrieval. Updated 5/10

Natural Language Processing. Topics in Information Retrieval. Updated 5/10 Natural Language Processing Topics in Information Retrieval Updated 5/10 Outline Introduction to IR Design features of IR systems Evaluation measures The vector space model Latent semantic indexing Background

More information

Eigenvalue Problems Computation and Applications

Eigenvalue Problems Computation and Applications Eigenvalue ProblemsComputation and Applications p. 1/36 Eigenvalue Problems Computation and Applications Che-Rung Lee cherung@gmail.com National Tsing Hua University Eigenvalue ProblemsComputation and

More information

9 Searching the Internet with the SVD

9 Searching the Internet with the SVD 9 Searching the Internet with the SVD 9.1 Information retrieval Over the last 20 years the number of internet users has grown exponentially with time; see Figure 1. Trying to extract information from this

More information

PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211

PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211 PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211 IIR 18: Latent Semantic Indexing Handout version Petr Sojka, Hinrich Schütze et al. Faculty of Informatics, Masaryk University,

More information

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD DATA MINING LECTURE 8 Dimensionality Reduction PCA -- SVD The curse of dimensionality Real data usually have thousands, or millions of dimensions E.g., web documents, where the dimensionality is the vocabulary

More information

RETRIEVAL MODELS. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

RETRIEVAL MODELS. Dr. Gjergji Kasneci Introduction to Information Retrieval WS RETRIEVAL MODELS Dr. Gjergji Kasneci Introduction to Information Retrieval WS 2012-13 1 Outline Intro Basics of probability and information theory Retrieval models Boolean model Vector space model Probabilistic

More information

Data Mining and Matrices

Data Mining and Matrices Data Mining and Matrices 10 Graphs II Rainer Gemulla, Pauli Miettinen Jul 4, 2013 Link analysis The web as a directed graph Set of web pages with associated textual content Hyperlinks between webpages

More information

Lecture 5: Web Searching using the SVD

Lecture 5: Web Searching using the SVD Lecture 5: Web Searching using the SVD Information Retrieval Over the last 2 years the number of internet users has grown exponentially with time; see Figure. Trying to extract information from this exponentially

More information

CS 572: Information Retrieval

CS 572: Information Retrieval CS 572: Information Retrieval Lecture 11: Topic Models Acknowledgments: Some slides were adapted from Chris Manning, and from Thomas Hoffman 1 Plan for next few weeks Project 1: done (submit by Friday).

More information

13 Searching the Web with the SVD

13 Searching the Web with the SVD 13 Searching the Web with the SVD 13.1 Information retrieval Over the last 20 years the number of internet users has grown exponentially with time; see Figure 1. Trying to extract information from this

More information

Information Retrieval and Topic Models. Mausam (Based on slides of W. Arms, Dan Jurafsky, Thomas Hofmann, Ata Kaban, Chris Manning, Melanie Martin)

Information Retrieval and Topic Models. Mausam (Based on slides of W. Arms, Dan Jurafsky, Thomas Hofmann, Ata Kaban, Chris Manning, Melanie Martin) Information Retrieval and Topic Models Mausam (Based on slides of W. Arms, Dan Jurafsky, Thomas Hofmann, Ata Kaban, Chris Manning, Melanie Martin) Sec. 1.1 Unstructured data in 1620 Which plays of Shakespeare

More information

Boolean and Vector Space Retrieval Models

Boolean and Vector Space Retrieval Models Boolean and Vector Space Retrieval Models Many slides in this section are adapted from Prof. Joydeep Ghosh (UT ECE) who in turn adapted them from Prof. Dik Lee (Univ. of Science and Tech, Hong Kong) 1

More information

Generic Text Summarization

Generic Text Summarization June 27, 2012 Outline Introduction 1 Introduction Notation and Terminology 2 3 4 5 6 Text Summarization Introduction Notation and Terminology Two Types of Text Summarization Query-Relevant Summarization:

More information

Intelligent Data Analysis. Mining Textual Data. School of Computer Science University of Birmingham

Intelligent Data Analysis. Mining Textual Data. School of Computer Science University of Birmingham Intelligent Data Analysis Mining Textual Data Peter Tiňo School of Computer Science University of Birmingham Representing documents as numerical vectors Use a special set of terms T = {t 1, t 2,..., t

More information

CSE 494/598 Lecture-4: Correlation Analysis. **Content adapted from last year s slides

CSE 494/598 Lecture-4: Correlation Analysis. **Content adapted from last year s slides CSE 494/598 Lecture-4: Correlation Analysis LYDIA MANIKONDA HT TP://WWW.PUBLIC.ASU.EDU/~LMANIKON / **Content adapted from last year s slides Announcements Project-1 Due: February 12 th 2016 Analysis report:

More information

Dimensionality Reduction

Dimensionality Reduction Dimensionality Reduction Given N vectors in n dims, find the k most important axes to project them k is user defined (k < n) Applications: information retrieval & indexing identify the k most important

More information

Insights from Viewing Ranked Retrieval as Rank Aggregation

Insights from Viewing Ranked Retrieval as Rank Aggregation Insights from Viewing Ranked Retrieval as Rank Aggregation Holger Bast Ingmar Weber Max-Planck-Institut für Informatik Stuhlsatzenhausweg 85 66123 Saarbrücken, Germany {bast, iweber}@mpi-sb.mpg.de Abstract

More information

Semantics with Dense Vectors. Reference: D. Jurafsky and J. Martin, Speech and Language Processing

Semantics with Dense Vectors. Reference: D. Jurafsky and J. Martin, Speech and Language Processing Semantics with Dense Vectors Reference: D. Jurafsky and J. Martin, Speech and Language Processing 1 Semantics with Dense Vectors We saw how to represent a word as a sparse vector with dimensions corresponding

More information

Geometric and Quantum Methods for Information Retrieval

Geometric and Quantum Methods for Information Retrieval PAPER Geometric and Quantum Methods for Information Retrieval Yaoyong Li, Hamish Cunningham Department of Computer Science, University of Sheffield, Sheffield, UK {yaoyong;hamish}@dcs.shef.ac.uk Abstract

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

Singular Value Decompsition

Singular Value Decompsition Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost

More information

Latent Semantic Analysis (Tutorial)

Latent Semantic Analysis (Tutorial) Latent Semantic Analysis (Tutorial) Alex Thomo Eigenvalues and Eigenvectors Let A be an n n matrix with elements being real numbers. If x is an n-dimensional vector, then the matrix-vector product Ax is

More information

Updating the Partial Singular Value Decomposition in Latent Semantic Indexing

Updating the Partial Singular Value Decomposition in Latent Semantic Indexing Updating the Partial Singular Value Decomposition in Latent Semantic Indexing Jane E. Tougas a,1, a Faculty of Computer Science, Dalhousie University, Halifax, NS, B3H 1W5, Canada Raymond J. Spiteri b,2

More information

PROBABILISTIC LATENT SEMANTIC ANALYSIS

PROBABILISTIC LATENT SEMANTIC ANALYSIS PROBABILISTIC LATENT SEMANTIC ANALYSIS Lingjia Deng Revised from slides of Shuguang Wang Outline Review of previous notes PCA/SVD HITS Latent Semantic Analysis Probabilistic Latent Semantic Analysis Applications

More information

Polynomial Filtering in Latent Semantic Indexing for Information Retrieval

Polynomial Filtering in Latent Semantic Indexing for Information Retrieval Polynomial Filtering in Latent Semantic Indexing for Information Retrieval E. Koiopoulou Yousef Saad May 12, 24 Abstract Latent Semantic Indexing (LSI) is a well established and effective framewor for

More information

CSE 494/598 Lecture-6: Latent Semantic Indexing. **Content adapted from last year s slides

CSE 494/598 Lecture-6: Latent Semantic Indexing. **Content adapted from last year s slides CSE 494/598 Lecture-6: Latent Semantic Indexing LYDIA MANIKONDA HT TP://WWW.PUBLIC.ASU.EDU/~LMANIKON / **Content adapted from last year s slides Announcements Homework-1 and Quiz-1 Project part-2 released

More information

A few applications of the SVD

A few applications of the SVD A few applications of the SVD Many methods require to approximate the original data (matrix) by a low rank matrix before attempting to solve the original problem Regularization methods require the solution

More information

Updating the Centroid Decomposition with Applications in LSI

Updating the Centroid Decomposition with Applications in LSI Updating the Centroid Decomposition with Applications in LSI Jason R. Blevins and Moody T. Chu Department of Mathematics, N.C. State University May 14, 24 Abstract The centroid decomposition (CD) is an

More information

LSI vs. Wordnet Ontology in Dimension Reduction for Information Retrieval

LSI vs. Wordnet Ontology in Dimension Reduction for Information Retrieval LSI vs. Wordnet Ontology in Dimension Reduction for Information Retrieval Pavel Moravec, Michal Kolovrat, and Václav Snášel Department of Computer Science, FEI, VŠB - Technical University of Ostrava, 17.

More information

Boolean and Vector Space Retrieval Models CS 290N Some of slides from R. Mooney (UTexas), J. Ghosh (UT ECE), D. Lee (USTHK).

Boolean and Vector Space Retrieval Models CS 290N Some of slides from R. Mooney (UTexas), J. Ghosh (UT ECE), D. Lee (USTHK). Boolean and Vector Space Retrieval Models 2013 CS 290N Some of slides from R. Mooney (UTexas), J. Ghosh (UT ECE), D. Lee (USTHK). 1 Table of Content Boolean model Statistical vector space model Retrieval

More information

A randomized algorithm for approximating the SVD of a matrix

A randomized algorithm for approximating the SVD of a matrix A randomized algorithm for approximating the SVD of a matrix Joint work with Per-Gunnar Martinsson (U. of Colorado) and Vladimir Rokhlin (Yale) Mark Tygert Program in Applied Mathematics Yale University

More information

Regularized Locality Preserving Indexing via Spectral Regression

Regularized Locality Preserving Indexing via Spectral Regression Regularized Locality Preserving Indexing via Spectral Regression Deng Cai, Xiaofei He, Wei Vivian Zhang, Jiawei Han University of Illinois at Urbana-Champaign Yahoo! {dengcai2, hanj}@cs.uiuc.edu, {hex,

More information

An Empirical Study on Dimensionality Optimization in Text Mining for Linguistic Knowledge Acquisition

An Empirical Study on Dimensionality Optimization in Text Mining for Linguistic Knowledge Acquisition An Empirical Study on Dimensionality Optimization in Text Mining for Linguistic Knowledge Acquisition Yu-Seop Kim 1, Jeong-Ho Chang 2, and Byoung-Tak Zhang 2 1 Division of Information and Telecommunication

More information

Folding-up: A Hybrid Method for Updating the Partial Singular Value Decomposition in Latent Semantic Indexing

Folding-up: A Hybrid Method for Updating the Partial Singular Value Decomposition in Latent Semantic Indexing Folding-up: A Hybrid Method for Updating the Partial Singular Value Decomposition in Latent Semantic Indexing by Jane Elizabeth Bailey Tougas Submitted in partial fulfillment of the requirements for the

More information

USING SINGULAR VALUE DECOMPOSITION (SVD) AS A SOLUTION FOR SEARCH RESULT CLUSTERING

USING SINGULAR VALUE DECOMPOSITION (SVD) AS A SOLUTION FOR SEARCH RESULT CLUSTERING POZNAN UNIVE RSIY OF E CHNOLOGY ACADE MIC JOURNALS No. 80 Electrical Engineering 2014 Hussam D. ABDULLA* Abdella S. ABDELRAHMAN* Vaclav SNASEL* USING SINGULAR VALUE DECOMPOSIION (SVD) AS A SOLUION FOR

More information

Embeddings Learned By Matrix Factorization

Embeddings Learned By Matrix Factorization Embeddings Learned By Matrix Factorization Benjamin Roth; Folien von Hinrich Schütze Center for Information and Language Processing, LMU Munich Overview WordSpace limitations LinAlgebra review Input matrix

More information

Matrix Decomposition and Latent Semantic Indexing (LSI) Introduction to Information Retrieval INF 141/ CS 121 Donald J. Patterson

Matrix Decomposition and Latent Semantic Indexing (LSI) Introduction to Information Retrieval INF 141/ CS 121 Donald J. Patterson Matrix Decomposition and Latent Semantic Indexing (LSI) Introduction to Information Retrieval INF 141/ CS 121 Donald J. Patterson Latent Semantic Indexing Outline Introduction Linear Algebra Refresher

More information

LSI improves retrieval by improving the fit between a linear language model and non-linear data

LSI improves retrieval by improving the fit between a linear language model and non-linear data Miles J. Efron. The Dynamics of Dimensionality Reduction for Information Retrieval: A Study of Latent Semantic Indexing Using Simulated Data. A Master s Paper for the M.S. in I.S. degree. April, 2000,

More information

INFO 4300 / CS4300 Information Retrieval. IR 9: Linear Algebra Review

INFO 4300 / CS4300 Information Retrieval. IR 9: Linear Algebra Review INFO 4300 / CS4300 Information Retrieval IR 9: Linear Algebra Review Paul Ginsparg Cornell University, Ithaca, NY 24 Sep 2009 1/ 23 Overview 1 Recap 2 Matrix basics 3 Matrix Decompositions 4 Discussion

More information

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2016 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276, Stanford)

More information

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice 3. Eigenvalues and Eigenvectors, Spectral Representation 3.. Eigenvalues and Eigenvectors A vector ' is eigenvector of a matrix K, if K' is parallel to ' and ' 6, i.e., K' k' k is the eigenvalue. If is

More information

Linear Algebra and Eigenproblems

Linear Algebra and Eigenproblems Appendix A A Linear Algebra and Eigenproblems A working knowledge of linear algebra is key to understanding many of the issues raised in this work. In particular, many of the discussions of the details

More information

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2017 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276, Stanford)

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

Semantic Similarity from Corpora - Latent Semantic Analysis

Semantic Similarity from Corpora - Latent Semantic Analysis Semantic Similarity from Corpora - Latent Semantic Analysis Carlo Strapparava FBK-Irst Istituto per la ricerca scientifica e tecnologica I-385 Povo, Trento, ITALY strappa@fbk.eu Overview Latent Semantic

More information

Motivation. User. Retrieval Model Result: Query. Document Collection. Information Need. Information Retrieval / Chapter 3: Retrieval Models

Motivation. User. Retrieval Model Result: Query. Document Collection. Information Need. Information Retrieval / Chapter 3: Retrieval Models 3. Retrieval Models Motivation Information Need User Retrieval Model Result: Query 1. 2. 3. Document Collection 2 Agenda 3.1 Boolean Retrieval 3.2 Vector Space Model 3.3 Probabilistic IR 3.4 Statistical

More information

How Latent Semantic Indexing Solves the Pachyderm Problem

How Latent Semantic Indexing Solves the Pachyderm Problem How Latent Semantic Indexing Solves the Pachyderm Problem Michael A. Covington Institute for Artificial Intelligence The University of Georgia 2011 1 Introduction Here I present a brief mathematical demonstration

More information

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2014 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276, Stanford)

More information

SVD Subspace Projections for Term Suggestion Ranking and Clustering

SVD Subspace Projections for Term Suggestion Ranking and Clustering SVD Subspace Projections for Term Suggestion Ranking and Clustering David Gleich Harvey Mudd College Claremont,CA dgleich@cs.hmc.edu Leonid Zhukov Yahoo! Research Labs Pasadena, CA leonid.zhukov@overture.com

More information

vector space retrieval many slides courtesy James Amherst

vector space retrieval many slides courtesy James Amherst vector space retrieval many slides courtesy James Allan@umass Amherst 1 what is a retrieval model? Model is an idealization or abstraction of an actual process Mathematical models are used to study the

More information

MATRIX DECOMPOSITION AND LATENT SEMANTIC INDEXING (LSI) Introduction to Information Retrieval CS 150 Donald J. Patterson

MATRIX DECOMPOSITION AND LATENT SEMANTIC INDEXING (LSI) Introduction to Information Retrieval CS 150 Donald J. Patterson MATRIX DECOMPOSITION AND LATENT SEMANTIC INDEXING (LSI) Introduction to Information Retrieval CS 150 Donald J. Patterson Content adapted from Hinrich Schütze http://www.informationretrieval.org Latent

More information

Information Retrieval Basic IR models. Luca Bondi

Information Retrieval Basic IR models. Luca Bondi Basic IR models Luca Bondi Previously on IR 2 d j q i IRM SC q i, d j IRM D, Q, R q i, d j d j = w 1,j, w 2,j,, w M,j T w i,j = 0 if term t i does not appear in document d j w i,j and w i:1,j assumed to

More information

Assignment 3. Latent Semantic Indexing

Assignment 3. Latent Semantic Indexing Assignment 3 Gagan Bansal 2003CS10162 Group 2 Pawan Jain 2003CS10177 Group 1 Latent Semantic Indexing OVERVIEW LATENT SEMANTIC INDEXING (LSI) considers documents that have many words in common to be semantically

More information

ANLP Lecture 22 Lexical Semantics with Dense Vectors

ANLP Lecture 22 Lexical Semantics with Dense Vectors ANLP Lecture 22 Lexical Semantics with Dense Vectors Henry S. Thompson Based on slides by Jurafsky & Martin, some via Dorota Glowacka 5 November 2018 Henry S. Thompson ANLP Lecture 22 5 November 2018 Previous

More information

Sparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation.

Sparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation. ANLP Lecture 22 Lexical Semantics with Dense Vectors Henry S. Thompson Based on slides by Jurafsky & Martin, some via Dorota Glowacka 5 November 2018 Previous lectures: Sparse vectors recap How to represent

More information

Introduction to Information Retrieval (Manning, Raghavan, Schutze) Chapter 6 Scoring term weighting and the vector space model

Introduction to Information Retrieval (Manning, Raghavan, Schutze) Chapter 6 Scoring term weighting and the vector space model Introduction to Information Retrieval (Manning, Raghavan, Schutze) Chapter 6 Scoring term weighting and the vector space model Ranked retrieval Thus far, our queries have all been Boolean. Documents either

More information

Information Retrieval

Information Retrieval Introduction to Information Retrieval CS276: Information Retrieval and Web Search Christopher Manning and Prabhakar Raghavan Lecture 6: Scoring, Term Weighting and the Vector Space Model This lecture;

More information

Latent Semantic Indexing

Latent Semantic Indexing Latent Semantic Indexing 1 Latent Semantic Indexing Term-document matrices are very large, though most cells are zeros But the number of topics that people talk about is small (in some sense) Clothes,

More information

1 Feature Vectors and Time Series

1 Feature Vectors and Time Series PCA, SVD, LSI, and Kernel PCA 1 Feature Vectors and Time Series We now consider a sample x 1,..., x of objects (not necessarily vectors) and a feature map Φ such that for any object x we have that Φ(x)

More information

Text Mining for Economics and Finance Unsupervised Learning

Text Mining for Economics and Finance Unsupervised Learning Text Mining for Economics and Finance Unsupervised Learning Stephen Hansen Text Mining Lecture 3 1 / 46 Introduction There are two main divisions in machine learning: 1. Supervised learning seeks to build

More information

Matrices, Vector Spaces, and Information Retrieval

Matrices, Vector Spaces, and Information Retrieval Matrices, Vector Spaces, and Information Authors: M. W. Berry and Z. Drmac and E. R. Jessup SIAM 1999: Society for Industrial and Applied Mathematics Speaker: Mattia Parigiani 1 Introduction Large volumes

More information

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued)

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued) 1 A linear system of equations of the form Sections 75, 78 & 81 a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2 a m1 x 1 + a m2 x 2 + + a mn x n = b m can be written in matrix

More information

Text Analytics (Text Mining)

Text Analytics (Text Mining) http://poloclub.gatech.edu/cse6242 CSE6242 / CX4242: Data & Visual Analytics Text Analytics (Text Mining) Concepts, Algorithms, LSI/SVD Duen Horng (Polo) Chau Assistant Professor Associate Director, MS

More information

573 Schedule. Crosswords, Games, Visualization. Logistics. IR Models. Measuring Performance. Precision-recall curves CSE 573

573 Schedule. Crosswords, Games, Visualization. Logistics. IR Models. Measuring Performance. Precision-recall curves CSE 573 Crosswords, Games, Visualization CSE 573 Artificial Life 573 Schedule Supervised Learning Crossword Puzzles Intelligent Internet Systems Reinforcement Learning Planning Logic-Based Probabilistic Knowledge

More information

Integrating Logical Operators in Query Expansion in Vector Space Model

Integrating Logical Operators in Query Expansion in Vector Space Model Integrating Logical Operators in Query Expansion in Vector Space Model Jian-Yun Nie, Fuman Jin DIRO, Université de Montréal C.P. 6128, succursale Centre-ville, Montreal Quebec, H3C 3J7 Canada {nie, jinf}@iro.umontreal.ca

More information

Chap 2: Classical models for information retrieval

Chap 2: Classical models for information retrieval Chap 2: Classical models for information retrieval Jean-Pierre Chevallet & Philippe Mulhem LIG-MRIM Sept 2016 Jean-Pierre Chevallet & Philippe Mulhem Models of IR 1 / 81 Outline Basic IR Models 1 Basic

More information

Hyperlinked-Induced Topic Search (HITS) identifies. authorities as good content sources (~high indegree) HITS [Kleinberg 99] considers a web page

Hyperlinked-Induced Topic Search (HITS) identifies. authorities as good content sources (~high indegree) HITS [Kleinberg 99] considers a web page IV.3 HITS Hyperlinked-Induced Topic Search (HITS) identifies authorities as good content sources (~high indegree) hubs as good link sources (~high outdegree) HITS [Kleinberg 99] considers a web page a

More information

Information retrieval LSI, plsi and LDA. Jian-Yun Nie

Information retrieval LSI, plsi and LDA. Jian-Yun Nie Information retrieval LSI, plsi and LDA Jian-Yun Nie Basics: Eigenvector, Eigenvalue Ref: http://en.wikipedia.org/wiki/eigenvector For a square matrix A: Ax = λx where x is a vector (eigenvector), and

More information