ChengXiang ( Cheng ) Zhai Department of Computer Science University of Illinois at Urbana-Champaign

Similar documents
Lower-Bounding Term Frequency Normalization

CS630 Representing and Accessing Digital Information Lecture 6: Feb 14, 2006

Lecture 13: More uses of Language Models

Semantic Term Matching in Axiomatic Approaches to Information Retrieval

Midterm Examination Practice

A Study of the Dirichlet Priors for Term Frequency Normalisation

University of Illinois at Urbana-Champaign. Midterm Examination

A Risk Minimization Framework for Information Retrieval

Language Models. Hongning Wang

Is Document Frequency important for PRF?

Deposited on: 1 November 2007 Glasgow eprints Service

Language Models, Smoothing, and IDF Weighting

PV211: Introduction to Information Retrieval

Ranked Retrieval (2)

A Study of Smoothing Methods for Language Models Applied to Information Retrieval

Comparing Relevance Feedback Techniques on German News Articles

RETRIEVAL MODELS. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

Risk Minimization and Language Modeling in Text Retrieval

Generalized Inverse Document Frequency

Latent Semantic Analysis. Hongning Wang

Information Retrieval Models

Information Retrieval

Language Models. Web Search. LM Jelinek-Mercer Smoothing and LM Dirichlet Smoothing. Slides based on the books: 13

Fall CS646: Information Retrieval. Lecture 6 Boolean Search and Vector Space Model. Jiepu Jiang University of Massachusetts Amherst 2016/09/26

Chap 2: Classical models for information retrieval

Lecture 9: Probabilistic IR The Binary Independence Model and Okapi BM25

MRIM at ImageCLEF2012. From Words to Concepts: A New Counting Approach

Language Models. CS6200: Information Retrieval. Slides by: Jesse Anderton


Information Retrieval

11. Learning To Rank. Most slides were adapted from Stanford CS 276 course.

Query Performance Prediction: Evaluation Contrasted with Effectiveness

Query Propagation in Possibilistic Information Retrieval Networks

Modeling the Score Distributions of Relevant and Non-relevant Documents

Information Retrieval

Modern Information Retrieval

CS276A Text Information Retrieval, Mining, and Exploitation. Lecture 4 15 Oct 2002

Collaborative topic models: motivations cont

A Note on the Expectation-Maximization (EM) Algorithm

Why Language Models and Inverse Document Frequency for Information Retrieval?

CS 646 (Fall 2016) Homework 3

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology

On the Foundations of Diverse Information Retrieval. Scott Sanner, Kar Wai Lim, Shengbo Guo, Thore Graepel, Sarvnaz Karimi, Sadegh Kharazmi

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology

Variable Latent Semantic Indexing

Language Models and Smoothing Methods for Collections with Large Variation in Document Length. 2 Models

A risk minimization framework for information retrieval

1 Information retrieval fundamentals

Boolean and Vector Space Retrieval Models

Probabilistic Information Retrieval

Information Retrieval and Web Search Engines

Lecture 3: Probabilistic Retrieval Models

Positional Language Models for Information Retrieval

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

Score Distribution Models

A Note on the Effect of Term Weighting on Selecting Intrinsic Dimensionality of Data

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Intro to Learning Theory Date: 12/8/16

perplexity = 2 cross-entropy (5.1) cross-entropy = 1 N log 2 likelihood (5.2) likelihood = P(w 1 w N ) (5.3)

A Neural Passage Model for Ad-hoc Document Retrieval

Information Retrieval and Organisation

CS 277: Data Mining. Mining Web Link Structure. CS 277: Data Mining Lectures Analyzing Web Link Structure Padhraic Smyth, UC Irvine

An Algorithms-based Intro to Machine Learning

Predicting Neighbor Goodness in Collaborative Filtering

The Axiometrics Project

Chapter 4: Advanced IR Models

Notes on Latent Semantic Analysis

Term Weighting and the Vector Space Model. borrowing from: Pandu Nayak and Prabhakar Raghavan

Vector Space Model. Yufei Tao KAIST. March 5, Y. Tao, March 5, 2013 Vector Space Model

Matrix Factorization & Latent Semantic Analysis Review. Yize Li, Lanbo Zhang

Comparative Document Analysis for Large Text Corpora

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

Some Formal Analysis of Rocchio s Similarity-Based Relevance Feedback Algorithm

INFO 630 / CS 674 Lecture Notes

5/21/17. Machine learning for IR ranking? Machine learning for IR ranking. Machine learning for IR ranking. Introduction to Information Retrieval

Some Simple Effective Approximations to the 2 Poisson Model for Probabilistic Weighted Retrieval

PV211: Introduction to Information Retrieval

IR Models: The Probabilistic Model. Lecture 8

COMS 4721: Machine Learning for Data Science Lecture 18, 4/4/2017

Lecture 3: Pivoted Document Length Normalization

Term Filtering with Bounded Error

Models of collective inference

CS 188: Artificial Intelligence Spring Today

Retrieval by Content. Part 2: Text Retrieval Term Frequency and Inverse Document Frequency. Srihari: CSE 626 1

Ranking-II. Temporal Representation and Retrieval Models. Temporal Information Retrieval

Multi-theme Sentiment Analysis using Quantified Contextual

Term Weighting and Vector Space Model. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze

CAIM: Cerca i Anàlisi d Informació Massiva

Pivoted Length Normalization I. Summary idf II. Review

INFO 4300 / CS4300 Information Retrieval. slides adapted from Hinrich Schütze s, linked from

N-gram Language Modeling

Collaborative Topic Modeling for Recommending Scientific Articles

Yahoo! Labs Nov. 1 st, Liangjie Hong, Ph.D. Candidate Dept. of Computer Science and Engineering Lehigh University

Leverage Sparse Information in Predictive Modeling

Effectiveness of complex index terms in information retrieval

Information Retrieval

Latent Dirichlet Allocation Introduction/Overview

Compact Indexes for Flexible Top-k Retrieval

Ordinary Least Squares Linear Regression

arxiv: v1 [cs.ir] 25 Oct 2015

Transcription:

Axiomatic Analysis and Optimization of Information Retrieval Models ChengXiang ( Cheng ) Zhai Department of Computer Science University of Illinois at Urbana-Champaign http://www.cs.uiuc.edu/homes/czhai 1

Search is everywhere, and part of everyone s life Web Search Desk Search Enterprise Search Site Search Social Media Search 2

Search accuracy matters! # Queries /Day 4,700,000,000 X 1 sec ~1,300,000 hrs X 10 sec ~13,000,000 hrs 1,600,000,000 ~440,000 hrs ~4,400,000 hrs 2,000,000 ~550 hrs ~5,500 hrs How can we improve all search engines in a general way? Sources: Google, Twitter: http://www.statisticbrain.com/ PubMed: http://www.ncbi.nlm.nih.gov/about/tools/restable_stat_pubmed.html 3

Behind all the search boxes number of queries k search engines Query q Document collection d Ranked list Machine Learning How can we optimize a retrieval model? Score(q,d) Retrieval Model Natural Language Processing 4

Retrieval model = computational definition of relevance S( computer science CMU, ) s( computer, ) s( science, ) s( CMU, ) How many times does computer occur in d? Term Frequency (TF): c( computer, d) How long is d? Document length: d How often do we see computer in the entire collection? Document Frequency: df( computer) P( computer Collection) 5

This image cannot currently be displayed. Scoring based on bag of words in general Sum over matched query terms q w d s( q, d ) = f weight( w, q, d), a( q, d w q d ) g[ c( w, q), c( w, d ), d, df ( w)] Term Frequency (TF) Document length p( w C) Inverse Document Frequency (IDF) 6

Improving retrieval models is a long-standing challenge Vector Space Models: [Salton et al. 1975], [Singhal et al. 1996], Classic Probabilistic Models: [Maron & Kuhn 1960], [Harter 1975], [Robertson & Sparck Jones 1976], [van Rijsbergen 1977], [Robertson 1977], [Robertson et al. 1981], [Robertson & Walker 1994], Language Models: [Ponte & Croft 1998], [Hiemstra & Kraaij 1998], [Zhai & Lafferty 2001], [Lavrenko & Croft 2001], [Kurland & Lee 2004], Non-Classic Logic Models: [Rijsbergen 1986], [Wong & Yao 1991], Divergence from Randomness: [Amati & Rijsbergen 2002], [He & Ounis 2005], Learning to Rank: [Fuhr 1989], [Gey 1994],... Many different models were proposed and tested 7

Some are working very well (equally well) Pivoted length normalization (PIV) [Singhal et al. 96] BM25 [Robertson & Walker 94] PL2 [Amati & Rijsbergen 02] Query likelihood with Dirichlet prior (DIR) [Ponte & Croft 98], [Zhai & Lafferty] but many others failed to work well 8

State of the art retrieval models Pivoted Normalization Method w q d 1+ ln(1 + ln( cwd (, ))) N+ 1 cwq (, ) ln d (1 s) + s df ( w) avdl Dirichlet Prior Method w q d cwd (, ) µ cwq (, ) ln(1 + ) + q ln µ pwc ( ) µ + d Okapi Method w q d ln N df ( w) + 0.5 ( k1 + 1) cwd (, ) ( k3 + 1) cwq (, ) df ( w) + 0.5 d k 3 (, ) 1((1 b) + b ) + cwd (, ) k + c w q avdl PL2 is a bit more complicated, but implements similar heuristics 9

Questions Why do {BM25, PIV, PL, DIR, } tend to perform similarly even though they were derived in very different ways? Why are they better than many other variants? Why does it seem to be hard to beat these strong baseline methods? Are they hitting the ceiling of bag-of-words assumption? If yes, how can we prove it? If not, how can we find a more effective one? 10

Suggested Answers Why do {BM25, PIV, PL, DIR, } tend to perform similarly even though they were derived in very different ways? They share some nice common properties These properties are more important than how each is derived Why are they better than many other variants? Other variants don t have all the nice properties Why does it seem to be hard to beat these strong baseline methods? We don t have a good knowledge about their deficiencies Are they hitting the ceiling of bag-of-words assumption? If yes, how can we prove it? If not, how can we find a more effective one? Need to formally define the ceiling (= complete set of nice properties ) 11

Main Point of the Talk: Axiomatic Relevance Hypothesis (ARH) Relevance can be modeled by a set of formally defined constraints on a retrieval function If a function satisfies all the constraints, it will perform well empirically If function Fa satisfies more constraints than function Fb, Fa would perform bettter than Fb empirically Analytical evaluation of retrieval functions Given a set of relevance constraints C={c1,, ck} Function Fa is analytically more effective than function Fb iff the set of constraints satisfied by Fb is a proper subset of those satisfied by Fa A function F is optimal iff it satisfies all the constraints in C 12

Rest of the Talk 1. Modeling relevance with formal constraints 2. Testing the axiomatic relevance hypothesis 3. An axiomatic framework for optimizing retrieval models 4. Open challenge: seeking an ultimately optimal retrieval model 13

Outline 1. Modeling relevance with formal constraints 2. Testing the axiomatic relevance hypothesis 3. An axiomatic framework for optimizing retrieval models 4. Open challenge: seeking an ultimately optimal retrieval model 14

Motivation: different models, but similar Pivoted Normalization Method w q d 1+ ln(1 + ln( cwd (, ))) N+ 1 cwq (, ) ln d (1 s) + s df ( w) avdl Document Inversed Length Term Document Parameter Frequency Normalization Frequency sensitivity Dirichlet Prior Method w q d cwd (, ) µ cwq (, ) ln(1 + ) + q ln µ pwc ( ) µ + d Okapi Method heuristics w q d ln N df ( w) + 0.5 ( k1 + 1) cwd (, ) ( k3 + 1) cwq (, ) df ( w) + 0.5 d k 3 (, ) 1((1 b) + b ) + cwd (, ) k + c w q avdl PL2 is a bit more complicated, but implements similar heuristics 15

Are they performing well because they implement similar retrieval heuristics? Can we formally capture these necessary retrieval heuristics? 16

Term Frequency Constraints (TFC1) TF weighting heuristic I: Give a higher score to a document with more occurrences of a query term. TFC1 Let q be a query with only one term w. If d 1 = d2 and c w, d ) > c( w, ( 1 d2 then f ( d1, q) > f ( d2, q). ) q : d 1 : d 2 : w c( w, d1) f ( d1, q) > f ( d2, q) c( w, d2) 17

Term Frequency Constraints (TFC2) TF weighting heuristic II: Favor a document with more distinct query terms. TFC2 Let q be a query and w 1, w 2 be two query terms. Assume idf ( w1) = idf ( w2) and If and cw (, d) = cw (, d) + cw (, d) 1 2 1 1 2 1 d = d 1 2 cw (, d) = 0, cw (, d) 0, cw (, d) 0 2 2 1 1 2 1 q: d 1 : d 2 : w 1 w 2 cw ( 1, d1) cw ( 2, d1) then f( d, q) > f( d, q). 1 2 cw ( 1, d2) f( d, q) > f( d, q) 1 2 18

Length Normalization Constraints(LNCs) Document length normalization heuristic: Penalize long documents(lnc1); avoid over-penalizing long documents (LNC2). LNC1 Let q be a query. If for some word w q, c( w, d ) = c( w, d1) w c( w, d ) = c( w, ) 2 +, 2 d1 ( d1, q) f ( d2, q) but for other words then f 1 q: d 1 : d 2 : f ( 2 d1, q) f ( d, q) c( w, d1) c( w, d2) w q LNC2 Let q be a query. If k >, d = k and c( w, d1) = k c( w, d2) then f 1 1 d2 ( 2 d1, q) f ( d, q) q: d 1 : d 2 : f ( 2 d1, q) f ( d, q) 19

TF-LENGTH Constraint (TF-LNC) TF-LN heuristic: Regularize the interaction of TF and document length. TF-LNC Let q be a query with only one term w. q: w c( w, d1) d = d + cwd (, ) cwd (, ) If 1 2 1 2 d 1 : and c ( w, d1) > c( w, d2) d 2 : then f ( d1, q) > f ( d2, q). c( w, d2) f( d, q) > f( d, q) 1 2 20

Seven Basic Relevance Constraints [Fang et al. 2011] Hui Fang, Tao Tao, ChengXiang Zhai: Diagnostic Evaluation of Information Retrieval Models. ACM Trans. Inf. Syst. 29(2): 7 (2011) 21

Outline 1. Modeling relevance with formal constraints 2. Testing the axiomatic relevance hypothesis 3. An axiomatic framework for optimizing retrieval models 4. Open challenge: seeking an ultimately optimal retrieval model 22

Axiomatic Relevance Hypothesis (ARH) Relevance can be modeled by a set of formally defined constraints on a retrieval function If a function satisfies all the constraints, it will perform well empirically If function Fa satisfies more constraints than function Fb, Fa would perform bettter than Fb empirically Analytical evaluation of retrieval functions Given a set of relevance constraints C={c1,, ck} Function Fa is analytically more effective than function Fb iff the set of constraints satisfied by Fb is a proper subset of those satisfied by Fa A function F is optimal iff it satisfies all the constraints in C 23

Testing the Axiomatic Relevance Hypothesis Is the satisfaction of these constraints correlated with good empirical performance of a retrieval function? Can we use these constraints to analytically compare retrieval functions without experimentation? Yes! to both questions Constraint analysis reveals optimal ranges of parameter values When a formula does not satisfy the constraint, it often indicates non-optimality of the formula. Violation of constraints may pinpoint where a formula needs to be improved. 24

Bounding Parameters Pivoted Normalization Method LNC2 s<0.4 Optimal s (for average precision) Parameter sensitivity of s Avg. Prec. 0.4 s 25

Okapi Method w q d Analytical Comparison N df ( w) + 0.5 ( k1 + 1) cwd (, ) ( k3 + 1) cwq (, ) ln df ( w) + 0.5 d k 3 (, ) 1((1 b) + b ) + cwd (, ) k + c w q avdl Negative when df(w) is large Violate many constraints keyword query verbose query Avg. Prec Okapi Avg. Prec Pivoted s or b s or b 26

Fixing a deficiency in BM25 improves the effectiveness Modified Okapi Method w q d N + 1 ln df N df ( w) + 0.5 ( k1 + 1) cwd (, ) ( k3 + 1) cwq (, ) ln df ( w) + 0.5 d k 3 (, ) 1((1 b) + b ) + cwd (, ) k + c w q avdl Make Okapi satisfy more constraints; expected to help verbose queries keyword query verbose query Modified Okapi Avg. Prec. Okapi Avg. Prec. Pivoted s or b s or b 27

Systematic Analysis of 4 State of the Art Models [Fang et al. 11] Negative IDF Parameter s must be small Problematic with common terms; parameter c must be large Problematic when a query term occurs less frequently in a doc than expected Question: why are Dirichlet and PL2 still competitive despite their inherent problems that can t be fixed through parameter tuning? 28

Outline 1. Modeling relevance with formal constraints 2. Testing the axiomatic relevance hypothesis 3. An axiomatic framework for optimizing retrieval models 4. Open challenge: seeking an ultimately optimal retrieval model 29

How can we leverage constraints to find an optimal retrieval model? 30

Basic Idea of the Axiomatic Framework (Optimization Problem Setup) Our target Function space space C 2 S 2 S 2 C 3 S 3 S 3 S 1 S 1 C 1 Retrieval constraints 31

Three Questions How do we define the constraints? We ve talked about that; more later How do we define the function space? One possibility: leverage existing state of the art functions How do we search in the function space? One possibility: search in the neighborhood of existing state of the art functions 32

Inductive Definition of Function Space S : Q D Q = q 1,q 2,...,q m ; D = d 1,d 2,...,d n Q: D: cat dog big big Define the function space inductively Primitive weighting function (f) S(Q,D) = S(, ) = f (, ) Query growth function (h) S(Q,D) = S(, ) = S(, )+h(,, ) Document growth function (g) S(Q,D) = S(, ) = S(, )+g(,, ) 33

Derivation of New Retrieval Functions decompose S(Q,D) existing function generalize constrain f g h F G H C 1 S S C 3 f ' g' h' C 2 assemble S'(Q,D) new function 34

A Sample Derived Function based on BM25 [Fang & Zhai 05] QTF IDF TF S(Q,D) = N c(t,q) ( df (t) )0.35 t Q D c(t,d) c(t,d) + s + s D avdl length normalization 35

The derived function is less sensitive to the parameter setting better Axiomatic Model 36

Inevitability of heuristic thinking and necessity of axiomatic analysis The theory-effectiveness gap Theoretically motivated models don t automatically perform well empirically Heuristic adjustment seems always necessary Cause: inaccurate modeling of relevance How can we bridge the gap? The answer lies in axiomatic analysis Use constraints to help identify the error in modeling relevance, thus obtaining insights about how to improve a model 37

Systematic Analysis of 4 State of the Art Models [Fang et al. 11] Negative IDF Parameter s must be small Modified BM25 satisfies all the constraints! Problematic with common terms; parameter c must be large Without knowing its deficiency, we can t easily propose a new model working better than BM25 Problematic when a query term occurs less frequently in a doc than expected 38

A Recent Success of Axiomatic Analysis: Lower Bounding TF Normalization [Lv & Zhai 11] Existing retrieval functions lack a lower bound for normalized TF with document length Long documents overly penalized A very long document matching two query terms can have a lower score than a short document matching only one query term Proposed two constraints for lower bounding TF Proposed a general solution to fix the problem that worked for BM25, PL2, Dirichlet, and Piv, leading to improved versions of them (BM25+, PL2+, Dir+, Piv+) 39

New Constraints: LB1 & LB2 LB1: Let Q be a query. Assume D1 and D2 are two documents such that S(Q,D1) = S(Q,D2). If we reformulate the query by adding another term q Q into Q, where c(q,d1) = 0 and c(q,d2) > 0, then S(Q {q},d1) < S(Q {q},d2). The presence absence gap (0-1 gap) shouldn t be closed due to length normalization LB2: Let Q = {q1, q2} be a query with two terms q1 and q2. Assume td(q1) = td(q2), where td(t) can be any reasonable measure of term discrimination value. If D1 and D2 are two documents such that c(q2,d1) = c(q2,d2) = 0, c(q1,d1) > 0, c(q1,d2) > 0, and S(Q,D1) = S(Q,D2), then S(Q,D1 {q1} {t1}) < S(Q,D2 {q2} {t2}), for all t1 and t2 such that t1 D1, t2 D2, t1 Q and t2 Q. Repeated occurrence of an already matched query term isn t as important as the first occurrence of an otherwise absent query term 40

It turns out none of BM25, PL2, Dirichlet, PIV satisfies both constraints A general heuristic solution: add a small constant lower bound Worked well for improving all the four models 41

BM25+ Improves over BM25 For details, see Yuanhua Lv, ChengXiang Zhai, Lower Bounding Term Frequency Normalization, Proceedings of the 20 th ACM International Conference on Information and Knowledge Management (CIKM'11), page 7-16, 2011. 42

More examples of theory-effectiveness gap and the need for axiomatic analysis The derivation of the query likelihood retrieval function relies on 3 assumptions: (1) query likelihood scoring; (2) independency of query terms; (3) collection LM for smoothing; however, it can t explain why some apparently reasonable smoothing methods perform poorly In statistical translation model for retrieval [Berger & Lafferty 99], we must ensure sufficient self translation probability to avoid unreasonable retrieval results, but such a constraint can t be explained by estimation of translation model No explanation why other divergence-based similarity function doesn t work well as the asymmetric KL-divergence function D(Q D) 43

Outline 1. Modeling relevance with formal constraints 2. Testing the axiomatic relevance hypothesis 3. An axiomatic framework for optimizing retrieval models 4. Open challenge: seeking an ultimately optimal retrieval model 44

Open Challenges Does there exist a complete set of constraints? If yes, how can we define them? If no, how can we prove it? How do we evaluate the constraints? How do we evaluate a constraint? (e.g., should the score contribution of a term be bounded? In BM25, it is.) How do we evaluate a set of constraints? How do we define the function space? Search in the neighborhood of an existing function? Search in a new function space? 45

Open Challenges How do we check a function w.r.t. a constraint? How can we quantify the degree of satisfaction? How can we put constraints in a machine learning framework? Something like maximum entropy? How can we go beyond bag of words? Model pseudo feedback? Cross-lingual IR? Conditional constraints on specific type of queries? Specific type of documents? 46

Possible Future Scenario 1: Impossibility Theorems for IR We will find inconsistency among constraints Will be able to prove impossibility theorems for IR Similar to Kleinberg s impossibility theorem for clustering J. Kleinberg. An Impossibility Theorem for Clustering. Advances in Neural Information Processing Systems (NIPS) 15, 2002 47

Future Scenario 2: Sufficiently Restrictive Constraints We will be able to propose a comprehensive set of constraints that are sufficient for deriving a unique (optimal) retrieval function Similar to the derivation of the entropy function 48

Future Scenario 3 (most likely): Open Set of Insufficient Constraints We will have a large set of constraints without conflict, but insufficient for ensuring good retrieval performance Room for new constraints, but we ll never be sure what they are We need to combine axiomatic analysis with a constructive retrieval functional space and supervised machine learning 49

Summary: Axiomatic Relevance Hypothesis Formal retrieval function constraints for modeling relevance Axiomatic analysis as a way to assess optimality of retrieval models Inevitability of heuristic thinking in developing retrieval models for bridging the theory-effectiveness gap Possibility of leveraging axiomatic analysis to improve the state of the art models Axiomatic Framework = constraints + constructive function space based on existing or new models and theories 50

Updated Answers Why do {BM25, PIV, PL, DIR, } tend to perform similarly even though they were derived in very different ways? Relevance more accurately modeled with constraints They share some nice common properties These properties are more important than how each is derived Why are they better than many other variants? Other variants don t have all the nice properties Why does it seem to be hard to beat these strong baseline methods? We didn t find a constraint that they fail to satisfy We don t have a good knowledge about their deficiencies Are they hitting the ceiling of bag-of-words assumption? If yes, how can we prove it? If not, how can we find a more effective one? No, they have NOT hit the ceiling yet! Need to formally define the ceiling (= complete set of nice properties ) 51

Future Work: Putting All Together! Axiomatic Analysis Optimal Model We are here This strategy seems to have worked well in the past Vector Space Model? Optimal Model Cranfield You are 2013.0322 miles from destination 52

Acknowledgments Collaborators: Hui Fang, Yuanhua Lv, Tao Tao, Maryam Karimzadehgan, and others Funding 53

Thank You! Questions/Comments? 54

Overview of all Research Projects Web, Email, and Biomedical informatics Search Applications Information Access Summarization Filtering Search Information Organization Visualization Mining Extraction Mining Applications Knowledge Acquisition Categorization Clustering - Personalized - Retrieval models - Topic map - Recommender Natural Language Content Analysis Text Entity/Relation Extraction -Contextual text mining -Opinion integration & summarization - Information quality More information can be found at http://timan.cs.uiuc.edu/