GeoTime Retrieval through Passage-based Learning to Rank

Similar documents
A Study of the Dirichlet Priors for Term Frequency Normalisation

Effectiveness of complex index terms in information retrieval

Ranked Retrieval (2)

Latent Dirichlet Allocation Based Multi-Document Summarization

Latent Semantic Tensor Indexing for Community-based Question Answering

Variable Latent Semantic Indexing

IMU Experiment in IR4QA at NTCIR-8

Ranked IR. Lecture Objectives. Text Technologies for Data Science INFR Learn about Ranked IR. Implement: 10/10/2017. Instructor: Walid Magdy

Citation for published version (APA): Andogah, G. (2010). Geographically constrained information retrieval Groningen: s.n.

Probabilistic Field Mapping for Product Search

TALP at GeoQuery 2007: Linguistic and Geographical Analysis for Query Parsing

Prediction of Citations for Academic Papers

arxiv: v1 [cs.ir] 22 Dec 2012

Learning Topical Transition Probabilities in Click Through Data with Regression Models

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology

Outline for today. Information Retrieval. Cosine similarity between query and document. tf-idf weighting

DM-Group Meeting. Subhodip Biswas 10/16/2014

A Neural Passage Model for Ad-hoc Document Retrieval

Hub, Authority and Relevance Scores in Multi-Relational Data for Query Search

Sparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation.

ANLP Lecture 22 Lexical Semantics with Dense Vectors

Natural Language Processing. Topics in Information Retrieval. Updated 5/10

Comparing Relevance Feedback Techniques on German News Articles

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology

A Hierarchical, Multi-modal Approach for Placing Videos on the Map using Millions of Flickr Photographs

St. Kitts and Nevis Heritage and Culture

Chap 2: Classical models for information retrieval

Generalized Inverse Document Frequency

Scoring (Vector Space Model) CE-324: Modern Information Retrieval Sharif University of Technology

RETRIEVAL MODELS. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

COMS F18 Homework 3 (due October 29, 2018)

Information Retrieval

Comparative Document Analysis for Large Text Corpora

Retrieval by Content. Part 2: Text Retrieval Term Frequency and Inverse Document Frequency. Srihari: CSE 626 1

KSU Team s System and Experience at the NTCIR-11 RITE-VAL Task

On the Impact of Objective Function Transformations on Evolutionary and Black-Box Algorithms

Text mining and natural language analysis. Jefrey Lijffijt

The Use of Dependency Relation Graph to Enhance the Term Weighting in Question Retrieval

Semantics with Dense Vectors. Reference: D. Jurafsky and J. Martin, Speech and Language Processing

A Framework of Detecting Burst Events from Micro-blogging Streams

CS630 Representing and Accessing Digital Information Lecture 6: Feb 14, 2006

Convergence, Steady State, and Global Gains of Agent-Based Coalition Formation in E-markets

February 7, Jay Krafthefer, L.S.

Information Retrieval and Organisation

CS276A Text Information Retrieval, Mining, and Exploitation. Lecture 4 15 Oct 2002

Fielded Sequential Dependence Model for Ad-Hoc Entity Retrieval in the Web of Data

Modeling User s Cognitive Dynamics in Information Access and Retrieval using Quantum Probability (ESR-6)

DiscoveryGate SM Version 1.4 Participant s Guide

Language Models. Web Search. LM Jelinek-Mercer Smoothing and LM Dirichlet Smoothing. Slides based on the books: 13

Hyperlinked-Induced Topic Search (HITS) identifies. authorities as good content sources (~high indegree) HITS [Kleinberg 99] considers a web page

Spatial Role Labeling CS365 Course Project

Text Mining. Dr. Yanjun Li. Associate Professor. Department of Computer and Information Sciences Fordham University

Classification Based on Logical Concept Analysis

Chapter 10: Information Retrieval. See corresponding chapter in Manning&Schütze

THE POLAR DECOMPOSITION FOR ADJOINTABLE OPERATORS ON HILBERT C -MODULES AND CENTERED OPERATORS

AIMS Education Foundation

Trade Space Exploration with Ptolemy II

Language Models. CS6200: Information Retrieval. Slides by: Jesse Anderton

No. of Days. Building 3D cities Using Esri City Engine ,859. Creating & Analyzing Surfaces Using ArcGIS Spatial Analyst 1 7 3,139

Using the Past To Score the Present: Extending Term Weighting Models Through Revision History Analysis

Modeling the Score Distributions of Relevant and Non-relevant Documents

A Surface-Similarity Based Two-Step Classifier for RITE-VAL

Comparative Summarization via Latent Dirichlet Allocation

arxiv: v1 [cs.ir] 25 Oct 2015

No. of Days. ArcGIS 3: Performing Analysis ,431. Building 3D cities Using Esri City Engine ,859

No. of Days. ArcGIS Pro for GIS Professionals ,431. Building 3D cities Using Esri City Engine ,859

Predictive Analytics on Accident Data Using Rule Based and Discriminative Classifiers

Combining Vector-Space and Word-based Aspect Models for Passage Retrieval

ArcGIS 9 ArcGIS StreetMap Tutorial

Preliminary Results on Social Learning with Partial Observations

From Research Objects to Research Networks: Combining Spatial and Semantic Search

Multi-theme Sentiment Analysis using Quantified Contextual

Ranked IR. Lecture Objectives. Text Technologies for Data Science INFR Learn about Ranked IR. Implement: 10/10/2018. Instructor: Walid Magdy

Automatic Evaluation in Text Summarization

Query Propagation in Possibilistic Information Retrieval Networks

Text Analytics (Text Mining)

A Study of Smoothing Methods for Language Models Applied to Information Retrieval


CLRG Biocreative V

Maximum Achievable Gain of a Two Port Network

Blame and Coercion: Together Again for the First Time

Economic and Social Council 2 July 2015

M E R C E R W I N WA L K T H R O U G H

COMPARISON OF FEATURES FOR DP-MATCHING BASED QUERY-BY-HUMMING SYSTEM

Non-Boolean models of retrieval: Agenda

Correlation Autoencoder Hashing for Supervised Cross-Modal Search

Landsat-based Global Urban Area Map

Toponym Disambiguation using Ontology-based Semantic Similarity

1 Information retrieval fundamentals

Introduction to ArcGIS Server Development

N-gram N-gram Language Model for Large-Vocabulary Continuous Speech Recognition

Machine Learning for natural language processing

Spatial Information Retrieval

RaRE: Social Rank Regulated Large-scale Network Embedding

Modeling User Rating Profiles For Collaborative Filtering

Tables B to D (April 2008) Keep using Tables A 1993 issue - Pay Adjustment Tables


Total Ordering on Subgroups and Cosets

Boolean and Vector Space Retrieval Models

Comparative Document Analysis for Large Text Corpora

Transcription:

GeoTime Retrieval through Passage-based Learning to Rank Xiao-Lin Wang 1,2, Hai Zhao 1,2, Bao-Liang Lu 1,2 1 Center for Brain-Like Computing and Machine Intelligence Department of Computer Science and Engineering, Shanghai Jiao Tong University 2 MOE-Microsoft Key Laboratory for Intelligent Computing and Intelligent Systems Shanghai Jiao Tong University 800 Dong Chuan Rd., Shanghai 200240, China ك. لم. ي. ك ق ىفو فو»«, ك. ىف ه ه ف 属.. و ف ABSTRACT The NTCIR-9 GeoTime task is to retrieve documents to answer such questions as when and where certain events happened. In this paper we propose a Passage-Based Learning to Rank (PGLR) method to address this task. The proposed method recognizes texts both strongly related to the target topics and containing geographic and temporal expressions. The implemented system provides more accurate search results than a system without PGLR. Performance, according to the official evaluation, is average among submitted systems. Categories and Subject Descriptors H.3.3 [Information Search and Retrieval]: Search process General Terms Algorithms, Performance, Experimentation Keywords GeoTime Retrieval, BM25F, Learning to Rank 1. INTRODUCTION Queries that request geographic and temporal information, such as where and when did... happen?, are very common in information retrieval. Therefore, a customized This work was partially supported by the National Natural Science Foundation of China (Grant No. 60903119, Grant No. 61170114 and Grant No. 90820018), the National Basic Research Program of China (Grant No. 2009CB320901), the Science and Technology Commission of Shanghai Municipality (Grant No. 09511502400), and the European Union Seventh Framework Programme (Grant No. 247619). Corresponding author search method processing geographic and temporal expressions may benefit a large number of users. The NTCIR-9 GeoTime task provides researchers with a platform for testing such methods. The Center for Brain-like Computing and Machine Intelligence, Shanghai Jiao Tong University (SJTU-BCMI) participates in the NTCIR-9 GeoTime English subtask. We propose a Passage-based GeoTime Learning to Rank method (PGLR) to address this task. PGLR first evaluates the relevance and GeoTime informativeness of each passage in candidate documents, which are retrieved by a BM25F algorithm, and then re-ranks the documents according to the maximum score of their passages. This paper takes paragraphs as passages, and various methods of feature extraction and ranking are applied. The underlying intuition is that a document could contain both query terms and GeoTime expressions without any semantic relation between them, and such a document would not be a correct answer for GeoTime Retrieval. On the other hand, if the query terms and GeoTime expressions appear in the same passage, they are more likely to have a semantic relation. A document containing such a passage is more likely to be a correct answer for GeoTime Retrieval. This paper describes our proposed method and discusses its evaluation results. The rest of the paper is organized as follows: the system is presented in Sec. 2; then the submitted runs are described and discussed in Sec. 3; finally this paper in concluded in Sec. 4. 2. SYSTEM DESCRIPTION The core of our system for NTCIR-9 GeoTime is to employ a passage-based GeoTime Learning to Rank (PGLR) method to improve the relevance related to GeoTime retrieval. Figure 1 presents its framework where PGLR is used as post processing after page retrieval and page rank. The system works as follows: 1. Parse the GeoTime topics into query terms Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. NTCIR-9 Workshop Meeting Tokyo, Japan. Copyright 2011 ACM X-XXXXX-XX-X/XX/XX...$10.00. 2. Find the top-n relevant documents by BM25F similarity; 3. Re-rank with the PGLR algorithm; The following three subsections give details. 64

GeoTime Question Query Parser Query Unigrams Query Unigrams, Bigrams, Dependancies BM25F Search Passage-based Learning to Rank Articles Figure 1: Framework of SJTU-BCMI s system for NTCIR-9 GeoTime task 2.1 GeoTime Topic Parsing The representation of a GeoTime topic is normally a question that starts with when and where noted as DESCRIP- TION, a small piece of narrowing description noted as NAR- RATIVE. In this step they are parsed into queries. Three kinds of representations unigram, bigram and dependency are employed in order to catch the semantics of a GeoTime topic. Tab. 1 gives four examples in NTCIR9-GeoTime. The query unigrams, bigrams and dependencies are generated according to the following rules : Unigrams of DESCRIPTION(UoD) are the unigrams of the DESCRIPTION not in a customized stop word list. This list consists of common English stop words plus words with high frequency in GeoTime topics such as when and where. Bigrams of DESCRIPTION (BoD) are pairs of UoDs which are neighbors, with stop words skipped. Dependencies of DESCRIPTION (DoD) are the results of the Stanford parser 1 [3, 1] that have at least one UoD. Unigrams of NARRATIVE (UoN) are the unigrams which appear in the bigrams and dependencies of NARRATIVE (BoN and DoN). Bigrams of NARRATIVE (BoN) are those neighboring non-stop unigrams in NARRATIVE that have at least one UoD. Dependencies of NARRATIVE (DoN) are those dependencies in NARRATIVE that have at least one UoD. 2.2 Document Retrieval with BM25F The document retrieval component is a modified Lucene search engine. Lucene is an open source search engine from the Apache Software Foundation 2. The default similarity model of Lucene is a combination of a boolean model and a vector space model 3. We implemented the Standard B- M25F similarity formula on Lucene [6, 9], and this slightly raises the accuracy according to our pilot experiments. The implementation of BM25F follows Eq. 1 presented at [4]. و. م ف - م /م ف 属 ن / لم.ل ن ف. //: و 1 /ه.موكف ف.م مك //: و 2 /ه /م ك/ى ف/ ٢ ٩ ٠ /ف في/ه.موكف ف.م مك //: و 3 و. ى ف ى ىس/وك فم /م مك /موكف ف BM25F (d, q) = t q tf(t, d) idf(t) (1) k 1 + tf(t, d) tf(t, d) = c d w c tf c (t, d) (2) tf c (t, d) = occur c (t, d) l 1 b c + b d,c c l c (3) idf(t) = N n(t) + 0.5 log n(t) + 0.5 (4) where d is a document, q is a query, t represents a word, c represents a field contained in d, occur c(t, d) is the number of occurrences of t in c of d, l d,c is the length of c at d, l c is the average length of c, N is the number of documents, n(t) is the number of documents containing t, and k 1,w c and b c are parameters. tf c (t, d) is called the field term frequency function, tf(t, d) is called the term frequency function, and idf(t) is called the inverse document frequency. In our system, the title and body of a document are taken as the two fields in BM25F (please see Sec. 3 for details). 2.3 Passage-based GeoTime Learning to Rank (PGLR) The intuition of PGLR is that if a document has a small piece of text both strongly related to the GeoTime topic, and containing temporal and geographic expressions, this document is highly likely to be a correct one. By observing many documents of real-world news, the unit of passage is taken as the granule for computation. Learning to rank [7] is taken to evaluate the rightness of each passage given a GeoTime Topic. The passages of previously retrieved documents are first converted to vector representations based on a group of features; then a trained SVM rank [2] 4 predicts the rightness of these passages. In the end the documents are re-ranked according to the maximum rightness of their passages. The following nine features are used to represent passages. the document s BM25F score the TF IDF similarity of DESCRIPTION s unigrams the TF IDF similarity of DESCRIPTION s bigrams the TF IDF similarity of DESCRIPTION s dependencies / وهى /ي /م م / لم. م ك. ك. 属属属 //: و 4 و. ف 65

Table 1: Examples of Parsing GeoTime Questions into Queries ID Topic Query unigram Query bigram Query dependency 26 (D)Where and when did the space space shuttle columbia space shuttle shuttle space shuttle Columbia disaster take disaster shuttle columbia disaster columbia place? columbia disaster shuttle disaster and in what state the space shuttle Columbia exploded. state space shuttle columbia explode state space shuttle columbia columbia explode space shuttle columbia space columbia shuttle explode columbia 27 (D)When was the last flight of Concorde and where did it land? was the last time that the supersonic airliner Concorde flew and in what city it landed. last flight concorde land last flight flight concorde concorde land last time supersonic last time airliner concorde airliner concorde fly city concorde fly city land land flight last flight concorde time last concorde supersonic concorde airliner fly concorde time land fly land 28 (D)When and where were the Washington beltway snipers arrested? and in what state were the Washington snipers arrested who killed ten people and critically injured three more. washington sniper arrest beltway state washington sniper arrest kill washington beltway beltway sniper sniper arrest state washington washington sniper sniper arrest arrest kill arrest washington sniper beltway washington sniper sniper state sniper washington sniper arrest sniper kill 29 (D)When was the euro put in circulation and which three member states of the eurozone by that time declined its use? and in what state were the Washington snipers arrested who killed ten people and critically injured three more. euro put circulation three member state eurozone time decline use exact euro official currency eurozone enter circulation information 15 member state european union adopt euro require euro put put circulation circulation three three member member state state eurozone eurozone time time decline decline use exact euro euro official currency eurozone eurozone enter enter circulation circulation information 15 member member state state european adopt euro euro require put euro put circulation state three state member decline state state eurozone eurozone time put decline decline use enter euro euro currency currency eurozone enter circulation state member adopt state state union adopt euro 66

the TF IDF similarity of NARRATIVE s unigrams the TF IDF similarity of NARRATIVE s bigrams the TF IDF similarity of NARRATIVE s dependencies whether it contains temporary named entities recognized by Stanford s parser whether it contains geographic named entities recognized by Stanford s parser 3. EXPERIMENTS 3.1 Experimental Setting Four runs of SJTU-BCMI were submitted to NTCIR-9 GeoTime task. Table 2 summarizes the detailed approach of each run. The parameters of BM25F are tuned to maximum performance on the dataset of the NTCIR-8 GeoTime task. The features are transferred into vectors for classification. SVM rank is used to learn the weights of features, and the results are presented in the last column of Table 2. 3.2 Experimental Results NTCIR-9 INTENT-DR employs five evaluation matrices: MAP,Q, NDCG@10, NDCG@100 and NDCG@1000. The evaluation results for SJTU-BCMI s runs are presented in Table 3. The runs using the PGLR component provide higher accuracies as expected, since scores of RUN2 and RUN4 are higher than those of RUN1 and RUN3. [5] D. Ravichandran and E. Hovy. Learning surface text patterns for a question answering system. In Proceedings of the 40th Annual Meeting on Association for Computational Linguistics, pages 41 47. Association for Computational Linguistics, 2002. [6] S. Robertson, S. Walker, S. Jones, M. Hancock-Beaulieu, and M. Gatford. Okapi at trec-3. In Overview of the Third Text REtrieval Conference (TREC-3), pages 109 126. NIST, 1995. [7] A. Trotman. Learning to rank. Information Retrieval, 8(3):359 381, 2005. [8] E. Voorhees and H. Dang. Overview of the trec 2002 question answering track. NIST SPECIAL PUBLICATION SP, pages 57 68, 2003. [9] H. Zaragoza, N. Craswell, M. Taylor, S. Saria, and S. Robertson. Microsoft cambridge at trec-13: Web and hard tracks. In Proceedings of TREC 2004. Citeseer, 2004. 4. CONCLUSIONS In this paper, we proposed a Passage-based GeoTime Learning to Rank (PGLR) method to address the problem of Geo- Time retrieval. The implemented system achieves above average performances on the NTCIR-9 GeoTime task. The proposed PGLR obviously improves the GeoTime-related relevance according to the official evaluations of comparison experiments. The idea of PGLR is actually to recognize the rightness of a piece of text serving as the answer to a GeoTime topic (or question). In this paper we employ a linear model with the input of TF IDF similarities and presence of temporal and geographical expressions. In the future, we plan to incorporate more technologies from the domain of automatically question answering [5, 8] 5. REFERENCES [1] M. De Marneffe, B. MacCartney, and C. Manning. Generating typed dependency parses from phrase structure parses. In LREC 2006. Citeseer, 2006. [2] T. Joachims. Training linear svms in linear time. In Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 217 226. ACM, 2006. [3] D. Klein and C. Manning. Accurate unlexicalized parsing. In Proceedings of the 41st Annual Meeting on Association for Computational Linguistics-Volume 1, pages 423 430. Association for Computational Linguistics, 2003. [4] J. Pérez-Agüera, J. Arroyo, J. Greenberg, J. Iglesias, and V. Fresno. Using bm25f for semantic search. In Proceedings of the 3rd International Semantic Search Workshop, pages 1 8. ACM, 2010. 67

Table 2: Description of SJTU-BCMI submitted runs to GeoTime English subtask Runs NARRATIVE? Doc. Retrieval PGLR?(weight-s) SJTUBCMI-EN-EN-01-DN Yes BM25F a No SJTUBCMI-EN-EN-02-DN Yes BM25F a Yes(6.8,1.1,0.1,0.0,1.2,0.3,0.0,0.6,0.5) b SJTUBCMI-EN-EN-03-D No BM25F a No SJTUBCMI-EN-EN-04-D No BM25F a Yes(6.8,1.1,0.1,0.0,0.6,0.5) c a The parameters are k 1=2, b title =0.1, w title =4, b main=0.1,w main=2. b corresponding to the features in Sec. 2.3. c corresponding to the features in Sec. 2.3 with three NARRATIVE s similarities removed. Table 3: Evaluation of SJTU-BCMI submitted runs to NTCIR-9 Intent-DR Chinese subtask RunName MAP Q ndcg@10 ndcg@100 ndcg@1000 SJTUBCMI-EN-EN-01-DN 0.3326 0.3557 0.4498 0.4511 0.5772 SJTUBCMI-EN-EN-02-DN 0.3648 0.3884 0.5127 0.4977 0.6045 SJTUBCMI-EN-EN-03-D 0.2885 0.3111 0.4217 0.4085 0.5338 SJTUBCMI-EN-EN-04-D 0.3141 0.3370 0.4429 0.4523 0.5567 MAXIMUM 0.5549 0.5738 0.7379 0.6883 0.7654 AVERAGE 0.3517 0.3710 0.4794 0.4676 0.5684 MEDIAN 0.3326 0.3512 0.4591 0.4563 0.5772 MINIMUM 0.1392 0.1474 0.1712 0.2409 0.3195 68