Integrating linguistic knowledge in passage retrieval for question answering J¨org Tiedemann Alfa Informatica, University of Groningen HLT/EMNLP 2005.

Slides:



Advertisements
Similar presentations
Chapter 5: Introduction to Information Retrieval
Advertisements

Improved TF-IDF Ranker
QA-LaSIE Components The question document and each candidate answer document pass through all nine components of the QA-LaSIE system in the order shown.
ANNIC ANNotations In Context GATE Training Course 27 – 28 April 2006 Niraj Aswani.
January 12, Statistical NLP: Lecture 2 Introduction to Statistical NLP.
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Information Retrieval in Practice
Ang Sun Ralph Grishman Wei Xu Bonan Min November 15, 2011 TAC 2011 Workshop Gaithersburg, Maryland USA.
NaLIX: A Generic Natural Language Search Environment for XML Data Presented by: Erik Mathisen 02/12/2008.
WMES3103 : INFORMATION RETRIEVAL
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
1 Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang, Assistant Professor Dept. of Computer Science & Information Engineering National Central.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Overview of Search Engines
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 30, (2014) BERLIN CHEN, YI-WEN CHEN, KUAN-YU CHEN, HSIN-MIN WANG2 AND KUEN-TYNG YU Department of Computer.
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
Chapter 4 Query Languages.... Introduction Cover different kinds of queries posed to text retrieval systems Keyword-based query languages  include simple.
1 LOMGen: A Learning Object Metadata Generator Applied to Computer Science Terminology A. Singh, H. Boley, V.C. Bhavsar National Research Council and University.
1 A study on automatically extracted keywords in text categorization Authors:Anette Hulth and Be´ata B. Megyesi From:ACL 2006 Reporter: 陳永祥 Date:2007/10/16.
Survey of Semantic Annotation Platforms
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Information Extraction From Medical Records by Alexander Barsky.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Chapter 2 Architecture of a Search Engine. Search Engine Architecture n A software architecture consists of software components, the interfaces provided.
Querying Structured Text in an XML Database By Xuemei Luo.
1 University of Palestine Topics In CIS ITBS 3202 Ms. Eman Alajrami 2 nd Semester
MIRACLE Multilingual Information RetrievAl for the CLEF campaign DAEDALUS – Data, Decisions and Language, S.A. Universidad Carlos III de.
Péter Schönhofen – Ad Hoc Hungarian → English – CLEF Workshop 20 Sep 2007 Performing Cross-Language Retrieval with Wikipedia Participation report for Ad.
A Language Independent Method for Question Classification COLING 2004.
Answering Definition Questions Using Multiple Knowledge Sources Wesley Hildebrandt, Boris Katz, and Jimmy Lin MIT Computer Science and Artificial Intelligence.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
NTCIR /21 ASQA: Academia Sinica Question Answering System for CLQA (IASL) Cheng-Wei Lee, Cheng-Wei Shih, Min-Yuh Day, Tzong-Han Tsai, Tian-Jian Jiang,
Playing Biology ’ s Name Game: Identifying Protein Names In Scientific Text Daniel Hanisch, Juliane Fluck, Heinz-Theodor Mevissen and Ralf Zimmer Pac Symp.
Comparing and Ranking Documents Once our search engine has retrieved a set of documents, we may want to Rank them by relevance –Which are the best fit.
SINGULAR VALUE DECOMPOSITION (SVD)
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
1 Multi-Perspective Question Answering Using the OpQA Corpus (HLT/EMNLP 2005) Veselin Stoyanov Claire Cardie Janyce Wiebe Cornell University University.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
A Novel Pattern Learning Method for Open Domain Question Answering IJCNLP 2004 Yongping Du, Xuanjing Huang, Xin Li, Lide Wu.
Mining Binary Constraints in Feature Models: A Classification-based Approach Yi Li.
1 Automatic indexing Salton: When the assignment of content identifiers is carried out with the aid of modern computing equipment the operation becomes.
From Text to Image: Generating Visual Query for Image Retrieval Wen-Cheng Lin, Yih-Chen Chang and Hsin-Hsi Chen Department of Computer Science and Information.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
1 Language Specific Crawler for Myanmar Web Pages Pann Yu Mon Management and Information System Engineering Department Nagaoka University of Technology,
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
4. Relationship Extraction Part 4 of Information Extraction Sunita Sarawagi 9/7/2012CS 652, Peter Lindes1.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Exploiting Named Entity Taggers in a Second Language Thamar Solorio Computer Science Department National Institute of Astrophysics, Optics and Electronics.
Concepts and Realization of a Diagram Editor Generator Based on Hypergraph Transformation Author: Mark Minas Presenter: Song Gu.
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
ENHANCING CLUSTER LABELING USING WIKIPEDIA David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab SIGIR’09.
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Analysis of Experiments on Hybridization of different approaches in mono and cross-language information retrieval DAEDALUS – Data, Decisions and Language,
Feature Assignment LBSC 878 February 22, 1999 Douglas W. Oard and Dagobert Soergel.
Information Retrieval in Practice
Korean version of GloVe Applying GloVe & word2vec model to Korean corpus speaker : 양희정 date :
Linguistic Graph Similarity for News Sentence Searching
Search Engine Architecture
Text Based Information Retrieval
Web News Sentence Searching Using Linguistic Graph Similarity
CS 430: Information Discovery
Multimedia Information Retrieval
Introduction to XML IR XML Group.
Presentation transcript:

Integrating linguistic knowledge in passage retrieval for question answering J¨org Tiedemann Alfa Informatica, University of Groningen HLT/EMNLP 2005

Abstract This paper use annotation produced by a deep syntactic dependency parser for Dutch, Alpino, to extract various kinds of linguistic features and syntactic units to be included in a multi-layer index Use a genetic algorithm to optimize the selection of features and syntactic units to be included in a query The system is trained on questions from the competition on Dutch QA within CLEF Results show an improvement of about 15% in mean total reciprocal rank (MTRR) compared to traditional IR using plain text keywords

Introduction The goal is to utilize information that can be extracted from the analyzed question in order to match linguistic features and syntactic units in analyzed documents Furthermore, we have to find an appropriate way of combining query terms to optimize IR performance –Apply an iterative learning approach based on example questions annotated with their answers

QA with Dependency Relations This Dutch QA System, Joost, consists of two streams: Off-line strategy –Table look-up strategy –Use syntactic patterns to extract information from unrestricted text to be stored in fact tables On-line strategy –Use passage retrieval and on-the-fly answer extraction Furthermore, use strategies for reasoning over dependency rules to capture semantic relationships that are expressed by different syntactic patterns

QA with Dependency Relations (cont) We have parsed the entire corpus provided by CLEF with about 4,000,000 sentences in about 190,000 documents The dependency trees are stored in XML and are directly accessible from the QA system Syntactic patterns for off-line information extraction are run on the entire corpus For on-line strategy, we use traditional information retrieval to select relevant passages from the corpus to be processed by the answer extraction modules Questions are parsed in the same manner. According to the expected answer type (1) find the answer first in the fact database (if an appropriate table exists) and then (2) in the corpus using the on-line QA strategy

Passage Retrieval in Joost IR engine: Lucene –Widely-used open-source Java library with several extensions and useful features Baseline –Lucene with standard settings and a Dutch text analyzer for stemming and stopword removal Extension –Figure below shows a dependency tree for one sentence in the CLEF corpus

Passage Retrieval in Joost (cont) Extension –From the parse trees, we extract various kinds of linguistic features and syntactic units to be stored in the index –Besides the dependency relations, the parser also produces part-of-speech (POS) tags, named entity labels and linguistic root forms –It also recognizes compositional compounds and particle verbs

Passage Retrieval in Joost (cont) Lucene supports multiple index fields, called multi-layer index –Token layers, type layers and annotation layers

Passage Retrieval in Joost (cont) Token layers –Features are simply concatenated (using special delimiting symbols between the various parts) to create individual items in the layer –Ex. RootHead layer contains concatenated dependent-head bigrams taken from the dependency relations in the tree –Tokens in the text layer and in the root layer have been split at hyphens and underscores to split compositional compounds and particle verbs Type layers –Include only specific types of tokens in the corpus Annotation layers –This layer may contain the items ’ORG’, ’PER’ or ’LOC’ if such a named entity occurs in the text passage

Query Formulation Questions are analyzed in the same way, so we can extract appropriate units from analyzed questions to be matched with the various layers in the index. –For example, we can extract root-head word pairs to be matched with the RootHead layer Possible keyword types in passage retrieval component

Query Formulation (cont) Query keywords from all types can be combined into a single query (connect them in a disjunctive way) For example, form a query from Figure 2 with the following elements: (1) all plain text tokens; (2) named entities (ne) boosted with factor 2; (3) RootHead bigrams where the root is in an object relation; (4) RootRel keywords for all nouns

Query Formulation (cont) Add “required-marker” (’+’) and two elements to the query: (5) plain text keywords in an object relation with boost factor 3 and (6) plain text keywords labeled as names marked as required Altogether we have 109 different keyword types using the layers and the restrictions defined above Now the question is to select appropriate keyword types among them with the optimal parameters (weights) to maximize retrieval performance

Optimization of Query Parameters Training and evaluation data: results from the CLEF competition on Dutch QA from the years 2003 and 2004 –Contain natural language questions annotated with their answers found in the CLEF corpus –Altogether there are 631 questions with 851 answers we use the mean of the total reciprocal ranks (MTRR) to measure the performance

Optimization of Query Parameters (cont) In the experiments, provided answer was used string rather than the document ID to judge if a retrieved passage was relevant or not Obviously, this introduces errors where the matching string does not correspond to a valid answer in the context However, we believe that this does NOT influence the global evaluation figure significantly and therefore we use this approach as a reasonable compromise when doing automatic evaluation

Learning Query Parameters There is a large variety of possible keyword types that can be combined to query the multi- layer index Furthermore, we have a number of parameters to be set when formulating a query, e.g. the keyword weights We like to carry out a systematic search for optimizing parameters rather than using our intuition. Hence, we need an iterative approach for testing several settings to optimize query parameters

Learning Query Parameters (cont) We decided to use a simplified genetic algorithm to optimize query parameters. The optimization loop works as follows:

Learning Query Parameters (cont) Some additional parameters: –Population size, i.e. the number of IR runs to be kept for the next iteration: 25 –Fitness function: MTRR measure –Selection: top-N search –Max number of settings to be created: 50 –Combination: simply merge the settings of two previous runs (parents) to produce a new setting (a child)

Learning Query Parameters (cont) –Mutation All these parametes are intuitively chosen Assigned rather high probabilities to reduce the risk of local maximum traps

Experiments 420 questions from the CLEF data for training and used the remaining 150 questions for evaluation Retrieved a maximum of 20 passages per question For each setting we computed the fitness scores for the training set and the evaluation set using MTRR The top scores have been printed after each 10 runs and compared to the evaluation scores Figure 3 shows a plot of the fitness score development throughout the optimization process in comparison with the evaluation scores

Experiments (cont) The base-line of refers to the retrieval result on evaluation data when using traditional IR with plain text keywords only (i.e. using the text layer, Dutch stemming and stop word removal) The base-line performance on training data is slightly worse with MTRR After 1130 settings the MTRR scores increased to for training data and for evaluation data Thereafter we can observe a surprising drop in evaluation scores to around 0.97 in MTRR. This might be due to overfitting although the drop seems to be rather radical

Experiments (cont) The MTRR score after 3200 settings is at on evaluation data which is a statistically significant (tested using the Wilcoxon matched-pairs signed-ranks test at p < 0.01) MTRR measured on document IDs and evaluation data did also increase from to which is statistically significant at p¡0.02. Coverage went up from 78.68% to 81.62% on evaluation data Redundancy was improved from to (significance tests have not been carried out) Finally, the QA performance with only the IR based strategy was increased from (using CLEF scores) to This, however, is not statistically significant and may be due to chance. Table 3 shows the features and weights selected in the training process

Experiments (cont)

The largest weights are given to (1)names in the text layer, to (2)root forms of names in modifier relations and to (3)plain text adjectives A surprisingly large number of keyword types are marked as required. Some of them overlap with each other and are therefore redundant –For example, all RootPOS keywords are marked as required and therefore, the restrictions of RootPOS keywords are useless because they do not alter the query

Conclusions Integrating linguistic information derived from dependency analyses in passage retrieval for question answering Demonstrated a genetic algorithm for optimizing query parameters to improve the retrieval performance Results show a significant improvement of about 15% in mean total reciprocal rank using extended queries with optimized parameters