A Trainable Multi-factored QA System Radu Ion, Dan Ştefănescu, Alexandru Ceauşu, Dan Tufiş, Elena Irimia, Verginica Barbu-Mititelu Research Institute for.

Slides:



Advertisements
Similar presentations
On-line Compilation of Comparable Corpora and Their Evaluation Radu ION, Dan TUFIŞ, Tiberiu BOROŞ, Alexandru CEAUŞU and Dan ŞTEFĂNESCU Research Institute.
Advertisements

Chapter 5: Introduction to Information Retrieval
TI: An Efficient Indexing Mechanism for Real-Time Search on Tweets Chun Chen 1, Feng Li 2, Beng Chin Ooi 2, and Sai Wu 2 1 Zhejiang University, 2 National.
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
Information Retrieval IR 7. Recap of the last lecture Vector space scoring Efficiency considerations Nearest neighbors and approximations.
Web Search - Summer Term 2006 II. Information Retrieval (Basics Cont.)
Search Engines and Information Retrieval
IR Challenges and Language Modeling. IR Achievements Search engines  Meta-search  Cross-lingual search  Factoid question answering  Filtering Statistical.
T.Sharon - A.Frank 1 Internet Resources Discovery (IRD) IR Queries.
Modern Information Retrieval Chapter 2 Modeling. Can keywords be used to represent a document or a query? keywords as query and matching as query processing.
Information Retrieval in Practice
A machine learning approach to improve precision for navigational queries in a Web information retrieval system Reiner Kraft
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
Computer comunication B Information retrieval. Information retrieval: introduction 1 This topic addresses the question on how it is possible to find relevant.
Information retrieval Finding relevant data using irrelevant keys Example: database of photographic images sorted by number, date. DBMS: Well structured.
INEX 2003, Germany Searching in an XML Corpus Using Content and Structure INEX 2003, Germany Yiftah Ben-Aharon, Sara Cohen, Yael Grumbach, Yaron Kanza,
Chapter 5: Information Retrieval and Web Search
CSCI 5417 Information Retrieval Systems Jim Martin Lecture 6 9/8/2011.
CS344: Introduction to Artificial Intelligence Vishal Vachhani M.Tech, CSE Lecture 34-35: CLIR and Ranking in IR.
Evaluation David Kauchak cs458 Fall 2012 adapted from:
Evaluation David Kauchak cs160 Fall 2009 adapted from:
COMP423: Intelligent Agent Text Representation. Menu – Bag of words – Phrase – Semantics – Bag of concepts – Semantic distance between two words.
The use of machine translation tools for cross-lingual text-mining Blaz Fortuna Jozef Stefan Institute, Ljubljana John Shawe-Taylor Southampton University.
Search Engines and Information Retrieval Chapter 1.
CLEF Ǻrhus Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Oier Lopez de Lacalle, Arantxa Otegi, German Rigau UVA & Irion: Piek Vossen.
By : Garima Indurkhya Jay Parikh Shraddha Herlekar Vikrant Naik.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Michael Cafarella Alon HalevyNodira Khoussainova University of Washington Google, incUniversity of Washington Data Integration for Relational Web.
Intelligent Database Systems Lab Advisor : Dr. Hsu Graduate : Chien-Shing Chen Author : Satoshi Oyama Takashi Kokubo Toru lshida 國立雲林科技大學 National Yunlin.
Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.
Interactive Probabilistic Search for GikiCLEF Ray R Larson School of Information University of California, Berkeley Ray R Larson School of Information.
A Language Independent Method for Question Classification COLING 2004.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
Chapter 6: Information Retrieval and Web Search
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
A Novel Pattern Learning Method for Open Domain Question Answering IJCNLP 2004 Yongping Du, Xuanjing Huang, Xin Li, Lide Wu.
Evaluation of (Search) Results How do we know if our results are any good? Evaluating a search engine  Benchmarks  Precision and recall Results summaries:
CLEF Kerkyra Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Arantxa Otegi UNIPD: Giorgio Di Nunzio UH: Thomas Mandl.
Improving Named Entity Translation Combining Phonetic and Semantic Similarities Fei Huang, Stephan Vogel, Alex Waibel Language Technologies Institute School.
Information Retrieval using Word Senses: Root Sense Tagging Approach Sang-Bum Kim, Hee-Cheol Seo and Hae-Chang Rim Natural Language Processing Lab., Department.
What Does the User Really Want ? Relevance, Precision and Recall.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
1 13/05/07 1/20 LIST – DTSI – Interfaces, Cognitics and Virtual Reality Unit The INFILE project: a crosslingual filtering systems evaluation campaign Romaric.
AQUAINT AQUAINT Evaluation Overview Ellen M. Voorhees.
Improving QA Accuracy by Question Inversion John Prager, Pablo Duboue, Jennifer Chu-Carroll Presentation by Sam Cunningham and Martin Wintz.
Information Retrieval Quality of a Search Engine.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Information Retrieval Lecture 3 Introduction to Information Retrieval (Manning et al. 2007) Chapter 8 For the MSc Computer Science Programme Dell Zhang.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Information Retrieval in Practice
Language Identification and Part-of-Speech Tagging
An Efficient Algorithm for Incremental Update of Concept space
Tolerant Retrieval Review Questions
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Language Technologies Institute Carnegie Mellon University
Search Engine Architecture
Information Retrieval and Web Search
A research literature search engine with abbreviation recognition
Multimedia Information Retrieval
of the Artificial Neural Networks.
Data Integration for Relational Web
CS246: Information Retrieval
Search Engine Architecture
Question Answer System Deliverable #2
WSExpress: A QoS-Aware Search Engine for Web Services
Introduction to Search Engines
Machine Reading.
CoXML: A Cooperative XML Query Answering System
Presentation transcript:

A Trainable Multi-factored QA System Radu Ion, Dan Ştefănescu, Alexandru Ceauşu, Dan Tufiş, Elena Irimia, Verginica Barbu-Mititelu Research Institute for Artificial Intelligence, Romanian Academy

ResPubliQA We participated in the Romanian-Romanian ResPubliQA task 500 juridical questions to be answered from the Romanian JRC Acquis (10714 docs) Questions have been translated from other languages => a more difficult QA task since translated terms are not necessarily expressed the same in the actual Romanian documents

Corpus processing and indexing POS tagging, lemmatization, chunking. Only the ‘body’ part of a document was indexed (no annexes, no headers) We have two Lucene indexes: a document index and a paragraph index What’s in the index: lemmas and paragraph classes for the paragraph index

QA flow Web services based: –Question preprocessing using TTL ( –Question classification using a ME classifier ( –Query generation (2 types: TFIDF and chunk based) ( –Search engine interrogation ( –Paragraph relevance score computation and paragraph reordering

The combined QA system In order to account for NOA strings (which, when given, will increase the overall performance measure) we decided to combine 2 results: –The QA system using the TFIDF query –The QA system using the chunk query When the same paragraph was returned among the top K (=3) paragraphs by the two systems, it was the answer For the other case, we returned the NOA string

Paragraph relevance s 1 to s 5 are paragraph relevance scores λ i are trained weights by iteratively computing MRR scores on a 200 questions test set using sets of weights for which the sum is 1. Retaining the value of the weights that account for the largest obtained MRR, results in a MERT- like training procedure Increment step was 0.01

Relevance scores Lucene scores for the document and paragraph retrieval One BLUE-like relevance score which is high if a candidate paragraph contains keywords much in same order as in the question One indicator variable that is 1 if the candidate paragraph has the same class as the question (0 otherwise) One lexical chains based score (a real number quantified semantic distance between the question and the candidate paragraph)

Evaluations Official results Second run: query contained the question class

Post CLEF2009 Evaluations Results with all questions (500) answered (no NOA strings) With trained parameters for every question class, we obtain an overall accuracy of (29 additional correctly answered questions)

Post CLEF2009 Evaluations (II) Some other informative measures: –Answering precision: correct / answered –Rejection precision: (1 – correct) / unanswered AP(icia092roro) = 75.58% RP(icia092roro) = 86.53% So, the system is able to reject giving wrong answers at a high rate which is a merit in itself (discovered due to the calculus) even if it cannot offer the same answering precision in the unanswered

Conclusions A multi-factored QA system may be easily extended with new paragraph relevance scores It’s also easily adaptable on new domains and/or languages Update: better correlation between documents and paragraph relevance scores Future plans: to develop the English QA system along the same lines and combine the En-Ro outputs

Conclusions (II) Competition drives innovation but let’s not forget that these tools are there to help users Useful requirement: QA systems to be on the Web Ours is at