Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Slides:



Advertisements
Similar presentations
Improved TF-IDF Ranker
Advertisements

SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Vikas BhardwajColumbia University NLP for the Web – Spring 2010 Improving QA Accuracy by Question Inversion Prager et al. IBM T.J. Watson Res. Ctr. 02/18/2010.
Query Operations: Automatic Local Analysis. Introduction Difficulty of formulating user queries –Insufficient knowledge of the collection –Insufficient.
The Informative Role of WordNet in Open-Domain Question Answering Marius Paşca and Sanda M. Harabagiu (NAACL 2001) Presented by Shauna Eggers CS 620 February.
Gimme’ The Context: Context- driven Automatic Semantic Annotation with CPANKOW Philipp Cimiano et al.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 30, (2014) BERLIN CHEN, YI-WEN CHEN, KUAN-YU CHEN, HSIN-MIN WANG2 AND KUEN-TYNG YU Department of Computer.
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
Hang Cui et al. NUS at TREC-13 QA Main Task 1/20 National University of Singapore at the TREC- 13 Question Answering Main Task Hang Cui Keya Li Renxu Sun.
Query Processing: Query Formulation Ling573 NLP Systems and Applications April 14, 2011.
An Integrated Approach to Extracting Ontological Structures from Folksonomies Huairen Lin, Joseph Davis, Ying Zhou ESWC 2009 Hyewon Lim October 9 th, 2009.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Querying Structured Text in an XML Database By Xuemei Luo.
Annotating Words using WordNet Semantic Glosses Julian Szymański Department of Computer Systems Architecture, Faculty of Electronics, Telecommunications.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
1 Query Operations Relevance Feedback & Query Expansion.
Probabilistic Query Expansion Using Query Logs Hang Cui Tianjin University, China Ji-Rong Wen Microsoft Research Asia, China Jian-Yun Nie University of.
Retrieval Models for Question and Answer Archives Xiaobing Xue, Jiwoon Jeon, W. Bruce Croft Computer Science Department University of Massachusetts, Google,
QUALIFIER in TREC-12 QA Main Task Hui Yang, Hang Cui, Min-Yen Kan, Mstislav Maslennikov, Long Qiu, Tat-Seng Chua School of Computing National University.
A Language Independent Method for Question Classification COLING 2004.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
10/22/2015ACM WIDM'20051 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis Voutsakis.
INTERESTING NUGGETS AND THEIR IMPACT ON DEFINITIONAL QUESTION ANSWERING Kian-Wei Kor, Tat-Seng Chua Department of Computer Science School of Computing.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
A Novel Pattern Learning Method for Open Domain Question Answering IJCNLP 2004 Yongping Du, Xuanjing Huang, Xin Li, Lide Wu.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
LOGO 1 Corroborate and Learn Facts from the Web Advisor : Dr. Koh Jia-Ling Speaker : Tu Yi-Lang Date : Shubin Zhao, Jonathan Betz (KDD '07 )
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
August 17, 2005Question Answering Passage Retrieval Using Dependency Parsing 1/28 Question Answering Passage Retrieval Using Dependency Parsing Hang Cui.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Unsupervised Auxiliary Visual Words Discovery for Large-Scale Image Object Retrieval Yin-Hsi Kuo1,2, Hsuan-Tien Lin 1, Wen-Huang Cheng 2, Yi-Hsuan Yang.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Acquisition of Categorized Named Entities for Web Search Marius Pasca Google Inc. from Conference on Information and Knowledge Management (CIKM) ’04.
Query Suggestions in the Absence of Query Logs Sumit Bhatia, Debapriyo Majumdar,Prasenjit Mitra SIGIR’11, July 24–28, 2011, Beijing, China.
Relevance Models and Answer Granularity for Question Answering W. Bruce Croft and James Allan CIIR University of Massachusetts, Amherst.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
AQUAINT AQUAINT Evaluation Overview Ellen M. Voorhees.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
2/10/2016Semantic Similarity1 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis.
Keyword Translation Accuracy and Cross-Lingual Question Answering in Chinese and Japanese Teruko Mitamura Mengqiu Wang Hideki Shima Frank Lin In CMU EACL.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
2016/3/11 Exploiting Internal and External Semantics for the Clustering of Short Texts Using World Knowledge Xia Hu, Nan Sun, Chao Zhang, Tat-Seng Chu.
Enhanced hypertext categorization using hyperlinks Soumen Chakrabarti (IBM Almaden) Byron Dom (IBM Almaden) Piotr Indyk (Stanford)
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
1 Predicting Answer Location Using Shallow Semantic Analogical Reasoning in a Factoid Question Answering System Hapnes Toba, Mirna Adriani, and Ruli Manurung.
University Of Seoul Ubiquitous Sensor Network Lab Query Dependent Pseudo-Relevance Feedback based on Wikipedia 전자전기컴퓨터공학 부 USN 연구실 G
Queensland University of Technology
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Multimedia Information Retrieval
Web IR: Recent Trends; Future of Web Search
Semantic Similarity Methods in WordNet and their Application to Information Retrieval on the Web Yizhe Ge.
Connecting the Dots Between News Article
Presentation transcript:

Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University of Singapore Sigir’03

Introduction Main task in recent TREC-11: consists of 500 short factoid questions posed over more than one million news articles Traditional “bag-of-words” representation: effective for IR research in retrieving large number of relevant documents, but ineffective for QA where precise answers are needed Moldovan et al: illustrated the power of extensive WordNet to perform logic proof relied on knowledge reasoning Lin gave a detailed comparison of employing the Web and WordNet

Introduction (2) This paper: investigates the integration and structured use of both linguistic and web knowledge for the TREC-style QA, which was called Event-based QA With a high performance QA system called QUALIFIER (QUestion Answering by LexIcal FabrIc and External Resources) and analyze its effectiveness using the TREC-11benchmark. Results show that: the Event-based approach, together with the use of Successive Constraint Relaxation, gives rise to an effective question answering system

Event-based QA Two basic types of things: entities and events From the definitions in WordNet: An entity is “anything having an existence (living or nonliving)” An event is “something that happens at a given place and time”

Event-based QA (2) Table 1 shows the correspondences of the most common WHquestion:

Event-based QA (3) Consider an event to be a point within the multidimensional QA event space:

Event-based QA (4) Difficult to find a correct answer, two major problems: Insufficient known elements (k<<n) Inexact known elements. Solutions: Use of world knowledge (from the Web and WordNet glosses) Exploit the lexical knowledge from WordNet synsets and morphemics and employ a new technique called Successive Constraint Relaxation In addition, perform structured semantic analysis based on what we already know in order to discover the event structure

QUALIFIER System Architecture

QUALIFIER System Architecture (2) Question Analysis: Extract the known QA elements embedded in the question and the expected answer (or targeted unknown element) type. Extracts several sets of words (known elements) from the original question and employs a rule- based question classifier to identify the answer target (unknown element) type. The rule-based question classifier adopts a two- tiered question taxonomy, which achieves a classification accuracy of over 98% on recent year’s TREC questions

QUALIFIER System Architecture (3) Query Formulation: Integrates the knowledge of both the Web and WordNet to extract context information Two ways to processing Linear approach of extracting highly correlated terms from the relevant information derived from the Web/WordNet Structured analysis to discover different facets or elements of events, and employ a structured (event- based) approach to perform query formulation. In this paper, they will concentrate more on the event-based query expansion

QUALIFIER System Architecture (4) Given the newly formulated query: Employs the MG tool to perform Boolean search for top ranked documents in the QA corpus so that we can narrow down the scope of unknown elements. Then selects candidate answer sentences based on the similarity between the sentences and the query in terms of original/expanded query terms, noun phrases, and quotations, etc Finally performs linguistic and named entity analysis on the candidate sentences, and picks a named entity string with high possibility to be the targeted unknown element as the exact answer. Successive Constraint Relaxation is used here to improve the answer coverage

Structured Use of External Knowledge Given a short, fact-based question, it can be transformed to a query in a form of : For example, given the question: “What Spanish explorer discovered the Mississippi river?”, the query q(0) will be [Spanish,explorer, Mississippi, river] In order to constrain the query and to bridge the gap between the query and the QA corpus, we resort to using general open resources to overcome this problem. In this system, they focus on the amalgamation and structured use of the external knowledge like the Web and WordNet

Integrating External Knowledge QUALIFIER uses the original content words q(0) to retrieve the top Nw documents using the Web search engine (E.g. Google)

Integrating External Knowledge (2) Web lacks information on lexical relationships, such as synonyms Solution : employs WordNet to get the glosses and synsets of the original query terms

Integrating External Knowledge (3) Two possible approaches to investigate the potential of Event-based QA: Kq as a list without any structure (Linear Query Formulation) Form semantic word groups before building the new query (Structured Query Formulation)

Formulating Linear Queries If we simply append all the terms, the resulting expanded query will be too broad and contain too many terms out of context. Solution: restrict the glosses and synonyms to only those terms found in the web documents

Formulating Linear Queries (2)

Formulating Structured Queries In order to discover the structure of terms within the external knowledge, we perform structural analysis on Kq to form semantic groups of terms Given any two distinct terms ti, tj ∈ Kq, we compute their lexical, co-occurrence, and distance correlation as follows:

Formulating Structured Queries (2) Lexical correlation: consider only synonym relation

Formulating Structured Queries (3) Co-occurrence correlation:

Formulating Structured Queries (4) Distance correlation:

Formulating Structured Queries (5)

Formulating Structured Queries (6) One promising advantage of this approach: Be able to answer any factual questions about the elements in this QA event other than just “What Spanish explorer discovered the Mississippi river?”. For instance, can easily handle questions like “When was Mississippi river discovered?” and “Which river were discovered by Hernando De Soto?” etc with the same set of knowledge

Successive Constraint Relaxation (SCR) Given the newly formulated linear or structured query, QUALIFIER uses the MG tool to retrieve up to N (N=50) relevant documents within the QA corpus. Boolean retrieval : One of its disadvantages: returns either too many or too few documents However, this can be an advantage for QA as it ensures precision. Successive Constraint Relaxation (SCR): Avoid over-constraining the query; improve answer coverage while preserving precision

Answer Extraction With the set of the top K sentences obtained after document retrieval and sentence ranking, QUALIFIER performs fine-grained named entity tagging before extracting the string that matches the question class as the answer How to detect fine-grained named entities? Using a two-tiered hierarchy as shown in Table 2. It adopts a rule-based algorithm to detect the named entities by utilizing the lexical and semantic features such as capitalization, punctuation, context, part-of- speech tagging and phrase chunking Tests indicate that it could achieve a detection accuracy of over 90%

Answer Extraction (2)

Evaluation – TREC-11 The main task in TREC-11: 500 factoid questions from a QA corpus over one million documents Systems are required to return exactly one response per question A response is either a [answer-string; document-id] pair or the string “NIL”. The pair will be judged as follows: correct: the answer-string consists of exactly a correct answer and that answer is supported by the document returned; unsupported: the answer-string contains a correct answer but the document returned does not support that answer; inexact: the answer-string contains a correct answer and the document supports that answer, but the string contains either partial answer or more than just the answer; and wrong: the answer-string does not contain a correct answer or the answer is not responsive.

Evaluation – TREC-11 (2) TREC-11 employed the Confidence Weighted Score (CWS) to reward systems that successfully rank questions that they answered correctly higher than those answered incorrectly. By using the linear query expansion, it places them among one of the top performing systems.

Linear vs. Structured Query Formulation Several tests: 1) Baseline: perform QA without using the external resources. 2) Baseline + Web: add up to top m context words from Cq into q(1) 3) Baseline + Web + WordNet: combine both Web and WordNet knowledge, but do not constrain the new terms from WordNet. This is to test the effects of adding some WordNet terms out of context. 4) Baseline + Web + WordNet + context constraint: i.e. linear query formulation. 5) Baseline + Web + WordNet + structure analysis: i.e. structured query formulation.

Linear vs. Structured Query Formulation (2) For each test, also compare the results obtained without performing Successive Constraint Relaxation (SCR) to that obtained after 5 iterations of SCR. In these tests, they examine the top 75 web snippets returned by Google with a cut-off threshold σof 0.2. This set of parameters has been found in H.Yang and T.S. Chua to be most effective

Linear vs. Structured Query Formulation (3) Observations from Table 4: Web-based query formulation (Test 2) improves the baseline performance by 25.1%. This confirms: Using Web to extract highly correlated terms generally improves the QA performance. The use of lexical knowledge from WordNet without context constraint (Test 3) does not seem to be effective for QA, especially for the test w/o SCR: It tends to add too many terms out of context. The linear query formulation (Test 4) achieves an accuracy of after 5 SCR iterations. This gives an improvement of more than 7.3% over just using the Web (Test 2). The best performance is achieved (0.634) when carrying out the structured use of external knowledge (Test 5): An improvement of 7.8% over the linear query formulation (Test 4) with SCR. The impact is much more significant for tests without SCR with a corresponding improvement of 36.4%.

Conclusion QUALIFIER: A novel approach for Event-based QA Discover the info about a QA entity or different aspects of a QA event Structural analysis: extract semantic groups Semantic groups: expand query and achieve an answer accuracy of 63.4% The results demonstrated: this approach is feasible and effective for open domain QA