INTERESTING NUGGETS AND THEIR IMPACT ON DEFINITIONAL QUESTION ANSWERING Kian-Wei Kor, Tat-Seng Chua Department of Computer Science School of Computing.

Slides:



Advertisements
Similar presentations
ThemeInformation Extraction for World Wide Web PaperUnsupervised Learning of Soft Patterns for Generating Definitions from Online News Author Cui, H.,
Advertisements

Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
1 Entity Ranking Using Wikipedia as a Pivot (CIKM 10’) Rianne Kaptein, Pavel Serdyukov, Arjen de Vries, Jaap Kamps 2010/12/14 Yu-wen,Hsu.
IR Challenges and Language Modeling. IR Achievements Search engines  Meta-search  Cross-lingual search  Factoid question answering  Filtering Statistical.
Re-ranking Documents Segments To Improve Access To Relevant Content in Information Retrieval Gary Madden Applied Computational Linguistics Dublin City.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 30, (2014) BERLIN CHEN, YI-WEN CHEN, KUAN-YU CHEN, HSIN-MIN WANG2 AND KUEN-TYNG YU Department of Computer.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
CSC 9010 Spring Paula Matuszek A Brief Overview of Watson.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
Hang Cui et al. NUS at TREC-13 QA Main Task 1/20 National University of Singapore at the TREC- 13 Question Answering Main Task Hang Cui Keya Li Renxu Sun.
TREC 2009 Review Lanbo Zhang. 7 tracks Web track Relevance Feedback track (RF) Entity track Blog track Legal track Million Query track (MQ) Chemical IR.
Leveraging Conceptual Lexicon : Query Disambiguation using Proximity Information for Patent Retrieval Date : 2013/10/30 Author : Parvaz Mahdabi, Shima.
1 A study on automatically extracted keywords in text categorization Authors:Anette Hulth and Be´ata B. Megyesi From:ACL 2006 Reporter: 陳永祥 Date:2007/10/16.
1 Wikification CSE 6339 (Section 002) Abhijit Tendulkar.
Estimating Importance Features for Fact Mining (With a Case Study in Biography Mining) Sisay Fissaha Adafre School of Computing Dublin City University.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
1 Retrieval and Feedback Models for Blog Feed Search SIGIR 2008 Advisor : Dr. Koh Jia-Ling Speaker : Chou-Bin Fan Date :
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
Linking Wikipedia to the Web Antonio Flores Bernal Department of Computer Sciencies San Pablo Catholic University 2010.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
1 A Unified Relevance Model for Opinion Retrieval (CIKM 09’) Xuanjing Huang, W. Bruce Croft Date: 2010/02/08 Speaker: Yu-Wen, Hsu.
Estimating Topical Context by Diverging from External Resources SIGIR’13, July 28–August 1, 2013, Dublin, Ireland. Presenter: SHIH, KAI WUN Romain Deveaud.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.
Péter Schönhofen – Ad Hoc Hungarian → English – CLEF Workshop 20 Sep 2007 Performing Cross-Language Retrieval with Wikipedia Participation report for Ad.
Retrieval Models for Question and Answer Archives Xiaobing Xue, Jiwoon Jeon, W. Bruce Croft Computer Science Department University of Massachusetts, Google,
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Detecting Dominant Locations from Search Queries Lee Wang, Chuang Wang, Xing Xie, Josh Forman, Yansheng Lu, Wei-Ying Ma, Ying Li SIGIR 2005.
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
Improving Web Search Results Using Affinity Graph Benyu Zhang, Hua Li, Yi Liu, Lei Ji, Wensi Xi, Weiguo Fan, Zheng Chen, Wei-Ying Ma Microsoft Research.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
Automatic Set Instance Extraction using the Web Richard C. Wang and William W. Cohen Language Technologies Institute Carnegie Mellon University Pittsburgh,
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
21/11/20151Gianluca Demartini Ranking Clusters for Web Search Gianluca Demartini Paul–Alexandru Chirita Ingo Brunkhorst Wolfgang Nejdl L3S Info Lunch Hannover,
Intelligent Database Systems Lab N.Y.U.S.T. I. M. An information-pattern-based approach to novelty detection Presenter : Lin, Shu-Han Authors : Xiaoyan.
Chapter 23: Probabilistic Language Models April 13, 2004.
LOGO 1 Corroborate and Learn Facts from the Web Advisor : Dr. Koh Jia-Ling Speaker : Tu Yi-Lang Date : Shubin Zhao, Jonathan Betz (KDD '07 )
Authors: Marius Pasca and Benjamin Van Durme Presented by Bonan Min Weakly-Supervised Acquisition of Open- Domain Classes and Class Attributes from Web.
Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.
Carnegie Mellon Novelty and Redundancy Detection in Adaptive Filtering Yi Zhang, Jamie Callan, Thomas Minka Carnegie Mellon University {yiz, callan,
A Word Clustering Approach for Language Model-based Sentence Retrieval in Question Answering Systems Saeedeh Momtazi, Dietrich Klakow University of Saarland,Germany.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
A Classification-based Approach to Question Answering in Discussion Boards Liangjie Hong, Brian D. Davison Lehigh University (SIGIR ’ 09) Speaker: Cho,
AQUAINT IBM PIQUANT ARDACYCORP Subcontractor: IBM Question Answering Update piQuAnt ARDA/AQUAINT December 2002 Workshop This work was supported in part.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Improving the performance of personal name disambiguation.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Query Suggestions in the Absence of Query Logs Sumit Bhatia, Debapriyo Majumdar,Prasenjit Mitra SIGIR’11, July 24–28, 2011, Beijing, China.
Extracting and Ranking Product Features in Opinion Documents Lei Zhang #, Bing Liu #, Suk Hwan Lim *, Eamonn O’Brien-Strain * # University of Illinois.
Divided Pretreatment to Targets and Intentions for Query Recommendation Reporter: Yangyang Kang /23.
The Loquacious ( 愛說話 ) User: A Document-Independent Source of Terms for Query Expansion Diane Kelly et al. University of North Carolina at Chapel Hill.
CONTEXTUAL SEARCH AND NAME DISAMBIGUATION IN USING GRAPHS EINAT MINKOV, WILLIAM W. COHEN, ANDREW Y. NG SIGIR’06 Date: 2008/7/17 Advisor: Dr. Koh,
A Generation Model to Unify Topic Relevance and Lexicon-based Sentiment for Opinion Retrieval Min Zhang, Xinyao Ye Tsinghua University SIGIR
Personalizing Web Search Jaime Teevan, MIT with Susan T. Dumais and Eric Horvitz, MSR.
2016/3/11 Exploiting Internal and External Semantics for the Clustering of Short Texts Using World Knowledge Xia Hu, Nan Sun, Chao Zhang, Tat-Seng Chu.
ENHANCING CLUSTER LABELING USING WIKIPEDIA David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab SIGIR’09.
Toward Entity Retrieval over Structured and Text Data Mayssam Sayyadian, Azadeh Shakery, AnHai Doan, ChengXiang Zhai Department of Computer Science University.
UIC at TREC 2006: Blog Track Wei Zhang Clement Yu Department of Computer Science University of Illinois at Chicago.
AQUAINT Mid-Year PI Meeting – June 2002 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Compact Query Term Selection Using Topically Related Text
Presentation transcript:

INTERESTING NUGGETS AND THEIR IMPACT ON DEFINITIONAL QUESTION ANSWERING Kian-Wei Kor, Tat-Seng Chua Department of Computer Science School of Computing National University of Singapore (SIGIR’07) Speaker : Yi-Ling Tai Date : 2009/07/28

OUTLINE Introduction Definitional Question Answer Topic nuggets Human Interest Model Web Resources Implementation Initial Experiments Refinements Weighting Interesting terms Selecting Web resources Unifying Conclusion

DEFINITIONAL QUESTION ANSWERING TREC Question Answering Track main task in Given Topic X, “What is X?” 、 ”Who is X?” The definitional QA system is to search through a corpus and return a set of answers best describes the topic. Each answer should be a unique topic-specific nugget that makes up one facet in the definition of topic.

T HE TWO ASPECTS OF TOPIC NUGGETS Informative nuggets A sentence fragment that describes some factual information Including topic properties, relationships with other related entity, or events that happened to the topic Interesting nuggets Trivia-like quality Typically, these can pique a human reader’s interesting “What is X famous for?”, “What is extraordinary about X?”

THE TWO ASPECTS OF TOPIC NUGGETS TREC 2005 topic of “George Foreman” Informative nuggets Was graduate of Job Corps. Became oldest world champion in boxing history. Interesting nuggets Has lent his name to line of food preparation products. Waved American flag after winning 1968 Olympics championship. Returned to boxing after 10 yr hiatus.

HUMAN INTEREST MODEL To indentify sentences that a human reader would find interesting is a tall order. Assume that most sentences within web documents will contain interesting facets about the topic. The articles are written by humans, for human readers and thus contain the critical human world knowledge.

WEB RESOURCES To build the “Interest Corpus” by collecting articles from the following external resources Wikipedia We use a snapshot of Wikipedia taken in March 2006 Include the most relevant article NewsLibrary For each topic, we download the 50 most relevant articles and include the title and first paragraph Google Snippets For each topic, extract the top 100 snippets.

WEB RESOURCES Some are more specific in nature, and we do not always get any single relevant document Biography.com Bartleby.com s9.com Google Definitions WordNet

MULTIPLE INTERESTING CENTROIDS Relevance-based approaches are focused on identifying highly relevant sentences The use of only a single collection of centroid words has over-emphasized topic relevance. Perform a pairwise sentence comparison between Interest Corpus and candidate sentences retrieved from AQUAINT corpus. An answer can only be highly ranked if it is strongly similar to a sentence in the Interest Corpus, and is also strongly relevant to the topic.

IMPLEMENTATION

The AQUAINT Retrieval module Given a set of words describing the topic, the AQUAINT Retrieval module does query expansion using Google and searches an index of AQUAINT documents to retrieve the 800 most relevant documents Web Retrieval module searches the online resources to populate the Interest Corpus.

IMPLEMENTATION HIM Ranker First build the unigram language model, I, from the collected web documents. This language model will be used to weight the importance of terms within sentences. Segment all 800 retrieved documents into individual sentences. Perform a pairwise similarity comparison between a candidate sentence and sentences in our external documents using a weighted-term edit distance algorithm. Select the top 12 highest ranked and non redundant sentences as definitional answers.

INITIAL EXPERIMENTS Compare with soft-pattern bigram model To ensure comparable results identical input data same web articles retrieved by our Web Retrieval module rank the same set of candidate sentences retrieved by our AQUAINT Retrieval module.

INITIAL EXPERIMENTS TREC provides a list of vital and okay nuggets for each question topic. Every question is scored on nugget recall (NR) and nugget precision (NP) and a single final score is computed using F-Measure.

INITIAL EXPERIMENTS How well the Human Interest Model performs for different types of topics Manually divided the TREC 2005 topics into four broad categories PERSON ORGANIZATION THING EVENT

REFINEMENTS Weighting Interesting Terms Interesting nuggets often has a trivia-like quality we hypothesize that interesting nuggets are likely to occur rarely in a text corpora. Three different term weighting schemes that can provide more weight to low-frequency terms. TFIDF Kullback-Leiber divergence A : the AQUAINT corpus as a unigram language model of general English I : Interest Corpus as a unigram language model consisting of topic specific terms and general English terms Jensen-Shannon divergence

REFINEMENTS

TFIDF performed the worst, The reason is that most terms only appear once within each sentence, resulting in a term frequency of 1. Both KL and JS divergence, we observed that high frequency relevant terms still dominate the top of the weighted term list.

SELECTING WEB RESOURCES The quality of web resources included in the Interest Corpus may have a direct impact on the results. split web resources into four major groups N - News: Title and first paragraph of the top 50 most relevant articles found in NewsLibrary W - Wikipedia: Text from the most relevant article found in Wikipedia. S - Snippets: Snippets extracted from the top 100 most relevant links after querying Google. M-Miscellaneous sources

SELECTING WEB RESOURCES

UNIFYING INFORMATIVENESS WITH INTERESTINGNESS From the perspective of a human reader, both informative and interesting nuggets are useful and definitional. A good definitional question answering system should provide the reader with a combined mixture of both nugget types. We had initially hoped to unify The Soft Pattern Bigram Model and The Human Interest Model. However, none of the ensemble learning methods we attempted could outperform our Human Interest Model.

UNIFYING INFORMATIVENESS WITH INTERESTINGNESS The two models are disagreeing on which sentences are definitional. In the top 10 sentences from both systems, only 4.4% of these sentences appeared in both answer sets. To verify if both systems are selecting the same answer nuggets randomly select a subset of 10 topics from the TREC 2005 question set manually identified correct answer nuggets from both systems the nugget agreement rate between both systems was 16.6%. Definitions are indeed made up of a mixture of informative and interesting nuggets

UNIFYING INFORMATIVENESS WITH INTERESTINGNESS The best approach we found for combining both answer sets is to merge and rerank both answer sets with boosting agreements. When both systems agree agree that the sentence is definitional, the sentence’s score is boosted Using the approach described here, we achieve a F3 score of

CONCLUSION A novel perspective for answering definitional questions through the identification of interesting nuggets. Using a combination of difference external corpus, we can build a definitional question answering module. The inherent differences between both types of nuggets seemingly caused by the low agreement rates between both models have made combination a difficult task.