Question Answering over Implicitly Structured Web Content

Slides:



Advertisements
Similar presentations
A Support Vector Method for Optimizing Average Precision
Advertisements

Improved TF-IDF Ranker
Optimizing search engines using clickthrough data
Jean-Eudes Ranvier 17/05/2015Planet Data - Madrid Trustworthiness assessment (on web pages) Task 3.3.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan Susan T.Dumains Eric Horvitz MIT,CSAILMicrosoft Researcher Microsoft.
Learning to Rank: New Techniques and Applications Martin Szummer Microsoft Research Cambridge, UK.
 Copyright 2005 Digital Enterprise Research Institute. All rights reserved. 1 The Architecture of a Large-Scale Web Search and Query Engine.
1 The Four Dimensions of Search Engine Quality Jan Pedersen Chief Scientist, Yahoo! Search 19 September 2005.
Time-dependent Similarity Measure of Queries Using Historical Click- through Data Qiankun Zhao*, Steven C. H. Hoi*, Tie-Yan Liu, et al. Presented by: Tie-Yan.
Presented by Li-Tal Mashiach Learning to Rank: A Machine Learning Approach to Static Ranking Algorithms for Large Data Sets Student Symposium.
Retrieval Evaluation. Brief Review Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Shared Ontology for Knowledge Management Atanas Kiryakov, Borislav Popov, Ilian Kitchukov, and Krasimir Angelov Meher Shaikh.
Sigir’99 Inside Internet Search Engines: Search Jan Pedersen and William Chang.
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
Adapting Deep RankNet for Personalized Search
Large-Scale Content-Based Image Retrieval Project Presentation CMPT 880: Large Scale Multimedia Systems and Cloud Computing Under supervision of Dr. Mohamed.
Quality-aware Collaborative Question Answering: Methods and Evaluation Maggy Anastasia Suryanto, Ee-Peng Lim Singapore Management University Aixin Sun.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Search Engines and Information Retrieval Chapter 1.
Hang Cui et al. NUS at TREC-13 QA Main Task 1/20 National University of Singapore at the TREC- 13 Question Answering Main Task Hang Cui Keya Li Renxu Sun.
1 Context-Aware Search Personalization with Concept Preference CIKM’11 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
An Analysis of Assessor Behavior in Crowdsourced Preference Judgments Dongqing Zhu and Ben Carterette University of Delaware.
Master Thesis Defense Jan Fiedler 04/17/98
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
PAUL ALEXANDRU CHIRITA STEFANIA COSTACHE SIEGFRIED HANDSCHUH WOLFGANG NEJDL 1* L3S RESEARCH CENTER 2* NATIONAL UNIVERSITY OF IRELAND PROCEEDINGS OF THE.
1 Mining User Behavior Mining User Behavior Eugene Agichtein Mathematics & Computer Science Emory University.
Michael Cafarella Alon HalevyNodira Khoussainova University of Washington Google, incUniversity of Washington Data Integration for Relational Web.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
Personalized Search Cheng Cheng (cc2999) Department of Computer Science Columbia University A Large Scale Evaluation and Analysis of Personalized Search.
« Pruning Policies for Two-Tiered Inverted Index with Correctness Guarantee » Proceedings of the 30th annual international ACM SIGIR, Amsterdam 2007) A.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Natural Language Based Reformulation Resource and Web Exploitation for Question Answering Ulf Hermjakob, Abdessamad Echihabi, Daniel Marcu University of.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Personalized Search Xiao Liu
Curtis Spencer Ezra Burgoyne An Internet Forum Index.
LOGO Finding High-Quality Content in Social Media Eugene Agichtein, Carlos Castillo, Debora Donato, Aristides Gionis and Gilad Mishne (WSDM 2008) Advisor.
WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries.
Google’s Deep-Web Crawl By Jayant Madhavan, David Ko, Lucja Kot, Vignesh Ganapathy, Alex Rasmussen, and Alon Halevy August 30, 2008 Speaker : Sahana Chiwane.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
Next Generation Search Engines Ehsun Daroodi 1 Feb, 2003.
IR, IE and QA over Social Media Social media (blogs, community QA, news aggregators)  Complementary to “traditional” news sources (Rathergate)  Grow.
Querying Web Data – The WebQA Approach Author: Sunny K.S.Lam and M.Tamer Özsu CSI5311 Presentation Dongmei Jiang and Zhiping Duan.
Web Information Retrieval Prof. Alessandro Agostini 1 Context in Web Search Steve Lawrence Speaker: Antonella Delmestri IEEE Data Engineering Bulletin.
Exploiting Relevance Feedback in Knowledge Graph Search
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Finding the Right Facts in the Crowd: Factoid Question Answering over Social Media J. Bian, Y. Liu, E. Agichtein, and H. Zha ACM WWW, 2008.
WIRED Future Quick review of Everything What I do when searching, seeking and retrieving Questions? Projects and Courses in the Fall Course Evaluation.
The Loquacious ( 愛說話 ) User: A Document-Independent Source of Terms for Query Expansion Diane Kelly et al. University of North Carolina at Chapel Hill.
Improving QA Accuracy by Question Inversion John Prager, Pablo Duboue, Jennifer Chu-Carroll Presentation by Sam Cunningham and Martin Wintz.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Information Retrieval Lecture 3 Introduction to Information Retrieval (Manning et al. 2007) Chapter 8 For the MSc Computer Science Programme Dell Zhang.
1 Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan, MIT Susan T. Dumais, Microsoft Eric Horvitz, Microsoft SIGIR 2005.
Image Retrieval and Ranking using L.S.I and Cross View Learning Sumit Kumar Vivek Gupta
SEARCH AND CONTEXT Susan Dumais, Microsoft Research INFO 320.
Collection Fusion in Carrot2
Search User Behavior: Expanding The Web Search Frontier
A research literature search engine with abbreviation recognition
The Four Dimensions of Search Engine Quality
Eugene Agichtein Mathematics & Computer Science Emory University
Evidence from Behavior
Combining Keyword and Semantic Search for Best Effort Information Retrieval  Andrew Zitzelberger 1.
Topic: Semantic Text Mining
Presentation transcript:

Question Answering over Implicitly Structured Web Content Eugene Agichtein* Emory University Chris Burges Microsoft Research Eric Brill Microsoft Research * Research done while at Microsoft Research

Questions are Problematic for Web Search What was the name of president Fillmore’s cat? Who invented crocs? … Agichtein et al., WI 2007

Web search: What was the name of president Fillmore’s cat? Agichtein et al., WI 2007

Web Question Answering Why are questions problematic for web search engines? Search engines treat questions as keyword queries, ignoring the semantic relationships between words, and the explicitly stated information need Poor performance for long (> 5 terms) queries Problem exacerbated when common keywords are included Agichtein et al., WI 2007

… and millions more of other tables and lists … Agichtein et al., WI 2007

Implicitly Structured Web Content HTML Tables, Lists Product descriptions Example: Lists of favorite things, “top 10” lists, etc. HTML Syntax (sometimes) reflects semantics Authors imply semantic relationships, entity types by grouping Can infer information about ambiguous entities from others in the same column Millions of HTML tables, lists on the “surface” web alone No common schema Keyword queries: primary access method. How to exploit this structured content for good (e.g., for Question Answering) at web scale? Agichtein et al., WI 2007

Related Work Web Question Answering Web-scale Information Extraction AskMSR (TREC 2001)  Aranea (TREC 2003) Mulder (WWW 2001) A No-Frills Architecture for Lightweight Answer Retrieval (WWW 2007) Web-scale Information Extraction QXtract (ICDE 2003): learn keyword queries to retrieve content KnowItAll (WWW 2004): minimal supervision, larger scale TextRunner (IJCAI 2007): single pass scan, disambiguate at query time Towards Domain-Independent Information Extraction from Web Tables (WWW 2007) Agichtein et al., WI 2007

Our System TQA: Overview Index all promising HTML tables Translate a question into select/project query Select table rows, project candidate answers Rank candidate answers Return top K answers Agichtein et al., WI 2007

TableQA: Indexing Crawl the Web Identify “promising” tables (heuristic, could be improved) Extract metadata for each table Context Document content Document metadata Index extracted metadata Agichtein et al., WI 2007

Table Metadata Combines information about the source document, and table context Agichtein et al., WI 2007

TQA Question Processing Agichtein et al., WI 2007

Table QA: Querying Overview Agichtein et al., WI 2007

Features for Ranking Candidate Answers Explain: Frequency, freq decayed, idf, type sim, page rank , column distance, tightness, overlap Agichtein et al., WI 2007

Ranking Answer Candidates Frequency-based (AskMSR): Heuristic weight assignment (AskMSR improved) Neither is robust or general Add animation – show one at a time Agichtein et al., WI 2007

Ranking Answer Candidates (cont) Solution: machine learning-based ranking Naïve Bayes: Score(answer) = RankNet (Burges et al. 2005): scalable Neural Net implementation: Optimized for ranking – predicting an ordering of items, not scores for each Trains on pairs (where first point is to be ranked higher or equal to second) Uses cross entropy cost and gradient descent to set weights Agichtein et al., WI 2007

Some Implementation Details Lucene, distributed indices (20M tables per index) NLP Tools: MS internal Named Entity tagger (many free ones exist) Porter Stemmer Relatively light-weight architecture: Client (question processing): desktop machine Table index server: dual-processor, 8 Gb RAM, WinNT Agichtein et al., WI 2007

Experimental Setup Queries: TREC QA 2002, 2003 questions Corpus: 100M web pages (a “random” subset of an MSN Search crawl, from 2005) Evaluation: TREC QA factoid patterns “Minimal” regular expressions to match only right answers Not comprehensive (based on judgement pool) Agichtein et al., WI 2007

Evaluation Metrics MRR (mean reciprocal rank): Recall @ K: MRR @ K = , averaged over all questions Recall @ K: The fraction of the questions for which a system returned a correct answer ranked at or above K. Agichtein et al., WI 2007

Results (1): Accuracy vs. Corpus Size Agichtein et al., WI 2007

Results (2): Comparing Ranking Methods If output consumed by another system, large K ok Agichtein et al., WI 2007

Results (3): Accuracy on Hard Questions TQA can retrieve answer in top 100 when best QA system not able to return any answer Agichtein et al., WI 2007

Result Summary Requires indexing more than 150M tables before respectable accuracy achieved Performance was around median on TREC 2002, 2003 benchmarks Can be helpful for questions difficult for traditional QA systems Agichtein et al., WI 2007

Promising Directions for Future Work Craw-time: aggressive pruning/classification Index-time: Integration of related tables Query-time: taxonomies integration/hypernimy User behavior modeling Past clickthrough to rerank candidate tables, answers Query reformulation Agichtein et al., WI 2007

Conclusions Implicitly structured web content can be useful for web question answering We demonstrated scalability of a lightweight table-based web QA approach Much room for improvement, future research Agichtein et al., WI 2007

Thank you! Questions? E-mail: eugene@mathcs.emory.edu Plug: User Interactions for Web Question Answering: http://www.mathcs.emory.edu/~eugene/uqa/ E. Agichtein, E. Brill, S. Dumais, Mining user behavior to improve web search ranking, SIGIR 2006 E. Agichtein, User Behavior Mining and Information Extraction: Towards closing the gap, IEEE Data Engineering Bulletin, Dec. 2006 E. Agichtein, C. Castillo, D. Donato, A. Gionis, and G. Mishne, Finding High Quality Content in Social Media with applications to Community-based Question Answering, to appear WSDM 2008 Agichtein et al., WI 2007