Automatic Question Generation from Queries Natural Language Computing, Microsoft Research Asia Chin-Yew LIN

Slides:



Advertisements
Similar presentations
Diversified Retrieval as Structured Prediction Redundancy, Diversity, and Interdependent Document Relevance (IDR ’09) SIGIR 2009 Workshop Yisong Yue Cornell.
Advertisements

1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
1 Web Search and Web Search Overlap: What the Deal? Amanda Spink Queensland University of Technology.
1 Question Answering in Biomedicine Student: Andreea Tutos Id: Supervisor: Diego Molla.
Evaluating Search Engine
Search Engines and Information Retrieval
Distributed Search over the Hidden Web Hierarchical Database Sampling and Selection Panagiotis G. Ipeirotis Luis Gravano Computer Science Department Columbia.
Modern Information Retrieval
Ryen W. White, Microsoft Research Jeff Huang, University of Washington.
INFO 624 Week 3 Retrieval System Evaluation
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
Investigation of Web Query Refinement via Topic Analysis and Learning with Personalization Department of Systems Engineering & Engineering Management The.
Online Learning for Web Query Generation: Finding Documents Matching a Minority Concept on the Web Rayid Ghani Accenture Technology Labs, USA Rosie Jones.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
Web Archive Information Retrieval Miguel Costa, Daniel Gomes (speaker) Portuguese Web Archive.
Searching the Web Dr. Frank McCown Intro to Web Science Harding University This work is licensed under a Creative Commons Attribution-NonCommercial- ShareAlike.
Personalization in Local Search Personalization of Content Ranking in the Context of Local Search Philip O’Brien, Xiao Luo, Tony Abou-Assaleh, Weizheng.
Cognitive Interviewing for Question Evaluation Kristen Miller, Ph.D. National Center for Health Statistics
Search Engines and Information Retrieval Chapter 1.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
Bringing Order to the Web: Automatically Categorizing Search Results Hao Chen, CS Division, UC Berkeley Susan Dumais, Microsoft Research ACM:CHI April.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
Searching the Web Dr. Frank McCown Intro to Web Science Harding University This work is licensed under Creative Commons Attribution-NonCommercial 3.0Attribution-NonCommercial.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
1 Can People Collaborate to Improve the relevance of Search Results? Florian Eiteljörge June 11, 2013Florian Eiteljörge.
11 Learning to Suggest Questions in Online Learning to Suggest Questions in Online Forums Tom Chao Zhou, Chin-Yew Lin, Irwin King Michael R.
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
Designing Ranking Systems for Consumer Reviews: The Economic Impact of Customer Sentiment in Electronic Markets Anindya Ghose Panagiotis Ipeirotis Stern.
Natural Language Based Reformulation Resource and Web Exploitation for Question Answering Ulf Hermjakob, Abdessamad Echihabi, Daniel Marcu University of.
INTRODUCING THE WEB INTELLIGENCE (WIT) GROUP Microsoft Research Asia.
Predicting Question Quality Bruce Croft and Stephen Cronen-Townsend University of Massachusetts Amherst.
Question Answering over Implicitly Structured Web Content
Modeling term relevancies in information retrieval using Graph Laplacian Kernels Shuguang Wang Joint work with Saeed Amizadeh and Milos Hauskrecht.
Analysis of Topic Dynamics in Web Search Xuehua Shen (University of Illinois) Susan Dumais (Microsoft Research) Eric Horvitz (Microsoft Research) WWW 2005.
1 Web-Page Summarization Using Clickthrough Data* JianTao Sun, Yuchang Lu Dept. of Computer Science TsingHua University Beijing , China Dou Shen,
WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
Qi Guo Emory University Ryen White, Susan Dumais, Jue Wang, Blake Anderson Microsoft Presented by Tetsuya Sakai, Microsoft Research.
Meet the web: First impressions How big is the web and how do you measure it? How many people use the web? How many use search engines? What is the shape.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
© 2004 Chris Staff CSAW’04 University of Malta of 15 Expanding Query Terms in Context Chris Staff and Robert Muscat Department of.
Retroactive Answering of Search Queries Beverly Yang Glen Jeh.
CoCQA : Co-Training Over Questions and Answers with an Application to Predicting Question Subjectivity Orientation Baoli Li, Yandong Liu, and Eugene Agichtein.
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
WERST – Methodology Group
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1 13/05/07 1/20 LIST – DTSI – Interfaces, Cognitics and Virtual Reality Unit The INFILE project: a crosslingual filtering systems evaluation campaign Romaric.
The Loquacious ( 愛說話 ) User: A Document-Independent Source of Terms for Query Expansion Diane Kelly et al. University of North Carolina at Chapel Hill.
11 A Classification-based Approach to Question Routing in Community Question Answering Tom Chao Zhou 1, Michael R. Lyu 1, Irwin King 1,2 1 The Chinese.
TO Each His Own: Personalized Content Selection Based on Text Comprehensibility Date: 2013/01/24 Author: Chenhao Tan, Evgeniy Gabrilovich, Bo Pang Source:
Information Retrieval Lecture 3 Introduction to Information Retrieval (Manning et al. 2007) Chapter 8 For the MSc Computer Science Programme Dell Zhang.
Navigation Aided Retrieval Shashank Pandit & Christopher Olston Carnegie Mellon & Yahoo.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
1 INFILE - INformation FILtering Evaluation Evaluation of adaptive filtering systems for business intelligence and technology watch Towards real use conditions.
Walid Magdy Gareth Jones
Evaluation of IR Systems
Personalizing Search on Shared Devices
IR Theory: Evaluation Methods
WorkShop on Community Question Answering on the Web
Web Mining Department of Computer Science and Engg.
Panagiotis G. Ipeirotis Luis Gravano
Category-Sensitive Question Routing in Community Question Answering
A Classification-based Approach to Question Routing in Community Question Answering Tom Chao Zhou 22, Feb, 2010 Department of Computer.
Presentation transcript:

Automatic Question Generation from Queries Natural Language Computing, Microsoft Research Asia Chin-Yew LIN

Generating Questions from Queries Where is the next Hannah Montana concert? Q2Q as a question generation shared task

Remember Ask Jeeves?  “How large is British Columbia?”

Live Search QnA (English)

Naver Knowledge iN (Korea) 5  Naver “Knowledge iN “Service  Opened at October 2002  70 Millions Knowledge iN DB are collected ( )  # of Users: 12 millions Upper level users (higher than Kosu): 6,648 (0.05%)  Distribution of knowledge Education, Learning: 17.78% Computer, Communication: 12.89% Entertainments, Arts: 11.42% Business, Economy: 11.42% Home, Life: 7.44%

Baidu Zhidao (China)  17,012,767 resolved questions in two years’ operation.  8,921,610 are knowledge related.  96.7% of questions are resolved.  10,000,000 daily visitors.  71,308 new questions per day.  3.14 answers per question.  ( 中国人搜索行为研究 /User Research Lab of Chinese Search)

Yahoo! Answers (Global; Marciniak)  Launched in December  20 million users in the U.S. (> 90 million worldwide).  33,557,437 resolved questions (US; April 2008).  ~70,000* new questions per day (US).  6.76* answers per question (US).

Question Taxonomy  ISI’s question answer typology ( Hovy et al & 2002 )  Results of analyzing over 20K online questions  140 different question types with examples  language/projects/webclopedia/Taxonomy/taxonomy_tople vel.html  Liu et al. (COLING 2008)’s cQA question taxonomy  Derived from Broder’s (SIGIR Forum 2002) web serach taxonomy  Results of analyzing 100 randomly sampled questions from top 4 Yahoo! Answers categories Entertainment & Music, Society & Culture, Health, and Computer & Internet

Main Task: Q2Q  Generate questions given a query  Query: “Hannah Montana concert”  Questions: “How do I get Hannah Montana concert tickets for a really good price?” “What should i wear to a hannah montana concert?” “How long is the Hannah Montana concert?” …  Subtasks  Predict user goals  Learn question templates  Normalize questions

Data Preparation  cQA archives  Live Search QnA  Yahoo! Answers  Ask.com  Other sources  Query logs  MSN/Live Search  Yahoo!  Ask.com  TREC and other sources  Possible process  Sample queries from search engine query logs  Ensure broad topic coverage  Find candidate questions from cQA archives given queries  Create mapped Q2Q corpus for training and testing

Intrinsic Evaluation  Given a query term  Generate a rank list of questions related to the query term  Open set – use pooling approach Pool all questions from participants Rate each question as relevant or not Compute recall/precision/F1 scores  Closed set – use test set data as gold standard  Metrics Diversity, interestingness, utility, and so on.

Extrinsic Evaluation  A straw man scenario  Task – online information seeking  Setup 1. A user select a topic (T) she is interested in. 2. Generate a set of N queries given T and a query log. 3. The user select a query (q) from the set. 4. Generate a set of M questions given q. 5. The user select the question (Q) that she has in mind. 6. If the user does not select any question, record it as not successful. 7. Send q to a search engine (S); get results X. 8. Send q, Q, and anything inferred from Q to S; get results Y. 9. Compare results X and Y using standard IR relevance metrics.

Summary  Task: Question generation from queries  Data:  Search engine query logs  cQA question answer archives  Question taxonomies  Evaluation:  Intrinsic – evaluate specific technology areas  Extrinsic – evaluate its effect on real world scenarios  Real data, real task, and real impact

Analyze cQA Questions (Liu et al. COLING 08) cQA QuestionNavigationalInformationalConstantDynamicOpinion Context- Dependent OpenTransactionalSocial