Multi-Perspective Question Answering Using the OpQA Corpus Veselin Stoyanov Claire Cardie Janyce Wiebe Cornell University University of Pittsburgh.

Slides:



Advertisements
Similar presentations
QA-LaSIE Components The question document and each candidate answer document pass through all nine components of the QA-LaSIE system in the order shown.
Advertisements

NYU ANLP-00 1 Automatic Discovery of Scenario-Level Patterns for Information Extraction Roman Yangarber Ralph Grishman Pasi Tapanainen Silja Huttunen.
LINGUISTICA GENERALE E COMPUTAZIONALE SENTIMENT ANALYSIS.
Playing the Telephone Game: Determining the Hierarchical Structure of Perspective and Speech Expressions Eric Breck and Claire Cardie Department of Computer.
Jan Wiebe University of Pittsburgh Claire Cardie Cornell University Ellen Riloff University of Utah Opinions in Question Answering.
Recognizing Contextual Polarity in Phrase-Level Sentiment Analysis Theresa Wilson Janyce Wiebe Paul Hoffmann University of Pittsburgh.
Sentiment Analysis An Overview of Concepts and Selected Techniques.
Annotating Topics of Opinions Veselin Stoyanov Claire Cardie.
A Framework for Automated Corpus Generation for Semantic Sentiment Analysis Amna Asmi and Tanko Ishaya, Member, IAENG Proceedings of the World Congress.
807 - TEXT ANALYTICS Massimo Poesio Lecture 4: Sentiment analysis (aka Opinion Mining)
Comparing Methods to Improve Information Extraction System using Subjectivity Analysis Prepared by: Heena Waghwani Guided by: Dr. M. B. Chandak.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Chapter 11 Beyond Bag of Words. Question Answering n Providing answers instead of ranked lists of documents n Older QA systems generated answers n Current.
Annotating Committed Belief Mona Diab, Lori Levin, Teruko Mitamura, Owen Rambow CMU/Columbia-CCLS
Annotating Expressions of Opinions and Emotions in Language Wiebe, Wilson, Cardie.
IR & Metadata. Metadata Didn’t we already talk about this? We discussed what metadata is and its types –Data about data –Descriptive metadata is external.
Event Extraction: Learning from Corpora Prepared by Ralph Grishman Based on research and slides by Roman Yangarber NYU.
1 Attributions and Private States Jan Wiebe (U. Pittsburgh) Theresa Wilson (U. Pittsburgh) Claire Cardie (Cornell U.)
Learning Subjective Nouns using Extraction Pattern Bootstrapping Ellen Riloff, Janyce Wiebe, Theresa Wilson Presenter: Gabriel Nicolae.
Combining Low-Level and Summary Representations of Opinions for Multi- Perspective Question Answering Claire Cardie, Janyce Wiebe, Theresa Wilson, Diane.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Learning Subjective Adjectives from Corpora Janyce M. Wiebe Presenter: Gabriel Nicolae.
PNC 2011: Pacific Neighborhood Consortium S-Sense: An Opinion Mining Tool for Market Intelligence Choochart Haruechaiyasak and Alisa Kongthon Speech and.
Mining and Summarizing Customer Reviews
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Automatic Extraction of Opinion Propositions and their Holders Steven Bethard, Hong Yu, Ashley Thornton, Vasileios Hatzivassiloglou and Dan Jurafsky Department.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Carmen Banea, Rada Mihalcea University of North Texas A Bootstrapping Method for Building Subjectivity Lexicons for Languages.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
1 A Unified Relevance Model for Opinion Retrieval (CIKM 09’) Xuanjing Huang, W. Bruce Croft Date: 2010/02/08 Speaker: Yu-Wen, Hsu.
On the Issue of Combining Anaphoricity Determination and Antecedent Identification in Anaphora Resolution Ryu Iida, Kentaro Inui, Yuji Matsumoto Nara Institute.
Automatic Detection of Tags for Political Blogs Khairun-nisa Hassanali Vasileios Hatzivassiloglou The University.
WebMining Web Mining By- Pawan Singh Piyush Arora Pooja Mansharamani Pramod Singh Praveen Kumar 1.
RESEARCH POSTER PRESENTATION DESIGN © © 2012 PosterPresentations.com 2117 Fourth Street, Unit C Berkeley CA 94710
1 Determining the Hierarchical Structure of Perspective and Speech Expressions Eric Breck and Claire Cardie Cornell University Department of Computer Science.
Exploiting Subjectivity Classification to Improve Information Extraction Ellen Riloff University of Utah Janyce Wiebe University of Pittsburgh William.
A Language Independent Method for Question Classification COLING 2004.
Experiments of Opinion Analysis On MPQA and NTCIR-6 Yaoyong Li, Kalina Bontcheva, Hamish Cunningham Department of Computer Science University of Sheffield.
A Bootstrapping Method for Building Subjectivity Lexicons for Languages with Scarce Resources Author: Carmen Banea, Rada Mihalcea, Janyce Wiebe Source:
Opinion Mining of Customer Feedback Data on the Web Presented By Dongjoo Lee, Intelligent Databases Systems Lab. 1 Dongjoo Lee School of Computer Science.
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
NTCIR-5, An Overview of Opinionated Tasks and Corpus Preparation Hsin-Hsi Chen Department of Computer Science and Information Engineering National.
Bootstrapping for Text Learning Tasks Ramya Nagarajan AIML Seminar March 6, 2001.
1 Multi-Perspective Question Answering Using the OpQA Corpus (HLT/EMNLP 2005) Veselin Stoyanov Claire Cardie Janyce Wiebe Cornell University University.
Summarization Focusing on Polarity or Opinion Fragments in Blogs Yohei Seki Toyohashi University of Technology Visiting Scholar at Columbia University.
What you have learned and how you can use it : Grammars and Lexicons Parts I-III.
1 Toward Opinion Summarization: Linking the Sources Veselin Stoyanov and Claire Cardie Department of Computer Science Cornell University Ithaca, NY 14850,
Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.
Multilingual Opinion Holder Identification Using Author and Authority Viewpoints Yohei Seki, Noriko Kando,Masaki Aono Toyohashi University of Technology.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Evaluating an Opinion Annotation Scheme Using a New Multi- perspective Question and Answer Corpus (AAAI 2004 Spring) Veselin Stoyanov Claire Cardie Diane.
Have we had Hard Times or Cosy Times? A Discourse Analysis of Opinions Expressed over Socio-political Events in News Editorials Bal Krishna Bal Information.
Learning Subjective Nouns using Extraction Pattern Bootstrapping Ellen Riloff School of Computing University of Utah Janyce Wiebe, Theresa Wilson Computing.
Creating Subjective and Objective Sentence Classifiers from Unannotated Texts Ellen Riloff University of Utah (Joint work with Janyce Wiebe at the University.
7/2003EMNLP031 Learning Extraction Patterns for Subjective Expressions Ellen Riloff Janyce Wiebe University of Utah University of Pittsburgh.
Learning Extraction Patterns for Subjective Expressions 2007/10/09 DataMining Lab 안민영.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Finding strong and weak opinion clauses Theresa Wilson, Janyce Wiebe, Rebecca Hwa University of Pittsburgh Just how mad are you? AAAI-2004.
Integrating linguistic knowledge in passage retrieval for question answering J¨org Tiedemann Alfa Informatica, University of Groningen HLT/EMNLP 2005.
AQUAINT Mid-Year PI Meeting – June 2002 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Identifying Expressions of Opinion in Context Eric Breck and Yejin Choi and Claire Cardie IJCAI 2007.
CSC 594 Topics in AI – Natural Language Processing
Paradigms, Corpora, and Tools in Discourse and Dialog Research
Parts of Speech Mr. White English I.
Automatic Detection of Causal Relations for Question Answering
Extracting Why Text Segment from Web Based on Grammar-gram
Presentation transcript:

Multi-Perspective Question Answering Using the OpQA Corpus Veselin Stoyanov Claire Cardie Janyce Wiebe Cornell University University of Pittsburgh

10/08/05HLT/EMNLP Multi-Perspective Question Answering Fact-based question answering (QA): When is the first day of spring? Do Lipton employees take coffee breaks? Vs Multi-perspective question answering (MPQA). How does the US regard the terrorist attacks in Iraq? Is Derek Jeter a bum?

10/08/05HLT/EMNLP Properties of Opinion vs. Fact answers –OpQA corpus –Traditional fact-based QA systems –Different properties of opinion questions Using fine-grained opinion information for MPQA –Annotation framework and automatic classifiers –QA experiments Talk Outline

10/08/05HLT/EMNLP Opinion Question & Answer (OpQA) Corpus 98 documents manually tagged for opinions (from the NRRC MPQA corpus [Wilson and Wiebe 2003]) 30 questions –15 fact –15 opinion [Stoyanov, Cardie, Litman, and Wiebe 2004]

10/08/05HLT/EMNLP OpQA corpus: Answer Annotations Two annotators Include every text segment contributing to an answer –Partial answers: When was the Kyoto protocol ratified? –… before May 2003 … Are the Japanese unanimous in their support of Koizumi? –… most Japanese support their prime minister … Minimum spans

10/08/05HLT/EMNLP Traditional Fact-based QA systems IR subsystem 1.Frag Frag Frag Frag Frag Frag Frag Frag 213 Linguistic filters Guesses: 1.Frag Frag ….. Documents (document fragments) Questions Syntactic filters Semantic filters

10/08/05HLT/EMNLP Characteristics of Opinion vs. Fact Answers Answer length –Syntactic and semantic class –Additional processing difficulties Partial answers –Answer generator Number of answers Length (tokens)Number of partials Fact (9.68%) Opinion (37.11%)

10/08/05HLT/EMNLP Fine-grained Opinion Information for MPQA Recent interest in the area of automatic opinion information extraction. –E.g. [Bethard, Yu, Thornton, Hatzivassiloglou, and Jurafsky 2004], [Pang and Lee 2004], [Riloff and Wiebe 2003], [Wiebe and Riloff 2005], [Wilson, Wiebe, and Hwa 2004], [Yu and Hatzivassiloglou 2003] In our evaluation: –Opinion annotation framework –Sentence-level automatic opinion classifiers –Subjectivity filters –Source filter

10/08/05HLT/EMNLP Described in [Wiebe, Wilson, and Cardie 2002] Accounts for both: –Explicitly stated opinions Joe believes that Sue dislikes the Red Sox. –Indirectly expressed opinions The aim of the report is to tarnish China’s image. Attributes include strength and source. Manual sentence-level classification –sentence subjective if it contains one or more opinions of strength >= medium Opinion Annotation Framework Described in [Wiebe, Wilson, and Cardie 2002] Accounts for both: –Explicitly stated opinions Joe believes that Sue dislikes the Red Sox. –Indirectly expressed opinions The aim of the report is to tarnish China’s image. Attributes include strength and source. Manual sentence-level classification –sentence subjective if it contains one or more opinions of strength >= medium

10/08/05HLT/EMNLP Automatic Opinion Classifiers Two sentence-level opinion classifiers from Wiebe and Riloff [2005] used for evaluation Both classifiers use unannotated data –Rulebased: Extraction patterns bootstrapped using word lists –NaiveBayes: Trained on data obtained from Rulebased PrecisionRecallF Rulebased NaiveBayes

10/08/05HLT/EMNLP Subjectivity Filters IR subsystem 1.Sent Sent Sent Sent Sent 324(o) 2.Sent 111(f) 3.Sent 431(f) 4.Sent 213(o) Subjectivity filters Document Sentences Opinion Questions Guesses 1.Sent Sent ….. Manual Rulebased NaiveBayes Baseline

10/08/05HLT/EMNLP Subjectivity Filters Cont’d Look for the rank of the first guess containing an answer Compute: 1.Sent Sent Sent 007 (ans) 4.Sent Sent 211 (ans) 6. … –Mean Reciprocal Rank (MRR) across the top 5 answers MRR = mean all_questions (1/Rank_of_first_answer) –Mean Rank of the First Answer MRFA = mean all_questions (Rank_of_first_answer)

10/08/05HLT/EMNLP Subjectivity Filters Results

10/08/05HLT/EMNLP Source Filter Manually identify the sources in the opinion questions Does France approve of the war in Iraq? Retains only sentences that contain opinions with sources matching sources in the question France has voiced some concerns with the situation.

10/08/05HLT/EMNLP Source Filter Results Performs well on the hardest questions in the corpus All questions answered within the first 25 sentences with one exception. MRRMRFA Baseline Source

10/08/05HLT/EMNLP Summary Properties of opinion vs. fact answers –Traditional architectures unlikely to be effective Use of fine-grained opinion information for MPQA –MPQA can benefit from fine-grained perspective information

10/08/05HLT/EMNLP Future Work Create summaries of all opinions in a document using fine-grained opinion information Methods used will be directly applicable to MPQA

10/08/05HLT/EMNLP Thank you. Questions?

10/08/05HLT/EMNLP Did something surprising happen when Chavez regained power in Venezuela after he was removed by a coup? What did South Africa want Mugabe to do after the 2002 elections? What’s Mugabe’s opinion about the West’s attitude and actions towards the 2002 Zimbabwe election?

10/08/05HLT/EMNLP Characteristics of Fact vs. Opinion Answers Cont’d Syntactic Constituent of the answers FactOpinion Answers in best matching category 31%16% Syntactic type of best match Verb Phrase 02 Noun Phrase 94 Clause26

10/08/05HLT/EMNLP All improvement significant using Wilcoxon Matched-Pairs Signed-Ranks Test (p<=0.05) except for source filter (p=0.81)