Automatic sense prediction for implicit discourse relations in text Emily Pitler, Annie Louis, Ani Nenkova University of Pennsylvania ACL 2009.

Slides:



Advertisements
Similar presentations
Using Syntax to Disambiguate Explicit Discourse Connectives in Text Source: ACL-IJCNLP 2009 Author: Emily Pitler and Ani Nenkova Reporter: Yong-Xiang Chen.
Advertisements

Emily Pitler, Annie Louis, Ani Nenkova University of Pennsylvania.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
Text Categorization Moshe Koppel Lecture 1: Introduction Slides based on Manning, Raghavan and Schutze and odds and ends from here and there.
Progress update Lin Ziheng. System overview 2 Components – Connective classifier Features from Pitler and Nenkova (2009): – Connective: because – Self.
Farag Saad i-KNOW 2014 Graz- Austria,
Automatically Evaluating Text Coherence Using Discourse Relations Ziheng Lin, Hwee Tou Ng and Min-Yen Kan Department of Computer Science National University.
Using Query Patterns to Learn the Durations of Events Andrey Gusev joint work with Nate Chambers, Pranav Khaitan, Divye Khilnani, Steven Bethard, Dan Jurafsky.
Problem Semi supervised sarcasm identification using SASI
© author(s) of these slides including research results from the KOM research network and TU Darmstadt; otherwise it is specified at the respective slide.
TEMPLATE DESIGN © Identifying Noun Product Features that Imply Opinions Lei Zhang Bing Liu Department of Computer Science,
A Self Learning Universal Concept Spotter By Tomek Strzalkowski and Jin Wang Original slides by Iman Sen Edited by Ralph Grishman.
D ETERMINING THE S ENTIMENT OF O PINIONS Presentation by Md Mustafizur Rahman (mr4xb) 1.
Recognizing Implicit Discourse Relations in the Penn Discourse Treebank Ziheng Lin, Min-Yen Kan, and Hwee Tou Ng Department of Computer Science National.
Text Specificity and Impact on Quality of News Summaries Annie Louis & Ani Nenkova University of Pennsylvania June 24, 2011.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
A Self Learning Universal Concept Spotter By Tomek Strzalkowski and Jin Wang Presented by Iman Sen.
Learning Subjective Adjectives from Corpora Janyce M. Wiebe Presenter: Gabriel Nicolae.
A Memory-Based Approach to Semantic Role Labeling Beata Kouchnir Tübingen University 05/07/04.
Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, Bing Qin
Top Ten Tips for teachers preparing students for the academic version of IELTS Sam McCarter Macmillan Online Conference 2013.
Mining and Summarizing Customer Reviews
A Joint Model of Feature Mining and Sentiment Analysis for Product Review Rating Jorge Carrillo de Albornoz Laura Plaza Pablo Gervás Alberto Díaz Universidad.
(ACM KDD 09’) Prem Melville, Wojciech Gryc, Richard D. Lawrence
McEnery, T., Xiao, R. and Y.Tono Corpus-based language studies. Routledge. Unit A 2. Representativeness, balance and sampling (pp13-21)
Lemmatization Tagging LELA /20 Lemmatization Basic form of annotation involving identification of underlying lemmas (lexemes) of the words in.
Automatic Extraction of Opinion Propositions and their Holders Steven Bethard, Hong Yu, Ashley Thornton, Vasileios Hatzivassiloglou and Dan Jurafsky Department.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
A Compositional Context Sensitive Multi-document Summarizer: Exploring the Factors That Influence Summarization Ani Nenkova, Stanford University Lucy Vanderwende,
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Discovery of Manner Relations and their Applicability to Question Answering Roxana Girju 1,2, Manju Putcha 1, and Dan Moldovan 1 University of Texas at.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
A Weakly-Supervised Approach to Argumentative Zoning of Scientific Documents Yufan Guo Anna Korhonen Thierry Poibeau 1 Review By: Pranjal Singh Paper.
Recognizing Names in Biomedical Texts: a Machine Learning Approach GuoDong Zhou 1,*, Jie Zhang 1,2, Jian Su 1, Dan Shen 1,2 and ChewLim Tan 2 1 Institute.
14/12/2009ICON Dipankar Das and Sivaji Bandyopadhyay Department of Computer Science & Engineering Jadavpur University, Kolkata , India ICON.
1 Multi-Perspective Question Answering Using the OpQA Corpus (HLT/EMNLP 2005) Veselin Stoyanov Claire Cardie Janyce Wiebe Cornell University University.
Summarization Focusing on Polarity or Opinion Fragments in Blogs Yohei Seki Toyohashi University of Technology Visiting Scholar at Columbia University.
Minimally Supervised Event Causality Identification Quang Do, Yee Seng, and Dan Roth University of Illinois at Urbana-Champaign 1 EMNLP-2011.
CSKGOI'08 Commonsense Knowledge and Goal Oriented Interfaces.
1 CSC 594 Topics in AI – Text Mining and Analytics Fall 2015/16 3. Word Association.
Recognizing Stances in Ideological Online Debates.
Social Tag Prediction Paul Heymann, Daniel Ramage, and Hector Garcia- Molina Stanford University SIGIR 2008.
Blog Summarization We have built a blog summarization system to assist people in getting opinions from the blogs. After identifying topic-relevant sentences,
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Automatic recognition of discourse relations Lecture 3.
1 Generating Comparative Summaries of Contradictory Opinions in Text (CIKM09’)Hyun Duk Kim, ChengXiang Zhai 2010/05/24 Yu-wen,Hsu.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Opinion Observer: Analyzing and Comparing Opinions on the Web
Automatic acquisition for low frequency lexical items Nuria Bel, Sergio Espeja, Montserrat Marimon.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
Discovering Relations among Named Entities from Large Corpora Takaaki Hasegawa *, Satoshi Sekine 1, Ralph Grishman 1 ACL 2004 * Cyberspace Laboratories.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Annotating and measuring Temporal relations in texts Philippe Muller and Xavier Tannier IRIT,Université Paul Sabatier COLING 2004.
UROP Research Update Citation Function Classification Eric Yulianto A B 22 February 2013.
BAYESIAN LEARNING. 2 Bayesian Classifiers Bayesian classifiers are statistical classifiers, and are based on Bayes theorem They can calculate the probability.
Twitter as a Corpus for Sentiment Analysis and Opinion Mining
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Relation Extraction (RE) via Supervised Classification See: Jurafsky & Martin SLP book, Chapter 22 Exploring Various Knowledge in Relation Extraction.
Progress Update Lin Ziheng 6/26/2016. Outline  Update Summarization  Opinion Summarization  Discourse Analysis 6/26/2016.
Identifying Expressions of Opinion in Context Eric Breck and Yejin Choi and Claire Cardie IJCAI 2007.
Language Identification and Part-of-Speech Tagging
Erasmus University Rotterdam
Improving a Pipeline Architecture for Shallow Discourse Parsing
Presentation transcript:

Automatic sense prediction for implicit discourse relations in text Emily Pitler, Annie Louis, Ani Nenkova University of Pennsylvania ACL 2009

Implicit discourse relations Explicit comparison – I am in Singapore, but I live in the United States. Implicit comparison – The main conference is over Wednesday. I am staying for EMNLP. Explicit contingency – I am here because I have a presentation to give at ACL. Implicit contingency – I am a little tired; there is a 13 hour time difference.

Related work Soricut and Marcu (2003) – Sentence level Wellner et al. (2006) – used GraphBank annotations that do not differentiate between implicit and explicit. – difficult to verify success for implicit relations. Marcu and Echihabi (2001) – Artificial implicit – Delete the connective to generate dataset. – [Arg1, but Arg2] => [Arg1, Arg2]

Word pairs investigation The most easily accessible features are the words in the two text spans of the relation. There is some relationship that hold between the words in the two arguments. – The recent explosion of country finds mirrors the “closed-end fund mania” of the 1920s. Mr. Foot says, when narrowly focused funds grew wildly popular. They fell into oblivion after the 1929 crash. – Popular ( 受歡迎 ) and oblivion ( 被遺忘 ) are almost antonyms. – Triggers the contrast relation between the sentences.

Word pairs selection Marcu and Echihabi (2001) – Only nouns, verbs, and others cue phrases. – Using all words were superior to those based on only non- functions words. Lapata and Lascarides (2004) – Only verbs, nouns, and adjectives. – Verb pairs are one of best features. – No useful information was obtained using nouns and adjectives. Blair-Goldensohn et al. (2007) – Stemming. – Small vocabulary. – Cutoff on the minimum frequency of a feature. – Filtering stop-words has a negative impact on the results.

Analysis of word pair features Finding the word pairs with highest information gain on the synthetic data. – The government says it has reached most isolated townships by now, but because roads are blocked, getting anything but basic food supplies to people remains difficult. – Remove but => comparison example – Remove because => contingency example

Features for sense prediction Polarity tags Inquirer tags Money/Percent/Number Verbs First-last/first 3 words Context

Similar to word pairs, but words replaced with polarity tags. Each word’s polarity was assigned according to its entry in the Multi-perspective Question Answering Opinion Corpus (Wilson et al., 2005) Each sentiment word is tagged as positive, negative, both, or neutral. Using the number of negated and non-negated positive, negative, and neutral sentiment word in the two spans as features. Executives at Time Inc. Magazine Co., a subsidiary of Time Warner, have said the joint venture with Mr. Lang wasn’t a good one. [Negated Positive] The venture, formed in 1986, was supposed to be Time’s low-cost, safe entry into women’s magazines. [Positive] Polarity Tags pairs

Look up what semantic categories each word falls into according to the General Inquirer lexicon (Stone et al., 1966). See more observation for each semantic class than for any particular word, reducing the data sparsity problem. Complementary classes – “Understatement” vs. “Overstatement” – “Rise” vs. “Fall” – “Pleasure” vs. “Pain” Only verbs. Inquirer Tags

If two adjacent sentences both contain numbers, dollar amounts, or percentages, it is likely that a comparison relation might hold between the sentences. Count of numbers, percentages, and dollar amounts in the two arguments. Number of times each combination of number/percent/dollar occurs in the two arguments. Newsweek's circulation for the first six months of 1989 was 3,288,453, flat from the same period last year U.S. News' circulation in the same time was 2,303,328, down 2.6% Money/Percent/Num

Number of pairs of verbs in Arg1 and Arg2 from the same verb class. – Two verbs are from the same verb class if each of their highest Levin verb class levels are the same. – The more related the verbs, the more likely the relation is an Expansion. Average length of verb phrases in each argument – They [are allowed to proceed] => Contingency – They [proceed] => Expansion, Temporal POS tags of the main verb – Same tense => Expansion – Different tense => Contingency, Temporal Verbs

Prior work found first and last words very helpful in predicting sense – Wellner et al., 2006 – Often explicit connectives First-Last, First3

Some implicit relations appear immediately before or immediately after certain explicit relations. Indicating if the immediately preceding/following relation was an explicit. – Connective – Sense of the connective Indicating if an argument begins a paragraph. Context

Penn Discourse Treebank Largest available annotated corpus of discourse relations – Penn Treebank WSJ articles – 16,224 implicit relations between adjacent sentences I am a little tired; [because] there is a 13 hour time difference. – Contingency.cause.reason Use only the top level of the sense annotations. Dataset

Top level discourse relations Comparison ( 轉折 ) – 但是、可是、卻、即使、竟然、然而 …… Contingency ( 因果 ) – 因為、由於、因此、於是 …… Expansion ( 並列 ) – 又、並且、而且 …… Temporal ( 時序 ) – 在此之前、之後 ……

Relation SenseProportion of implicits Expansion53% Contingency26% Comparison15% Temporal 6% Discourse relations

Developed features on sections 0-1 Trained on sections 2-20 Tested on sections Binary classification task for each sense Trained on equal numbers of positive and negative examples Tested on natural distribution Naïve Bayes classifier Experiment setting

Results: comparison Featuresf-score First-Last, First Context19.32 Money/Percent/Num19.04 Random9.91 Polarity is actually the worst feature 16.63

Distribution of opposite polarity pairs ComparisonNot Comparison Positive-Negative or Negative-Positive Pairs 30%31%

Results: contingency Featuresf-score First-Last, First Verbs36.59 Context29.55 Random19.11

Results: expansion Featuresf-score Polarity Tags71.29 Inquirer Tags70.21 Context67.77 Random64.74 Expansion is majority class precision more problematic than recall These features all help other senses

Results: temporal Featuresf-score First-Last, First Verbs12.61 Context12.34 Random5.38 Temporals often end with words like “Monday” or “yesterday”

Best feature sets Comparison – Selected word pairs. Contingency – Polarity, verb, first/last, modality, context, selected word pairs. Expansion – Polarity, inquirer tags, context. Temporal – First/last, selected word pairs.

Best results RelationF-scorebaseline Comparison Contingency Expansion Temporal

Sequence model for discourse relations Tried Conditional random field classifier. ModelAccuracy Naïve Bayes Model43.27% Conditional Random Fields44.58%

Conclusion First study that predicts implicit discourse relations in a realistic setting. Better understanding of word pairs. – The feature in fact do not capture opposite semantic relation but rather give information about function word co-occurrences. Empirical validation of new and old features. – Polarity, verb classes, context, and some lexical features indicate discourse relations.