ACL01 Workshop on Collocation1 Identifying Collocations for Recognizing Opinions Janyce Wiebe, Theresa Wilson, Matthew Bell University of Pittsburgh Office.

Slides:



Advertisements
Similar presentations
Farag Saad i-KNOW 2014 Graz- Austria,
Advertisements

Problem Semi supervised sarcasm identification using SASI
MINING FEATURE-OPINION PAIRS AND THEIR RELIABILITY SCORES FROM WEB OPINION SOURCES Presented by Sole A. Kamal, M. Abulaish, and T. Anwar International.
Playing the Telephone Game: Determining the Hierarchical Structure of Perspective and Speech Expressions Eric Breck and Claire Cardie Department of Computer.
Recognizing Contextual Polarity in Phrase-Level Sentiment Analysis Theresa Wilson Janyce Wiebe Paul Hoffmann University of Pittsburgh.
A Joint Model of Text and Aspect Ratings for Sentiment Summarization Ivan Titov (University of Illinois) Ryan McDonald (Google Inc.) ACL 2008.
A Brief Overview. Contents Introduction to NLP Sentiment Analysis Subjectivity versus Objectivity Determining Polarity Statistical & Linguistic Approaches.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Explorations in Tag Suggestion and Query Expansion Jian Wang and Brian D. Davison Lehigh University, USA SSM 2008 (Workshop on Search in Social Media)
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
1 Attributions and Private States Jan Wiebe (U. Pittsburgh) Theresa Wilson (U. Pittsburgh) Claire Cardie (Cornell U.)
Learning Subjective Nouns using Extraction Pattern Bootstrapping Ellen Riloff, Janyce Wiebe, Theresa Wilson Presenter: Gabriel Nicolae.
Learning Subjective Adjectives from Corpora Janyce M. Wiebe Presenter: Gabriel Nicolae.
Text Mining: Finding Nuggets in Mountains of Textual Data Jochen Dijrre, Peter Gerstl, Roland Seiffert Presented by Huimin Ye.
Forecasting with Twitter data Presented by : Thusitha Chandrapala MARTA ARIAS, ARGIMIRO ARRATIA, and RAMON XURIGUERA.
Extracting Opinions, Opinion Holders, and Topics Expressed in Online News Media Text Soo-Min Kim and Eduard Hovy USC Information Sciences Institute 4676.
Today Evaluation Measures Accuracy Significance Testing
Mining and Summarizing Customer Reviews
Mining the Peanut Gallery: Opinion Extraction and Semantic Classification of Product Reviews K. Dave et al, WWW 2003, citations Presented by Sarah.
Mining and Summarizing Customer Reviews Minqing Hu and Bing Liu University of Illinois SIGKDD 2004.
A Multivariate Biomarker for Parkinson’s Disease M. Coakley, G. Crocetti, P. Dressner, W. Kellum, T. Lamin The Michael L. Gargano 12 th Annual Research.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Automatic Extraction of Opinion Propositions and their Holders Steven Bethard, Hong Yu, Ashley Thornton, Vasileios Hatzivassiloglou and Dan Jurafsky Department.
Carmen Banea, Rada Mihalcea University of North Texas A Bootstrapping Method for Building Subjectivity Lexicons for Languages.
Short Introduction to Machine Learning Instructor: Rada Mihalcea.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
A Compositional Context Sensitive Multi-document Summarizer: Exploring the Factors That Influence Summarization Ani Nenkova, Stanford University Lucy Vanderwende,
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Incident Threading for News Passages (CIKM 09) Speaker: Yi-lin,Hsu Advisor: Dr. Koh, Jia-ling. Date:2010/06/14.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
1 Determining the Hierarchical Structure of Perspective and Speech Expressions Eric Breck and Claire Cardie Cornell University Department of Computer Science.
Persuasion Technique and Format. Activity I am in a very good and generous mood today. So I have decided to give away A’s for all the assignments this.
A Weakly-Supervised Approach to Argumentative Zoning of Scientific Documents Yufan Guo Anna Korhonen Thierry Poibeau 1 Review By: Pranjal Singh Paper.
1 Identifying Subjective Language Janyce Wiebe University of Pittsburgh.
Frames Icons. Over Time Means Issues of importance past, present and future Applying something historic to present knowledge Predicting something based.
A Bootstrapping Method for Building Subjectivity Lexicons for Languages with Scarce Resources Author: Carmen Banea, Rada Mihalcea, Janyce Wiebe Source:
1 Multi-Perspective Question Answering Using the OpQA Corpus (HLT/EMNLP 2005) Veselin Stoyanov Claire Cardie Janyce Wiebe Cornell University University.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
Chapter 23: Probabilistic Language Models April 13, 2004.
Automatic Identification of Pro and Con Reasons in Online Reviews Soo-Min Kim and Eduard Hovy USC Information Sciences Institute Proceedings of the COLING/ACL.
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
Blog Summarization We have built a blog summarization system to assist people in getting opinions from the blogs. After identifying topic-relevant sentences,
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Evaluating an Opinion Annotation Scheme Using a New Multi- perspective Question and Answer Corpus (AAAI 2004 Spring) Veselin Stoyanov Claire Cardie Diane.
Recognizing Stances in Online Debates Unsupervised opinion analysis method for debate-side classification. Mine the web to learn associations that are.
Learning Subjective Nouns using Extraction Pattern Bootstrapping Ellen Riloff School of Computing University of Utah Janyce Wiebe, Theresa Wilson Computing.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
1 Measuring the Semantic Similarity of Texts Author : Courtney Corley and Rada Mihalcea Source : ACL-2005 Reporter : Yong-Xiang Chen.
Data Mining and Decision Support
7/2003EMNLP031 Learning Extraction Patterns for Subjective Expressions Ellen Riloff Janyce Wiebe University of Utah University of Pittsburgh.
Learning Extraction Patterns for Subjective Expressions 2007/10/09 DataMining Lab 안민영.
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Annotating and measuring Temporal relations in texts Philippe Muller and Xavier Tannier IRIT,Université Paul Sabatier COLING 2004.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Twitter as a Corpus for Sentiment Analysis and Opinion Mining
Identifying Expressions of Opinion in Context Eric Breck and Yejin Choi and Claire Cardie IJCAI 2007.
Language Identification and Part-of-Speech Tagging
The 2nd Workshop on Argumentation Mining
CSC 594 Topics in AI – Natural Language Processing
Learning Subjective Adjectives From Corpora
Statistical NLP: Lecture 9
Evaluation and Its Methods
Statistical NLP : Lecture 9 Word Sense Disambiguation
Stance Classification of Ideological Debates
Presentation transcript:

ACL01 Workshop on Collocation1 Identifying Collocations for Recognizing Opinions Janyce Wiebe, Theresa Wilson, Matthew Bell University of Pittsburgh Office of Naval Research grant N

ACL01 Workshop on Collocation2 Introduction Subjectivity: aspects of language used to express opinions and evaluations (Banfield 1982) Relevant for many NLP applications, such as information extraction and text categorization This paper: identifying collocational clues of subjectivity

ACL01 Workshop on Collocation3 Outline Subjectivity Data and annotation Unigram features N-gram features Generalized N-gram features Document classification

ACL01 Workshop on Collocation4 Subjectivity Tagging Recognizing opinions and evaluations (Subjective sentences) as opposed to material objectively presented as true (Objective sentences) Banfield 1982, Fludernik 1993, Wiebe 1994, Stein & Wright 1995

ACL01 Workshop on Collocation5 Examples At several different levels, it’s a fascinating tale. subjective Bell Industries Inc. increased its quarterly to 10 cents from 7 cents a share. objective

ACL01 Workshop on Collocation6 Subjectivity “Complained” “You Idiot!” “Terrible product” “Speculated” “Maybe” “Enthused” “Wonderful!” “Great product”

ACL01 Workshop on Collocation7 Examples Strong addressee-oriented negative evaluation Recognizing flames (Spertus 1997) Personal filters (Kaufer 2000) I had in mind your facts, buddy, not hers. Nice touch. “Alleges” whenever facts posted are not in your persona of what is “real.”

ACL01 Workshop on Collocation8 Examples Opinionated, editorial language IR, text categorization (Kessler et al. 1997) Do the writers purport to be objective? Look, this is a man who has great numbers. We stand in awe of the Woodstock generation’s ability to be unceasingly fascinated by the subject of itself.

ACL01 Workshop on Collocation9 Examples Belief and speech reports Information extraction, summarization, intellectual attribution (Teufel & Moens 2000) Northwest Airlines settled the remaining lawsuits, a federal judge said. “The cost of health care is eroding our standard of living and sapping industrial strength”, complains Walter Maher.

ACL01 Workshop on Collocation10 Other Applications Review mining (Terveen et al. 1997) Clustering documents by ideology (Sack 1995) Style in machine translation and generation (Hovy 1987)

ACL01 Workshop on Collocation11 Potential Subjective Elements "The cost of health care is eroding standards of living and sapping industrial strength,” complains Walter Maher. Sap: potential subjective element Subjective element

ACL01 Workshop on Collocation12 Subjectivity Multiple types, sources, and targets We stand in awe of the Woodstock generation’s ability to be unceasingly fascinated by the subject of itself. Somehow grown-ups believed that wisdom adhered to youth.

ACL01 Workshop on Collocation13 Annotations Three levels: expression level sentence level document level Manually tagged + existing annotations

ACL01 Workshop on Collocation14 Expression Level Annotations [Perhaps you’ll forgive me] for reposting his response They promised [e+ 2 yet] more for [e+ 3 really good][e? 1 stuff]

ACL01 Workshop on Collocation15 Expression Level Annotations Difficult for manual and automatic tagging: detailed no predetermined classification unit To date: used for training and bootstrapping Probably the most natural level

ACL01 Workshop on Collocation16 Expression Level Data 1000 WSJ sentences (2J) 462 newsgroup messages (2J) words newsgroup data (1J) Single round of tagging; results promising Used to generate features, not for evaluation

ACL01 Workshop on Collocation17 Sentence Level Annotations “The cost of health care is eroding our standard of living and sapping industrial strength,’’ complains Walter Maher. “What an idiot,’’ the idiot presumably complained. A sentence is labeled subjective if any significant expression of subjectivity appears

ACL01 Workshop on Collocation18 Document Level Annotations This work: Opinion Pieces in the WSJ: editorials, letters to the editor, arts & leisure reviews + Free source of data + More directly related to applications Other work: flames 1-star to 5-star reviews

ACL01 Workshop on Collocation19 Document Level Annotations Opinion pieces contain objective sentences Non-opinion pieces contain subjective sentences Editorials contain facts supporting the argument News reports present reactions (van Dijk 1988) “Critics claim …” “Supporters argue …” Reviews contain information about the product

ACL01 Workshop on Collocation20 Class Proportions in WSJ Sample Non-Opinion Pieces Subjective sentences 43% Objective 57% Noise Opinion Pieces Objective 30%Subjective sentences 70% Noise

ACL01 Workshop on Collocation21 Word Distribution 13-17% of words are in opinion pieces 83-87% of words are in non-opinion pieces

ACL01 Workshop on Collocation22 Evaluation Metric for Feature S with Respect to Opinion Pieces Baseline for comparison # words in opinions / total # words Precision(S) = # instances of S in opinions / total # instances of S Given the distributions, precisions of even perfect subjectivity clues would be low Improvement over baseline taken as evidence of promising PSEs

ACL01 Workshop on Collocation23 Data Opinion Pieces Non-Opinion Pieces

ACL01 Workshop on Collocation24 Document Level Data Existing opinion-piece annotations used for training Manually refined classifications used for testing Identified editorials not marked as such 3 hours/edition Kappa =.93 for 2 judges 3 WSJ editions, each more than 150K words

ACL01 Workshop on Collocation25 Automatically Generated Unigram Features Adjective and verb features were generated using distributional similarity (Lin 1998, Wiebe 2000) Existing opinion-piece annotations used for training Manually refined annotations used for testing

ACL01 Workshop on Collocation26 Unigram Feature Results WSJ-10 WSJ-33 baseline 17% baseline 13% +prec/freq +prec/freq Adjs +21/ /2137 Verbs +16/ /3193

ACL01 Workshop on Collocation27 Example Adjective Feature conclusive, undiminished, brute, amazing, unseen, draconian, insurmountable, unqualified, poetic, foxy, vintage, jaded, tropical, distributional, discernible, adept, paltry, warm, reprehensible, astonishing, surprising, commonplace, crooked, dreary, virtuoso, trashy, sandy, static, virulent, desolate, ours, proficient, noteworthy, Insistent, daring, unforgiving, agreeable, uncritical, homicidal, comforting, erotic, resonant, ephemeral, believable, epochal, dense, exotic, topical, …

ACL01 Workshop on Collocation28 Unique Words hapax legomena More than expected single-instance words in subjective elements Unique-1-gram feature: all words that appear once in the test data Precision is 1.5 times baseline precision Frequent feature!

ACL01 Workshop on Collocation29 Unigram Feature Results WSJ-10 WSJ-33 baseline 17% baseline 13% Adjs +21/ /2137 Verbs +16/ /3193 Unique-1-gram +10/ /6048 Results are consistent, even with different identification procedures (similarly for WSJ-22)

ACL01 Workshop on Collocation30 Collocational PSEs get out what a for the last time just as well here we go again Started with the observation that low precision words often compose higher precision collocations

ACL01 Workshop on Collocation31 Identifying Collocational PSEs Searching for 2-grams, 3-grams, 4-grams No grammatical generalizations or constraints yet Train on the data annotated with subjective elements (expression level) Test on the manually-refined opinion-piece data (document level)

ACL01 Workshop on Collocation32 Identifying Collocational PSEs: Training Data (reminder) 1000 WSJ sentences (2J) 462 newsgroup messages (2J) words newsgroup data (1J) [Perhaps you’ll forgive me] for reposting his response They promised [e+ 2 yet] more for [e+ 3 really good] [e? 1 stuff]

ACL01 Workshop on Collocation33 N-Grams Each position is filled by a word POS pair in|prep the|det air|noun

ACL01 Workshop on Collocation34 Identifying Collocational PSEs: Training, Step 1 Precision(n-gram) = # subjective instances of n-gram / total # instances of n-gram Precision with respect to subjective elements calculated for all 1,2,3,4-grams in the training data An instance of an n-gram is subjective if each word in the instance is in a subjective element

ACL01 Workshop on Collocation35 Identifying Collocational PSEs: Training [Perhaps you’ll forgive me] for reposting his response They promised [e+ 2 yet] more for [e+ 3 really good] [e? 1 stuff] An instance of an n-gram is subjective if each word in the instance is in a subjective element

ACL01 Workshop on Collocation36 Identifying Collocational PSEs: Training, Step 2 N-gram PSEs selected based on their precisions, using two criteria: 1. Precision >= Precision >= maximum precision of its constituents

ACL01 Workshop on Collocation37 Identifying Collocational PSEs: Training, Step 2 prec (w1,w2) >= max (prec (w1), prec (w2)) prec (w1,w2,w3) >= max(prec(w1,w2),prec(w3)) or prec (w1,w2,w3) >= max(prec(w1),prec(w2,w3)) Precision >= maximum precision of its constituents

ACL01 Workshop on Collocation38 Results WSJ-10 WSJ-33 baseline 17% baseline 13% Adjs +21/ /2137 Verbs +16/ /3193 Unique-1-gram +10/ / grams +07/ / grams +09/ /262 4-grams +05/32 -03/30

ACL01 Workshop on Collocation39 Generalized Collocational PSEs Replace each single-instance word in the training data with “UNIQUE” Rerun the same training procedure, finding collocations such as highly|adverb UNIQUE|adj To test the new collocations on test data, first replace each single-instance word in the test data with “UNIQUE”

ACL01 Workshop on Collocation40 Results WSJ-10 WSJ-33 baseline 17% baseline 13% Adjs +21/ /2137 Verbs +16/ /3193 Unique-1-gram +10/ / grams +07/ / grams +09/ /262 4-grams +05/ /30 U-2-grams +24/ /288 U-3-grams +27/ /144 U-4-grams +83/3 +15/27

ACL01 Workshop on Collocation41 Example highly|adverb UNIQUE|adj highly unsatisfactory highly unorthodox highly talented highly conjectural highly erotic

ACL01 Workshop on Collocation42 Example UNIQUE|verb out|IN farm out chuck out ruling out crowd out flesh out blot out spoken out luck out

ACL01 Workshop on Collocation43 Examples UNIQUE|adj to|TO UNIQUE|verb impervious to reason strange to celebrate wise to temper UNIQUE|noun of|IN its|pronoun sum of its usurpation of its proprietor of its they|pronoun are|verb UNIQUE|noun they are fools they are noncontenders

ACL01 Workshop on Collocation44 How do Fixed and U-Collocations Compare? Started with the observation that low precision words often compose higher precision collocations Recall the original motivation for investigating fixed n-gram PSEs: But unique words are probably not low precision Are we finding the same collocations two different ways? Or are we finding new PSEs?

ACL01 Workshop on Collocation45 Comparison WSJ-10 2-grams 3-grams 4-grams Intersecting instances %overlap WSJ-33: all 0s

ACL01 Workshop on Collocation46 Opinion-Piece Recognition using Linear Regression %correct TP FP Adjs,verbs Ngrams Adjs,verbs,ngrams All features (+ max density) Max density: the maximum feature count in an 11-word window

ACL01 Workshop on Collocation47 Future Work Methods for recognizing non-compositional phrases (e.g., Lin 1999) Mutual bootstrapping (Rilof and Jones 1999) to alternatively recognize sequences and subjective fillers

ACL01 Workshop on Collocation48 Sentence Classification Binary Features: pronoun, adjective, number, modal ¬ “will “, adverb ¬ “not”, new paragraph Lexical feature: good for subj; good for obj; good for neither Probabilistic classifier 10-fold cross validation; 51% baseline 72% average accuracy across folds 82% average accuracy on sentences rated certain

ACL01 Workshop on Collocation49 Test for Bias: Marginal Homogeneity Worse the fit, greater the bias C1 C2 C4 C1 C3 C2C3C4 4+ = X4 3+ = X3 2+ = X2 1+ = X1 X1 +1 = X2 +2 = X3 +3 = X4 +4 = for all i

ACL01 Workshop on Collocation50 Test for Symmetric Disagreement: Quasi-Symmetry C1 C2 C4 C1 C3 C2C3C4 * * *** *** ** ** Tests relationships among the off-diagonal counts Better the fit, higher the correlation

ACL01 Workshop on Collocation51 Unigram PSEs Adjectives and Verbs identified using Lin’s distributional similarity (Lin 1998) Distributional similarity is often used in NLP to find synonyms Motivating hypothesis: words may be similar because they have similar subjective usages

ACL01 Workshop on Collocation52 Unigram Feature Generation AdjFeature = {} For all Adjectives A in the training data: S = A + N most similar words to A P = precision(S) in the training data if P > T: AdjFeature += S Many runs with various settings for N and T Choose values of N and T on a validation set Evaluate on a new test set

ACL01 Workshop on Collocation53 Lin’s Distributional Similarity Lin 1998 Ihaveabrowndog R1 R3 R2 R4 Word R W I R1 have have R2 dog brown R3 dog...

ACL01 Workshop on Collocation54 Filtering Seed Words Words+ Clusters Filtered Set Word + cluster removed if precision on training set < threshold

ACL01 Workshop on Collocation55 Parameters Seed Words Words+ Clusters Cluster size Threshold

ACL01 Workshop on Collocation56 Lin’s Distributional Similarity R W R W R W R W Word1 Word2 Pairs statistically correlated with Word1 Sum over RWint: I(Word1,RWint) + I(Word2,RWint) / Sum over RWw1: I(Word1,RWw1) + Sum over RWw2: I(Word2,RWw2)