Document-level Semantic Orientation and Argumentation Presented by Marta Tatu CS7301 March 15, 2005.

Slides:



Advertisements
Similar presentations
Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
Advertisements

Mustafa Cayci INFS 795 An Evaluation on Feature Selection for Text Clustering.
Improved TF-IDF Ranker
Distant Supervision for Emotion Classification in Twitter posts 1/17.
Sentiment Analysis Learning Sentiment Lexicons. Dan Jurafsky Semi-supervised learning of lexicons Use a small amount of information A few labeled examples.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
© author(s) of these slides including research results from the KOM research network and TU Darmstadt; otherwise it is specified at the respective slide.
Sentiment Analysis An Overview of Concepts and Selected Techniques.
Made with OpenOffice.org 1 Sentiment Classification using Word Sub-Sequences and Dependency Sub-Trees Pacific-Asia Knowledge Discovery and Data Mining.
A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts 04 10, 2014 Hyun Geun Soo Bo Pang and Lillian Lee (2004)
CSC 380 Algorithm Project Presentation Spam Detection Algorithms Kyle McCombs Bridget Kelly.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
The use of unlabeled data to improve supervised learning for text summarization MR Amini, P Gallinari (SIGIR 2002) Slides prepared by Jon Elsas for the.
Distributional Clustering of Words for Text Classification Authors: L.Douglas Baker Andrew Kachites McCallum Presenter: Yihong Ding.
CS Word Sense Disambiguation. 2 Overview A problem for semantic attachment approaches: what happens when a given lexeme has multiple ‘meanings’?
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
Approaches to automatic summarization Lecture 5. Types of summaries Extracts – Sentences from the original document are displayed together to form a summary.
Scalable Text Mining with Sparse Generative Models
Introduction to Machine Learning Approach Lecture 5.
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
An Automatic Segmentation Method Combined with Length Descending and String Frequency Statistics for Chinese Shaohua Jiang, Yanzhong Dang Institute of.
Mining the Peanut Gallery: Opinion Extraction and Semantic Classification of Product Reviews K. Dave et al, WWW 2003, citations Presented by Sarah.
Keyphrase Extraction in Scientific Documents Thuy Dung Nguyen and Min-Yen Kan School of Computing National University of Singapore Slides available at.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Search Engines and Information Retrieval Chapter 1.
1 A study on automatically extracted keywords in text categorization Authors:Anette Hulth and Be´ata B. Megyesi From:ACL 2006 Reporter: 陳永祥 Date:2007/10/16.
National Institute of Informatics Kiyoko Uchiyama 1 A Study for Introductory Terms in Logical Structure of Scientific Papers.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
PAUL ALEXANDRU CHIRITA STEFANIA COSTACHE SIEGFRIED HANDSCHUH WOLFGANG NEJDL 1* L3S RESEARCH CENTER 2* NATIONAL UNIVERSITY OF IRELAND PROCEEDINGS OF THE.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
A Weakly-Supervised Approach to Argumentative Zoning of Scientific Documents Yufan Guo Anna Korhonen Thierry Poibeau 1 Review By: Pranjal Singh Paper.
Sentiment Detection Naveen Sharma( ) PrateekChoudhary( ) Yashpal Meena( ) Under guidance Of Prof. Pushpak Bhattacharya.
Recognizing Names in Biomedical Texts: a Machine Learning Approach GuoDong Zhou 1,*, Jie Zhang 1,2, Jian Su 1, Dan Shen 1,2 and ChewLim Tan 2 1 Institute.
Bo Pang , Lillian Lee Department of Computer Science
Comparing and Ranking Documents Once our search engine has retrieved a set of documents, we may want to Rank them by relevance –Which are the best fit.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
Minimally Supervised Event Causality Identification Quang Do, Yee Seng, and Dan Roth University of Illinois at Urbana-Champaign 1 EMNLP-2011.
CSKGOI'08 Commonsense Knowledge and Goal Oriented Interfaces.
Processing of large document collections Part 5 (Text summarization) Helena Ahonen-Myka Spring 2005.
Blog Summarization We have built a blog summarization system to assist people in getting opinions from the blogs. After identifying topic-relevant sentences,
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Comparative Experiments on Sentiment Classification for Online Product Reviews Hang Cui, Vibhu Mittal, and Mayur Datar AAAI 2006.
Shallow Parsing for South Asian Languages -Himanshu Agrawal.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
11 Chapter 4 The Research Process – Theoretical Framework – (Stage 3 in Research Process) © 2009 John Wiley & Sons Ltd.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Single Document Key phrase Extraction Using Neighborhood Knowledge.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Jean-Yves Le Meur - CERN Geneva Switzerland - GL'99 Conference 1.
UROP Research Update Citation Function Classification Eric Yulianto A B 22 February 2013.
Concept-Based Analysis of Scientific Literature Chen-Tse Tsai, Gourab Kundu, Dan Roth UIUC.
COMP423 Summary Information retrieval and Web search  Vecter space model  Tf-idf  Cosine similarity  Evaluation: precision, recall  PageRank 1.
Twitter as a Corpus for Sentiment Analysis and Opinion Mining
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Thumbs up? Sentiment Classification using Machine Learning Techniques Jason Lewris, Don Chesworth “Okay, I’m really ashamed of it, but I enjoyed it. I.
Sentiment analysis algorithms and applications: A survey
Memory Standardization
Sentiment Analyzer Using a Multi-Level Classifier
An Overview of Concepts and Selected Techniques
Presentation transcript:

Document-level Semantic Orientation and Argumentation Presented by Marta Tatu CS7301 March 15, 2005

 or  ? Semantic Orientation Applied to Unsupervised Classification of Reviews Peter D. Turney ACL-2002

3 Overview  Unsupervised learning algorithm for classifying reviews as recommended or not recommended  The classification is based on the semantic orientation of the phrases in the review which contain adjectives and adverbs

4 Algorithm Input: review  Identify phrases that contain adjectives or adverbs by using a part-of-speech tagger  Estimate the semantic orientation of each phrase  Assign a class to the given review based on the average semantic orientation of its phrases Output: classification (  or  )

5 Step 1  Apply Brill’s part-of-speech tagger on the review  Adjective are good indicators of subjective sentences. In isolation: unpredictable steering (  ) / plot (  )  Extract two consecutive words: one is an adjective or adverb, the other provides the context First WordSecond WordThird Word (not extracted) 1.JJNN or NNSAnything 2.RB, RBR, or RBSJJNot NN nor NNS 3.JJ Not NN nor NNS 4.NN or NNSJJNot NN nor NNS 5.RB, RBR, or RBSVB, VBD, VBN, or VBGAnything

6 Step 2  Estimate the semantic orientation of the extracted phrases using PMI-IR (Turney, 2001)  Pointwise Mutual Information (Church and Hanks, 1989):  Semantic Orientation:  PMI-IR estimates PMI by issuing queries to a search engine (Altavista, ~350 million pages)

7 Step 2 – continued  Added 0.01 to hits to avoid division by zero  If hits(phrase NEAR “excellent”) and hits(phrase NEAR “poor”)≤4, then eliminate phrase  Added “AND (NOT host:epinions)” to the queries not to include the Epinions website

8 Step 3  Calculate the average semantic orientation of the phrases in the given review  If the average is positive, then   If the average is negative, then  PhrasePOS tagsSO direct depositJJ NN1.288 local branchJJ NN0.421 small partJJ NN0.053 online serviceJJ NN2.780 well otherRB JJ0.237 low feesJJ NNS0.333 … true serviceJJ NN other bankJJ NN inconveniently located RB VBN Average Semantic Orientation 0.322

9 Experiments  410 reviews from Epinions 170 (41%) (  ) 240 (59%) (  ) Average phrases per review: 26  Baseline accuracy: 59% DomainAccuracyCorrelation Automobiles84.00% Banks80.00% Movies65.83% Travel Destinations70.53% All74.39%0.5174

10 Discussion  What makes the movies hard to classify? The average SO tends to classify a recommended movies as not recommended Evil characters make good movies The whole is not necessarily the sum of the parts  Good beaches do not necessarily add up to a good vacation  But good automobile parts usually add up to a good automobile

11 Applications  Summary statistics for search engines  Summarization of reviews Pick out the sentence with the highest positive/negative semantic orientation given a positive/negative review  Filtering “flames” for newsgroups When the semantic orientation drops below a threshold, the message might be a potential flame

12 Questions ?  Comments ?  Observations ?

 ? Sentiment Classification using Machine Learning Techniques Bo Pang, Lillian Lee and Shivakumar Vaithyanathan EMNLP-2002

14 Overview  Consider the problem of classifying documents by overall sentiment  Three machine learning methods besides the human-generated lists of words Naïve Bayes Maximum Entropy Support Vector Machines

15 Experimental Data  Movie-review domain  Source: Internet Movie Database (IMDb)  Stars or numerical value ratings converted into positive, negative, or neutral » no need to hand label the data for training or testing  Maximum of 20 reviews/author/sentiment category 752 negative reviews 1301 positive reviews 144 reviewers

16 List of Words Baseline  Maybe there are certain words that people tend to use to express strong sentiments  Classification done by counting the number of positive and negative words in the document  Random-choice baseline: 50%

17 Machine Learning Methods  Bag-of-features framework: {f 1,…,f m } predefined set of m features n i (d) = number of times f i occurs in document d  (Naïve Bayes)

18 Machine Learning Methods – continued  (Maximum Entropy) where F i,c is a feature/class function:  Support vector machines: Find hyperplane that maximizes the margin. The constraint optimization problem: c j is the correct class of document d j

19 Evaluation  700 positive-sentiment and 700 negative- sentiment documents  3 equal-sized folds  The tag “NOT_” was added to every word between a negation word (“not”, “isn’t”, “didn’t”) and the first punctuation mark “good” is opposite to “not very good”  Features: 16,165 unigrams appearing at least 4 times in the 1400-document corpus 16,165 most often occurring bigrams in the same data

20 Results  POS information added to differentiate between: “I love this movie” and “This is a love story”

21 Conclusion  Results produced by the machine learning techniques are better than the human- generated baselines SVMs tend to do the best Unigram presence information is the most effective  Frequency vs. presence: “thwarted expectation”, many words indicative of the opposite sentiment to that of the entire review  Some form of discourse analysis is necessary

22 Questions ?  Comments ?  Observations ?

Summarizing Scientific Articles: Experiments with Relevance and Rhetorical Status Simone Teufel and Marc Moens CL-2002

24 Overview  Summarization of scientific articles: restore the discourse context of extracted material by adding the rhetorical status of each sentence in the document  Gold standard data for summaries consisting of computational linguistics articles annotated with the rhetorical status and relevance for each sentence  Supervised learning algorithm which classifies sentences into 7 rhetorical categories

25 Why?  Knowledge about the rhetorical status of the sentence enables the tailoring of the summaries according to user’s expertise and task Nonexpert summary: background information and the general purpose of the paper Expert summary: no background, instead differences between this approach and similar ones  Contrasts or complementarity among articles can be expressed

26 Rhetorical Status  Generalizations about the nature of scientific texts + information to enable the construction of better summaries  Problem structure: problems (research goals), solutions (methods), and results  Intellectual attribution: what the new contribution is, as opposed to previous work and background (generally accepted statements)  Scientific argumentation  Attitude toward other people’s work: rival approach, prior approach with a fault, or an approach contributing parts of the authors’ own solution

27 Metadiscourse and Agentivity  Metadiscourse is an aspect of scientific argumentation and a way of expressing attitude toward previous work “we argue that”, “in contrast to common belief, we”  Agent roles in argumentation: rivals, contributors of part of the solution (they), the entire research community, or the authors of the paper (we)

28 Citations and Relatedness  Just knowing that an article cites another is often not enough  One needs to read the context of the citation to understand the relation between the articles Article cited negatively or contrastively Article cited positively or in which the authors state that their own work originates from the cited work

29 Rhetorical Annotation Scheme  Only one category assigned to each full sentence  Nonoverlapping, nonhierarchical scheme  The rhetorical status is determined on the basis of the global context of the paper

30 Relevance  Select important content from text  Highly subjective » low human agreement  Sentence is considered relevant if it describes the research goal or states a difference with a rival approach  Other definitions: relevant sentence if it shows a high level of similarity with a sentence in the abstract

31 Corpus  80 conference articles Association for Computational Linguistics (ACL) European Chapter of the Association for Computational Linguistics (EACL) Applied Natural Language Processing (ANLP) International Joint Conference on Artificial Intelligence (IJCAI) International Conference on Computational Linguistics (COLING).  XML markups added

32 The Gold Standard  3 tasked-trained annotators  17 pages of guidelines  20 hours of training  No communication between annotators  Evaluation measures of the annotation: Stability Reproducibility

33 Results of Annotation  Kappa coefficient K (Siegel and Castellan, 1988) where P(A)= pairwise agreement and P(E)= random agreement  Stability: K=.82,.81,.76 (N=1,220 and k=2)  Reproducibility: K=.71

34 The System  Supervised machine learning Naïve Bayes

35 Features  Absolute location of a sentence Limitations of the author’s own method can be expected to be found toward the end, while limitations of other researchers’ work are discussed in the introduction

36 Features – continued  Section structure: relative and absolute position of sentence within section: First, last, second or third, second-last or third-last, or either somewhere in the first, second, or last third of the section  Paragraph structure: relative position of sentence within a paragraph Initial, medial, or final

37 Features – continued  Headlines: type of headline of current section Introduction, Implementation, Example, Conclusion, Result, Evaluation, Solution, Experiment, Discussion, Method, Problems, Related Work, Data, Further Work, Problem Statement, or Non-Prototypical  Sentence length Longer or shorter than 12 words (threshold)

38 Features – continued  Title word contents: does the sentence contain words also occurring in the title?  TF*IDF word contents High values to words that occur frequently in one document, but rarely in the overall collection of documents Do the 18 highest-scoring TF*IDF words belong to the sentence?  Verb syntax: voice, tense, and modal linguistic features

39 Features – continued  Citation Citation (self), citation (other), author name, or none + location of the citation in the sentence (beginning, middle, or end)  History: most probable previous category AIM tends to follow CONTRAST Calculated as a second pass process during training

40 Features – continued  Formulaic expressions: list of phrases described by regular expressions, divided into 18 classes, comprising a total of 644 patterns Clustering prevents data sparseness

41 Features – continued  Agent: 13 types, 167 patterns The placeholder WORK_NOUN can be replaced by a set of 37 nouns including theory, method, prototype, algorithm Agent classes with a distribution very similar with the overall distribution of target categories were excluded

42 Features – continued  Action: 365 verbs clustered into 20 classes based on semantic concepts such as similarity, contrast PRESENTATION_ACTIONs: present, report, state RESEARCH_ACTIONs: analyze, conduct, define, and observe Negation is considered

43 System Evaluation  10-fold-cross-validation

44 Feature Impact  The most distinctive single feature is Location, followed by SegAgent, Citations, Headlines, Agent and Formulaic

45 Questions ?  Comments ?  Observations ?

46 Thank You !