Sentiment analysis overview in the text area

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Neural networks Introduction Fitting neural networks
Farag Saad i-KNOW 2014 Graz- Austria,
Sentiment Analysis An Overview of Concepts and Selected Techniques.
Made with OpenOffice.org 1 Sentiment Classification using Word Sub-Sequences and Dependency Sub-Trees Pacific-Asia Knowledge Discovery and Data Mining.
A Brief Overview. Contents Introduction to NLP Sentiment Analysis Subjectivity versus Objectivity Determining Polarity Statistical & Linguistic Approaches.
A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts 04 10, 2014 Hyun Geun Soo Bo Pang and Lillian Lee (2004)
Techniques for Emotion Classification Julia Hirschberg COMS 4995/6998 Thanks to Kaushal Lahankar.
Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, Bing Qin
Distributed Representations of Sentences and Documents
Mining the Peanut Gallery: Opinion Extraction and Semantic Classification of Product Reviews K. Dave et al, WWW 2003, citations Presented by Sarah.
More than words: Social networks’ text mining for consumer brand sentiments A Case on Text Mining Key words: Sentiment analysis, SNS Mining Opinion Mining,
Opinion Mining of Customer Feedback Data on the Web Presented By Dongjoo Lee, Intelligent Databases Systems Lab. 1 Dongjoo Lee School of Computer Science.
Intelligent Database Systems Lab Advisor : Dr.Hsu Graduate : Keng-Wei Chang Author : Lian Yan and David J. Miller 國立雲林科技大學 National Yunlin University of.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Pattern Recognition. What is Pattern Recognition? Pattern recognition is a sub-topic of machine learning. PR is the science that concerns the description.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
A Document-Level Sentiment Analysis Approach Using Artificial Neural Network and Sentiment Lexicons Yan Zhu.
Mastering the Pipeline CSCI-GA.2590 Ralph Grishman NYU.
Unsupervised Learning Part 2. Topics How to determine the K in K-means? Hierarchical clustering Soft clustering with Gaussian mixture models Expectation-Maximization.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Big data classification using neural network
Sentiment analysis using deep learning methods
Jonatas Wehrmann, Willian Becker, Henry E. L. Cagnini, and Rodrigo C
Deep Feedforward Networks
Deep Learning for Bacteria Event Identification
Artificial Neural Networks
Sentiment analysis algorithms and applications: A survey
Deep Learning Amin Sobhani.
Randomness in Neural Networks
Sentence Modeling Representation of sentences is the heart of Natural Language Processing A sentence model is a representation and analysis of semantic.
Recurrent Neural Networks for Natural Language Processing
Erasmus University Rotterdam
A Hierarchical Model of Reviews for Aspect-based Sentiment Analysis
Intro to NLP and Deep Learning
Multimodal Learning with Deep Boltzmann Machines
Intelligent Information System Lab
Natural Language Processing of Knee MRI Reports
Machine Learning Basics
Aspect-based sentiment analysis
Data Mining Lecture 11.
Machine Learning Today: Reading: Maria Florina Balcan
Quanzeng You, Jiebo Luo, Hailin Jin and Jianchao Yang
A First Look at Music Composition using LSTM Recurrent Neural Networks
Chapter 3. Artificial Neural Networks - Introduction -
Word Embedding Word2Vec.
Creating Data Representations
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
The use of Neural Networks to schedule flow-shop with dynamic job arrival ‘A Multi-Neural Network Learning for lot Sizing and Sequencing on a Flow-Shop’
Emna Krichene 1, Youssef Masmoudi 1, Adel M
Parametric Methods Berlin Chen, 2005 References:
Attention.
Presentation By: Eryk Helenowski PURE Mentor: Vincent Bindschaedler
Word embeddings (continued)
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Attention for translation
CSE 291G : Deep Learning for Sequences
Introduction to Sentiment Analysis
Automatic Handwriting Generation
Topic: Semantic Text Mining
Recurrent Neural Networks
Random Neural Network Texture Model
CVPR 2019 Poster.
An introduction to neural network and machine learning
Presentation transcript:

Sentiment analysis overview in the text area --Yuanyuan Liu 纯文本领域

Sentiment analysis Sentiment analysis (also known as opinion mining) refers to the use of natural language processing, text analysis and computational linguistics to identify and extract subjective information in source materials. Sentiment analysis is widely applied to reviews and social media for a variety of applications, ranging from marketing to customer service. Generally speaking, sentiment analysis aims to determine the attitude of a speaker or a writer with respect to some topic or the overall contextual polarity of a document. The attitude may be his or her judgment or evaluation (see appraisal theory), affective state (that is to say, the emotional state of the author when writing), or the intended emotional communication (that is to say, the emotional effect the author wishes to have on the reader).

Introduction Goal Granularity Evaluation Document level Paragraph level Sentence level [feature/aspect level] Evaluation accuracy[precision and recall]

Methods knowledge-based techniques statistical methods classify text by affect categories based on the presence of unambiguous affect words such as happy, sad, afraid, and bored. assign arbitrary words a probable “affinity” to particular emotions. statistical methods Machine learning hybrid approaches

Measures using ML Classifier Neural networks Naïve Bayes Maximum Entropy (MaxEnt) Feature-based SVM … Neural networks Recurrent neural network(RNN) Convolutional neural network(CNN) Deep memory network and attention model

Sentiment Lexicons GI (The General Inquirer) LIWC (Linguistic Inquiry and Word Count) MPQA Subjectivity Cues Lexicon Bing Liu Opinion Lexicon SentiWordNet

Naïve Bayes assign to a given document d the class c∗ = arg maxc P (c | d) Assumption: the fi’s are conditionally independent given d’s class:

Naïve Bayes Advantages: Disadvantages: Simple Its conditional independence assumption clearly does not hold in real-world situations.

MaxEnt MaxEnt model when characterizing some unknown events with a statistical model, we should always choose the one that has Maximum Entropy http://homepages.inf.ed.ac.uk/lzhang10/maxent.html张宇博士

MaxEnt Advantages: Disadvantages: Adam Berger MaxEnt makes no assumptions about the relationships between features, and so might potentially perform better when conditional independence assumptions are not met. Disadvantages: A lot of computations. Adam Berger http://www.cs.cmu.edu/afs/cs/user/aberger/www/html/tutorial/tutorial.html

SVM Find a hyper plane and maximize the margin.

Datasets: movie reviews from the Internet Movie Database(IMDb) Accuracy comparison Thumbs up? Sentiment Classification using Machine Learning Techniques [Bo Pang and Lillian Lee ] Datasets: movie reviews from the Internet Movie Database(IMDb)

papers Survey: Thumbs up? Sentiment Classification using machine Learning Techniques (Pang & Lee) Opinion mining and sentiment analysis (Pang & Lee) Comprehensive Review Of Opinion Summarization (Kim et al) New Avenues in Opinion Mining and Sentiment Analysis (Cambria et al)

RNN A recurrent neural network (RNN) is a class of artificial neural network where connections between units form a directed cycle. This creates an internal state of the network which allows it to exhibit dynamic temporal behavior. Application: handwriting recognition  speech recognition

RNN

RNN

CNN A convolutional neural network (CNN, or ConvNet) is a type of feed-forward artificial neural network in which the connectivity pattern between its neurons is inspired by the organization of the animal visual cortex.

CNN

Aspect Level Sentiment Classification with Deep Memory Network Duyu Tang Bing Qin Ting Liu

Motivation Drawbacks of conventional neural models capture context information in an implicit way, and are incapable of explicitly exhibiting important context clues of an aspect. expensive computation Intuition: only some subset of context words are needed to infer the sentiment towards an aspect. E.g. “ great food but the service was dreadful! ”

Background:Memory network Question answering Central idea: inference with a long-term memory component Components: Memory m: an array of objects I: converts input to internal feature representation G: updates old memories with new input O: generates an output representation given a new input and the current memory state R: outputs a response based on the output representation IBM著名的Watson 用长期记忆(存储)部件来做推断,可以对其进行读写操作

Background: attention model One important property of human perception is that one does not tend to process a whole scene in its entirety at once. Instead, humans focus attention selectively on parts of the visual space to acquire information when and where it is needed, and combine information from different fixations over time to build up an internal representation of the scene, guiding future eye movements and decision making.

Deep memory network model aspect word sentence s = {w1, w2, … , wi, … , wn} Word embedding matrix: word embedding of wi : Task: determining the sentiment polarity of sentences towards the aspect wi. vocabulary size The dimension of the word vector

Overview of the approach Figure 1: An illustration of our deep memory network with three computational layers (hops) for aspect level sentiment classification

Attention model Content attention Location attention

Content attention Intuition: context words do not contribute equally to the semantic meaning of a sentence the importance of a word should be different if we focus on different aspect

Content attention Input: Output: external memory m: aspect vector vaspect : Output: mi is a piece of memory m αi ∈ [0,1] is the weight of mi and ∑i αi = 1

Calculation of αi Softmax function where

Location attention Intuition: a context word closer to the aspect should be more important than a farther one.

Location attention—model 1 The memory vector mi: vi ∈ Rdx1 is a location vector for word wi n is the sentence length k is the hop number li is the location of wi

Location attention—model 2 The memory vector mi: vi ∈ Rdx1 is a location vector for word wi

Location attention—model 3 The memory vector mi: vi is regarded as a parameter

Location attention—model 4 The memory vector mi: Different from Model 3, location representations are regarded as neural gates to control how many percent of word semantics is written into the memory.

The Need for Multiple Hops Computational models that are composed of multiple processing layers have the ability to learn representations of data with multiple levels of abstraction. In this work, the attention layer in one layer is essentially a weighted average compositional function, which is not powerful enough to handle the sophisticated computationality like negation, intensification and contrary in language.

Aspect level sentiment classification Regard the output vector in last hop as the feature, and feed it to a softmax layer for aspect level sentiment classification. Means; minimizing the cross entropy error of sentiment classification Loss function: gradient descent.

Experiments Datasets [from SemEval 2014]

Comparison to other methods accuracy runtime

Effects of location attention

Visualize Attention Models

Error Analysis 1. non-compositional sentiment expression. E.g. “dessert was also to die for!” 2. complex aspect expression consisting of many words. E.g. “ask for the round corner table next to the large window.” 3. sentimental relation between context words such as negation, comparison and condition. E.g. “but dinner here is never disappointing, even if the prices are a bit over the top”.

Conclusion develop deep memory networks that capture importance of context words for aspect level sentiment classification. leverage both content and location information. using multiple computational layers in memory network could obtain improved performance.

Thanks