Learn to Comment Lance Lebanoff Mentor: Mahdi. Emotion classification of text  In our neural network, one feature is the emotion detected in the image.

Slides:



Advertisements
Similar presentations
Feature Selection Advanced Statistical Methods in NLP Ling 572 January 24, 2012.
Advertisements

Latent Semantic Analysis
Introduction to Information Retrieval
2 Information Retrieval System IR System Query String Document corpus Ranked Documents 1. Doc1 2. Doc2 3. Doc3.
Scott Wen-tau Yih (Microsoft Research) Joint work with Vahed Qazvinian (University of Michigan)
Sentiment Analysis An Overview of Concepts and Selected Techniques.
Measuring Praise and Criticism: Inference of Semantic Orientation from Association Peter D. Turney National research Council Canada Michael L. Littman.
Feeler: Emotion Classification of Text Using Vector Space Model Presenter: Asif Salekin.
Comparison of information retrieval techniques: Latent semantic indexing (LSI) and Concept indexing (CI) Jasminka Dobša Faculty of organization and informatics,
Latent Semantic Analysis
Information Retrieval Ling573 NLP Systems and Applications April 26, 2011.
Search and Retrieval: More on Term Weighting and Document Ranking Prof. Marti Hearst SIMS 202, Lecture 22.
Intro to NLP - J. Eisner1 Words vs. Terms Taken from Jason Eisner’s NLP class slides:
1 Latent Semantic Indexing Jieping Ye Department of Computer Science & Engineering Arizona State University
Vector Space Information Retrieval Using Concept Projection Presented by Zhiguo Li
Information Retrieval in Text Part III Reference: Michael W. Berry and Murray Browne. Understanding Search Engines: Mathematical Modeling and Text Retrieval.
Scott Wen-tau Yih Joint work with Kristina Toutanova, John Platt, Chris Meek Microsoft Research.
1 CS 502: Computing Methods for Digital Libraries Lecture 12 Information Retrieval II.
TFIDF-space  An obvious way to combine TF-IDF: the coordinate of document in axis is given by  General form of consists of three parts: Local weight.
Information Retrieval: Models and Methods October 15, 2003 CMSC Gina-Anne Levow.
Dimension of Meaning Author: Hinrich Schutze Presenter: Marian Olteanu.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Distributed Representations of Sentences and Documents
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
1 Vector Space Model Rong Jin. 2 Basic Issues in A Retrieval Model How to represent text objects What similarity function should be used? How to refine.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
RuleML-2007, Orlando, Florida1 Towards Knowledge Extraction from Weblogs and Rule-based Semantic Querying Xi Bai, Jigui Sun, Haiyan Che, Jin.
Text mining.
Information Retrieval by means of Vector Space Model of Document Representation and Cascade Neural Networks Igor Mokriš, Lenka Skovajsová Institute of.
Indices Tomasz Bartoszewski. Inverted Index Search Construction Compression.
Feature selection LING 572 Fei Xia Week 4: 1/29/08 1.
Expressing Implicit Semantic Relations without Supervision ACL 2006.
10/22/2015ACM WIDM'20051 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis Voutsakis.
Pseudo-supervised Clustering for Text Documents Marco Maggini, Leonardo Rigutini, Marco Turchi Dipartimento di Ingegneria dell’Informazione Università.
Katrin Erk Vector space models of word meaning. Geometric interpretation of lists of feature/value pairs In cognitive science: representation of a concept.
A Bootstrapping Method for Building Subjectivity Lexicons for Languages with Scarce Resources Author: Carmen Banea, Rada Mihalcea, Janyce Wiebe Source:
Latent Semantic Indexing: A probabilistic Analysis Christos Papadimitriou Prabhakar Raghavan, Hisao Tamaki, Santosh Vempala.
14/12/2009ICON Dipankar Das and Sivaji Bandyopadhyay Department of Computer Science & Engineering Jadavpur University, Kolkata , India ICON.
Text mining. The Standard Data Mining process Text Mining Machine learning on text data Text Data mining Text analysis Part of Web mining Typical tasks.
SINGULAR VALUE DECOMPOSITION (SVD)
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
1 CSC 594 Topics in AI – Text Mining and Analytics Fall 2015/16 5. Document Representation and Information Retrieval.
Done this week  Getting familiar with corpus (100,000 words) of blog posts with corresponding emotions  NLP feature extraction  Multinomial logistic.
Gene Clustering by Latent Semantic Indexing of MEDLINE Abstracts Ramin Homayouni, Kevin Heinrich, Lai Wei, and Michael W. Berry University of Tennessee.
EMOTIONS. EMOTIONS What is an emotion? = a strong feeling.
A Clustering Method Based on Nonnegative Matrix Factorization for Text Mining Farial Shahnaz.
Introduction to String Kernels Blaz Fortuna JSI, Slovenija.
Jeff Howbert Introduction to Machine Learning Winter Machine Learning Natural Language Processing.
Web Search and Text Mining Lecture 5. Outline Review of VSM More on LSI through SVD Term relatedness Probabilistic LSI.
Application of latent semantic analysis to protein remote homology detection Wu Dongyin 4/13/2015.
Qual Presentation Daniel Khashabi 1. Outline  My own line of research  Papers:  Fast Dropout training, ICML, 2013  Distributional Semantics Beyond.
2/10/2016Semantic Similarity1 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis.
Sentiment Analysis Using Common- Sense and Context Information Basant Agarwal 1,2, Namita Mittal 2, Pooja Bansal 2, and Sonal Garg 2 1 Department of Computer.
Vector Semantics Dense Vectors.
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.

Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
From Frequency to Meaning: Vector Space Models of Semantics
Dimensionality Reduction and Principle Components Analysis
Kim Schouten, Flavius Frasincar, and Rommert Dekker
Semantic Processing with Context Analysis
Vector-Space (Distributional) Lexical Semantics
Efficient Estimation of Word Representation in Vector Space
Design open relay based DNS blacklist system
Word Embedding Word2Vec.
Introduction to JMP Text Explorer Platform
Introduction to Sentiment Analysis
Latent semantic space: Iterative scaling improves precision of inter-document similarity measurement Rie Kubota Ando. Latent semantic space: Iterative.
Vector Representation of Text
Presentation transcript:

Learn to Comment Lance Lebanoff Mentor: Mahdi

Emotion classification of text  In our neural network, one feature is the emotion detected in the image  Generated comment should show similar emotion  Studied 2 papers

Detecting Emotion in Text Kaitlyn Mulcrone

Vector Space Model  Represent corpus as a term-by-document matrix  Row -> word  Column -> document  Entry in the matrix -> td-idf score of the word for the document

VSM Example  Document 1: You got a new phone?  Document 2: Your new dog is a cute dog! Document 1Document 2... You.060 Got.060 A00 New00 Dog0.08 Your

Matrix Reduction  Term-by-document matrix is too large and sparse  Latent Semantic Analysis (LSA) using Singular Vector Decomposition (SVD)  Returns an approximated matrix with smaller dimensions

Emotional synset vectors  Synset – set of all synonyms of a given word  For each emotion, retrieve the synset using WordNet lexicon  Get vectors for each synonym from the term-by-document matrix  Save each synset as a list of vectors Emotion (happiness) WordNet Synset joy glee delight Term-by- document matrix Vector Vector Vector Emotional synset vectors

Testing  Given input sentence, compare the sentence to each emotional synset  Choose the emotion whose synset is most similar to the sentence Sentence dog has joy Term-by- document matrix Vector Vector Vector Sentence word vectors happiness sadness fear Emotions cosine

Iterative method  For each word in the sentence, get the word’s vector from the term-by- document matrix  Compare each word in the input sentence to each word in the first emotional synset  Find cosine similarity (cosine of the angle between the vectors)  Take the average of the cosines

Iterative method joy glee delight Sentence word vectors dog has joy cosine Emotional synset vectors (happiness)

Iterative method joy glee delight Sentence word vectors dog has joy Emotional synset vectors (happiness) cosine

Iterative method joy glee delight Emotional synset vectors (happiness) Sentence word vectors dog has joy Similarity = average of all 9 of these cosines cosine

Combinational method  Treat all of the words in the input sentence as if they were the same word for each document d in matrix for each word w in sentence vector d = ( Σ tf w,d ) * ( Σ idf w,d )  Similarly, treat all of the words in the emotional synset as the same word  Find cosine similarity between the sentence vector and the emotional synset vector

Combinational method Sentence words dog has joy Term-by- document matrix tf, idf Sentence vector ( Σ tf ) * ( Σ idf ) Emotional synset words (for happiness) joy delight glee Term-by- document matrix tf, idf Emotional synset vector ( Σ tf ) * ( Σ idf ) cosine Similarity

Classification  Find cosine similarity between input sentence and all emotion synsets  Choose the emotion with the max value  For our method, probably save all cosine similarities as a descriptor

Prior and contextual emotion of words in sentential context Diman Ghazi, Diana Inkpen, Stan Szpakowicz

Lexicons  NRC-Emotion lexicon  dictionary of words and their corresponding emotions  8 emotions  Prior polarity lexicon  polarity (positive / negative / neutral / both)  type (weak / strong) abandon angerantic.disgust fear joysadness trust surprise

Features  For the input sentence, find words that are in the emotional lexicon  Save features for each of these emotional words He was not sorrowful but instead excited. Features

Emotion-word features  Retrieved from lexicons  Emotion  Polarity

Part of speech features  Use Stanford tagger  Find the part of speech of the emotional word  Part of speech of 4 neighboring words

Dependency-tree features  Use Stanford parser

Dependency-tree features  neg – word is modified by negation word  amod – word is a noun modified by an adjective, or is an adjective modifying a noun  advmod – word is an adverb or head of adverbial phrase, or is modified by an adverb  modifies positive – word modifies a positive word form the prior-polarity lexicon  modifies-negative  modified-by-positive  modified-by-negative

Learning  Input these features into SVM or Logistic Regression  Logistic Regression had better results

Learning  For sentences with multiple emotional words:  For each emotional word, get probability distribution of emotional classes  For each emotional class, add up the probabilities for all of the emotional words  Select the class with the max probability

Future Work  Test Stanford CoreNLP code to extract features from input sentence  Organize features to be input into our neural network  Study other NLP methods