Milton King, Waseem Gharbieh, Sohyun Park, and Paul Cook

Slides:



Advertisements
Similar presentations
Text Categorization.
Advertisements

Mining User Similarity Based on Location History Yu Zheng, Quannan Li, Xing Xie Microsoft Research Asia.
Chapter 5: Introduction to Information Retrieval
Linked data: P redicting missing properties Klemen Simonic, Jan Rupnik, Primoz Skraba {klemen.simonic, jan.rupnik,
Comparing Twitter Summarization Algorithms for Multiple Post Summaries David Inouye and Jugal K. Kalita SocialCom May 10 Hyewon Lim.
| 1 › Gertjan van Noord2014 Zoekmachines Lecture 4.
Ranking models in IR Key idea: We wish to return in order the documents most likely to be useful to the searcher To do this, we want to know which documents.
Introduction to Information Retrieval (Manning, Raghavan, Schutze) Chapter 6 Scoring term weighting and the vector space model.
Information Retrieval Ling573 NLP Systems and Applications April 26, 2011.
Search and Retrieval: More on Term Weighting and Document Ranking Prof. Marti Hearst SIMS 202, Lecture 22.
Database Management Systems, R. Ramakrishnan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides.
Semantic text features from small world graphs Jure Leskovec, IJS + CMU John Shawe-Taylor, Southampton.
Hinrich Schütze and Christina Lioma
Formal Multinomial and Multiple- Bernoulli Language Models Don Metzler.
Distributed Representations of Sentences and Documents
Term weighting and vector representation of text Lecture 3.
Evaluation of N-grams Conflation Approach in Text-based Information Retrieval Serge Kosinov University of Alberta, Computing Science Department, Edmonton,
EVENT IDENTIFICATION IN SOCIAL MEDIA Hila Becker, Luis Gravano Mor Naaman Columbia University Rutgers University.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Advanced Multimedia Text Classification Tamara Berg.
Web search basics (Recap) The Web Web crawler Indexer Search User Indexes Query Engine 1 Ad indexes.
Personalisation Seminar on Unlocking the Secrets of the Past: Text Mining for Historical Documents Sven Steudter.
The Problem Finding information about people in huge text collections or on-line repositories on the Web is a common activity Person names, however, are.
Predicting Content Change on the Web Kira Radinsky Technion, Israel Paul Bennettt Microsoft Research.
Eric H. Huang, Richard Socher, Christopher D. Manning, Andrew Y. Ng Computer Science Department, Stanford University, Stanford, CA 94305, USA ImprovingWord.
LexRank: Graph-based Centrality as Salience in Text Summarization
2014 EMNLP Xinxiong Chen, Zhiyuan Liu, Maosong Sun State Key Laboratory of Intelligent Technology and Systems Tsinghua National Laboratory for Information.
1 Computing Relevance, Similarity: The Vector Space Model.
CPSC 404 Laks V.S. Lakshmanan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides at UC-Berkeley.
Vector Space Models.
Introduction to String Kernels Blaz Fortuna JSI, Slovenija.
Aim: What is the arithmetic series ? Do Now: Find the sum of each of the following sequences: a) b)
CIS 530 Lecture 2 From frequency to meaning: vector space models of semantics.
Ranked Retrieval INST 734 Module 3 Doug Oard. Agenda Ranked retrieval  Similarity-based ranking Probability-based ranking.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Learning in a Pairwise Term-Term Proximity Framework for Information Retrieval Ronan Cummins, Colm O’Riordan Digital Enterprise Research Institute SIGIR.
Utilizing vector models for automatic text lemmatization Ladislav Gallay Supervisor: Ing. Marián Šimko, PhD. Slovak University of Technology Faculty of.
Huffman code and Lossless Decomposition Prof. Sin-Min Lee Department of Computer Science.
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
Collaborative Filtering With Decoupled Models for Preferences and Ratings Rong Jin 1, Luo Si 1, ChengXiang Zhai 2 and Jamie Callan 1 Language Technology.
IR 6 Scoring, term weighting and the vector space model.
Sentiment Analysis CMPT 733. Outline What is sentiment analysis? Overview of approach Feature Representation Term Frequency – Inverse Document Frequency.
Rationalizing Neural Predictions
A Simple Approach for Author Profiling in MapReduce
Plan for Today’s Lecture(s)
Korean version of GloVe Applying GloVe & word2vec model to Korean corpus speaker : 양희정 date :
End-To-End Memory Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
A German Corpus for Similarity Detection
Indexing & querying text
Deep Compositional Cross-modal Learning to Rank via Local-Global Alignment Xinyang Jiang, Fei Wu, Xi Li, Zhou Zhao, Weiming Lu, Siliang Tang, Yueting.
An Empirical Study of Learning to Rank for Entity Search
A Deep Learning Technical Paper Recommender System
Distributed Representations of Words and Phrases and their Compositionality Presenter: Haotian Xu.
Vector-Space (Distributional) Lexical Semantics
Paraphrase Generation Using Deep Learning
Representation of documents and queries
Word Embedding Word2Vec.
MTBI Personality Predictor using ML
Word2Vec.
Hsien-Chin Lin, Chi-Yu Yang, Hung-Yi Lee, Lin-shan Lee
ARTS>PIEs>26 Lines
Presented by: Anurag Paul
Building Dynamic Knowledge Graphs From Text Using Machine Reading Comprehension Rajarshi Das, Tsendsuren Munkhdalai, Xingdi Yuan, Adam Trischler, Andrew.
Natural Language Processing Is So Difficult
Presented By: Harshul Gupta
Baseline Model CSV Files Pandas DataFrame Sentence Lists
Week 3 Presentation Ngoc Ta Aidean Sharghi.
Presented by Nick Janus
Presentation transcript:

Milton King, Waseem Gharbieh, Sohyun Park, and Paul Cook Semantic Textual Similarity: A Unified Framework for Semantic Processing and Evaluation Milton King, Waseem Gharbieh, Sohyun Park, and Paul Cook Semantic Textual Similarity Paragraph Vectors: An extension of word2vec to text of arbitrary length. The Distributed Memory Model of Paragraph Vectors (PV-DM) was used to represent each sentence as a vector. Given two sentences, determine their semantic similarity. Similarity is a score from 0 to 5. 0: no similarity. For example: - The farther away, the faster the galaxies move away from us - Here is my answer to a similar question posted on the physics stack exchange website 5: equivalent. For example: - BlackBerry loses US$965m in Q2 - BlackBerry loses $965M in 2nd quarter Performance: Pearson correlation with human judgements. Experimental Setup Sentence similarity is the cosine of their vector representations. Evaluated on SemEval 2016 STS datasets. 5 text types; 9183 sentence pairs total. Baseline Approaches Baseline Binary: The vectors hold binary values indicating whether the corresponding word occurs in the sentence. Baseline Frequency: The vectors hold the frequency of the corresponding word in the sentence. Tf-idf: Each vector holds the tf-idf weight for the corresponding word in the sentence. The aim of tf-idf is to de-emphasize frequent words by giving more weight to less frequent ones. Results Method Answer-answer Headlines Plagiarism Post-editing Question-question All Baseline Binary 0.50937 0.70636 0.80108 0.76370 0.61827 0.67881 Baseline Frequency 0.44204 0.72754 0.79604 0.79483 0.65749 0.68122 Tf-idf 0.45928 0.66593 0.75778 0.77204 0.61710 0.65271 Word2vec-Prod 0.39310 0.60667 0.71528 0.21306 0.10847 0.41322 Word2vec-Sum 0.13521 0.14328 0.23290 -0.02673 0.25153 0.14303 Paragraph-vectors 0.41123 0.69169 0.60488 0.75547 -0.02245 0.50206 Skip-thoughts 0.27148 0.23199 0.49643 0.48636 0.17749 0.33446 Skip-thoughts-Reg 0.28626 0.51019 0.66708 0.69947 0.40459 0.51299 Average 0.58520 0.69006 0.78923 0.82540 0.58605 0.69635 Regression 0.55254 0.71353 0.79769 0.81291 0.62037 0.69940 Vector Embeddings Word2vec: Each sentence is represented as the element-wise summation, and product, of the word embedding vectors for the words in that sentence. Skip-thoughts: An encoder-decoder model composed of gated recurrent units used for sequence modeling. Conclusions None of the vector embedding approaches improved over any of the baselines by itself. Combining vector embedding approaches via averaging and regression achieved modest improvements over the baselines.