Paraphrase Generation Using Deep Learning

Slides:



Advertisements
Similar presentations
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Advertisements

Addressing the Rare Word Problem in Neural Machine Translation
Haitham Elmarakeby.  Speech recognition
NOTE: To change the image on this slide, select the picture and delete it. Then click the Pictures icon in the placeholder to insert your own image. SHOW.
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
DeepWalk: Online Learning of Social Representations
Attention Model in NLP Jichuan ZENG.
R-NET: Machine Reading Comprehension With Self-Matching Networks
Neural Machine Translation
Convolutional Sequence to Sequence Learning
CNN-RNN: A Unified Framework for Multi-label Image Classification
End-To-End Memory Networks
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
Sentence Modeling Representation of sentences is the heart of Natural Language Processing A sentence model is a representation and analysis of semantic.
Recurrent Neural Networks for Natural Language Processing
Neural Machine Translation by Jointly Learning to Align and Translate
Attention Is All You Need
Show and Tell: A Neural Image Caption Generator (CVPR 2015)
An Overview of Machine Translation
Deep Learning: Model Summary
ICS 491 Big Data Analytics Fall 2017 Deep Learning
Intro to NLP and Deep Learning
Different Units Ramakrishna Vedantam.
mengye ren, ryan kiros, richard s. zemel
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Neural Machine Translation By Learning to Jointly Align and Translate
Word2Vec CS246 Junghoo “John” Cho.
Neural Language Model CS246 Junghoo “John” Cho.
Grid Long Short-Term Memory
RNN and LSTM Using MXNet Cyrus M Vahid, Principal Solutions Architect
Image Captions With Deep Learning Yulia Kogan & Ron Shiff
A First Look at Music Composition using LSTM Recurrent Neural Networks
Final Presentation: Neural Network Doc Summarization
Introduction to Text Generation
Understanding LSTM Networks
The Big Health Data–Intelligent Machine Paradox
Memory-augmented Chinese-Uyghur Neural Machine Translation
Introduction to Natural Language Processing
Lip movement Synthesis from Text
Machine Translation(MT)
Natural Language to SQL(nl2sql)
Report by: 陆纪圆.
Learning linguistic structure with simple recurrent neural networks
Attention.
实习生汇报 ——北邮 张安迪.
Word2Vec.
Presentation By: Eryk Helenowski PURE Mentor: Vincent Bindschaedler
Please enjoy.
Word embeddings (continued)
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Attention for translation
-- Ray Mooney, Association for Computational Linguistics (ACL) 2014
Type Topic in here! Created by Educational Technology Network
Learn to Comment Mentor: Mahdi M. Kalayeh
Automatic Handwriting Generation
Presented by: Anurag Paul
Neural Machine Translation - Encoder-Decoder Architecture and Attention Mechanism Anmol Popli CSE 291G.
Topic: Semantic Text Mining
Neural Machine Translation using CNN
Question Answering System
Neural Machine Translation
Presented By: Harshul Gupta
Sequence-to-Sequence Models
Bidirectional LSTM-CRF Models for Sequence Tagging
Week 7 Presentation Ngoc Ta Aidean Sharghi
Week 6 Presentation Ngoc Ta Aidean Sharghi.
Neural Machine Translation by Jointly Learning to Align and Translate
The experiment based on hier-attention
Presentation transcript:

Paraphrase Generation Using Deep Learning Prasanna Vaidya Co Founder DiscoveryAI

Agenda What is Paraphrase Generation? Use Cases Building Blocks Technologies Publicly Available Datasets & Compute Power Evaluation Metrics Important Research Papers Questions & Answers

What is Paraphrase Generation? Paraphrasing, the act to express the same meaning in different possible ways, is an important subtask in various Natural Language Processing (NLP) Applications. How old is your child? —> Age of your kid?

Why it is important & Use Cases Information Retrieval Conversational Systems Content Summarisation

Research Areas Recognition - Identify if two textual units are paraphrases of each other Extraction - Extract paraphrase instances from a thesaurus or a corpus Generation - Generate a reference paraphrase given a source text

Building Blocks

Word Embeddings Word embedding is a technique where words or phrases from the vocabulary are mapped to vectors of real numbers. King http://projector.tensorflow.org

Neural Networks

Limitations of Neural Networks Neural Networks don’t have memory. http://colah.github.io/posts/2015-08-Understanding-LSTMs/

Enter Recurrent Neural Nets They are networks with loops in them, allowing information to persist.

Limitations of RNNs I grew up in Pune…I speak fluent Marathi. In theory, RNNs are absolutely capable of handling such “long-term dependencies.” Sadly, in practice, RNNs don’t seem to be able to learn them.

Long Short Term Memory LSTMs are explicitly designed to avoid the long-term dependency problem. Remembering information for long periods of time is their default behavior.

Similarity with Machine Translation Paraphrasing Task can be modelled as Machine Translation Task. How are you? —> ¿cómo estás?

Encoder Decoder Model Encoder encodes the input sequence to an internal representation called 'context vector' which is used by the decoder to generate the output sequence. The lengths of input and output sequences can be different. import seq2seq from seq2seq.models import SimpleSeq2Seq model = SimpleSeq2Seq(input_dim=5, hidden_dim=10, output_length=8, output_dim=8, depth=3) model.compile(loss='mse', optimizer='rmsprop')

Publicly Available Datasets http://paraphrase.org/#/download https://www.kaggle.com/quora/question-pairs-dataset https://www.microsoft.com/en-us/download/details.aspx?id=52398

Compute Requirements Training lasted for 32 hours with on p2.xlarge on AWS for PPDB

Evaluation Metrics BLEU (bilingual evaluation understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another. METEOR (Metric for Evaluation of Translation with Explicit Ordering) is based on the harmonic mean of unigram precision and recall, with recall weighted higher than precision. https://github.com/jhclark/multeval

Results - How are you? how you doin ' , man uh , how are you how ya been how ya feelin ' , kid how the hell are you

Important Research Papers Neural Paraphrase Generation with Stacked Residual LSTM Networks https://arxiv.org/pdf/1610.03098.pdf Paraphrase Generation with Deep Reinforcement Learning https://arxiv.org/abs/1711.00279 A Deep Generative Framework for Paraphrase Generation https://www.cse.iitk.ac.in/users/piyush/papers/deep-paraphrase-aaai2018.pdf

Thank You! Questions? prasanna@discovery.ai @getprasannav