Lecture 23: More on Word Embeddings

Slides:



Advertisements
Similar presentations
Learning for Structured Prediction Overview of the Material TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A A.
Advertisements

CS621: Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 29– AI and Probability (exemplified through NLP) 4 th Oct, 2010.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
1 CS546: Machine Learning and Natural Language Preparation to the Term Project: - Dependency Parsing - Dependency Representation for Semantic Role Labeling.
Sum-Product Networks CS886 Topics in Natural Language Processing
27. May Topic Models Nam Khanh Tran L3S Research Center.
Deep Learning for Efficient Discriminative Parsing Niranjan Balasubramanian September 2 nd, 2015 Slides based on Ronan Collobert’s Paper and video from.
Haitham Elmarakeby.  Speech recognition
Daphne Koller Overview Maximum a posteriori (MAP) Probabilistic Graphical Models Inference.
NOTE: To change the image on this slide, select the picture and delete it. Then click the Pictures icon in the placeholder to insert your own image. SHOW.
Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
Bassem Makni SML 16 Click to add text 1 Deep Learning of RDF rules Semantic Machine Learning.
Fill-in-The-Blank Using Sum Product Network
S.Bengio, O.Vinyals, N.Jaitly, N.Shazeer
Brief Intro to Machine Learning CS539
Attention Model in NLP Jichuan ZENG.
Neural Machine Translation
Lecture 7: Constrained Conditional Models
Lecture 9: Part of Speech
Convolutional Sequence to Sequence Learning
CNN-RNN: A Unified Framework for Multi-label Image Classification
Who am I? Work in Probabilistic Machine Learning Like to teach 
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
Deep Learning for Program Repair
Neural Machine Translation by Jointly Learning to Align and Translate
Show and Tell: A Neural Image Caption Generator (CVPR 2015)
Lecture 24: Relation Extraction
Kai-Wei Chang University of Virginia
Are End-to-end Systems the Ultimate Solutions for NLP?
Lecture 24: NER & Entity Linking
Lecture 19: Question Answering
Attention Is All You Need
Trends and Future of NLP
Basic Intro Tutorial on Machine Learning and Data Mining
Deep Learning based Machine Translation
Conditional Random Fields
Variational Knowledge Graph Reasoning
RNN and LSTM Using MXNet Cyrus M Vahid, Principal Solutions Architect
Paraphrase Generation Using Deep Learning
Image Captions With Deep Learning Yulia Kogan & Ron Shiff
Lecture 9: Semantic Parsing
Probability and Time: Markov Models
Introduction to Text Generation
Bitcoin PRICE PREDICTION
Introduction to Natural Language Processing
Y2Seq2Seq: Cross-Modal Representation Learning for 3D Shape and Text by Joint Reconstruction and Prediction of View and Word Sequences 1, Zhizhong.
Machine Translation(MT)
Unsupervised Pretraining for Semantic Parsing
Probability and Time: Markov Models
RNN Encoder-decoder Architecture
Attention.
Word embeddings Text processing with current NNs requires encoding into vectors. One-hot encoding: N words encoded by length N vectors. A word gets a.
Please enjoy.
Word embeddings (continued)
Couse webpage: Final Project Kai-Wei Chang UCLA Couse webpage:
Natural Language Processing (NLP) Systems Joseph E. Gonzalez
Attention for translation
-- Ray Mooney, Association for Computational Linguistics (ACL) 2014
Presented by: Anurag Paul
Word representations David Kauchak CS158 – Fall 2016.
Neural Machine Translation using CNN
Question Answering System
Neural Machine Translation
Sequence-to-Sequence Models
Bidirectional LSTM-CRF Models for Sequence Tagging
Learning to Detect Human-Object Interactions with Knowledge
Neural Machine Translation by Jointly Learning to Align and Translate
Listen Attend and Spell – a brief introduction
The experiment based on hier-attention
Presentation transcript:

Lecture 23: More on Word Embeddings Kai-Wei Chang CS @ University of Virginia kw@kwchang.net Couse webpage: http://kwchang.net/teaching/NLP16 CS6501-NLP

Recap: Representation Learning How to represent words/phrases/sentences? CS6501-NLP

Auto-encoder and auto-decoder CS6501-NLP

Sequence to sequence models [Sutskever, Vinyals & Le 14] Have been shown effective in machine translation, image captioning and and many structured tasks CS6501-NLP

Structured prediction ⋂ Representation Learning NLP problems are structural Output variables are inter-correlated Need joint predictions Traditional approaches Graphical model approaches E.g., Probabilistic graphical models, structured perceptron Sequence of decisions E.g., incremental perceptron, L2S, transition- based methods SPNLP

Recent trends Landscape of methods in Deep⋂Structure Deep learning/hidden representation e.g., seq2seq, RNN Deep features into factors, traditional factor graph inference e.g., LSTM+CRF, graph transformer networks Globally optimized transitional-based approaches e.g., beam-search seq2seq, SyntaxNet … SPNLP

How to represent words? Treat words/n-gram as tokens Word Cluster – Brown clustering Lexical knowledge bases – WordNet, Thesaurus Word embeddings – continuous representation CS6501-NLP

Research on word embeddings Multi-sense word embeddings Multi-lingual word embeddings I got high interest on my savings from the bank. My interest lies in History. CS6501-NLP

Research on word embeddings Multi-sense multi-lingual word embeddings I got high interest on my savings from the bank. 我從銀行的儲蓄得到高額利息 My interest lies in History. 我的興趣在於歷史 CS6501-NLP

Research on word embeddings Encode lexical information in word embeddings CS6501-NLP