ECE599/692 - Deep Learning Lecture 14 – Recurrent Neural Network (RNN)

Slides:



Advertisements
Similar presentations
Introduction to Recurrent neural networks (RNN), Long short-term memory (LSTM) Wenjie Pei In this coffee talk, I would like to present you some basic.
Advertisements

Deep Learning Neural Network with Memory (1)
ECE 6504: Deep Learning for Perception Dhruv Batra Virginia Tech Topics: –Recurrent Neural Networks (RNNs) –BackProp Through Time (BPTT) –Vanishing / Exploding.
Predicting the dropouts rate of online course using LSTM method
NOTE: To change the image on this slide, select the picture and delete it. Then click the Pictures icon in the placeholder to insert your own image. SHOW.
Learning to Answer Questions from Image Using Convolutional Neural Network Lin Ma, Zhengdong Lu, and Hang Li Huawei Noah’s Ark Lab, Hong Kong
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
Lecture 6 Smaller Network: RNN
Welcome deep loria !.
Deep Learning RUSSIR 2017 – Day 3
Convolutional Sequence to Sequence Learning
SD Study RNN & LSTM 2016/11/10 Seitaro Shinagawa.
Best viewed with Computer Modern fonts installed
Faster R-CNN – Concepts
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
CSE 190 Neural Networks: The Neural Turing Machine
CSE 190 Neural Networks: How to train a network to look and see
Recurrent Neural Networks for Natural Language Processing
Show and Tell: A Neural Image Caption Generator (CVPR 2015)
Matt Gormley Lecture 16 October 24, 2016
Deep Learning: Model Summary
ICS 491 Big Data Analytics Fall 2017 Deep Learning
Different Units Ramakrishna Vedantam.
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Deep Learning Workshop
RNNs: Going Beyond the SRN in Language Prediction
RNN and LSTM Using MXNet Cyrus M Vahid, Principal Solutions Architect
Image Captions With Deep Learning Yulia Kogan & Ron Shiff
A First Look at Music Composition using LSTM Recurrent Neural Networks
Recurrent Neural Networks
ECE 599/692 – Deep Learning Lecture 1 - Introduction
ECE 599/692 – Deep Learning Lecture 9 – Autoencoder (AE)
ECE 599/692 – Deep Learning Lecture 4 – CNN: Practical Issues
Tips for Training Deep Network
RNNs & LSTM Hadar Gorodissky Niv Haim.
ECE 599/692 – Deep Learning Lecture 5 – CNN: The Representative Power
Hairong Qi, Gonzalez Family Professor
Hairong Qi, Gonzalez Family Professor
Understanding LSTM Networks
Introduction to RNNs for NLP
Recurrent Neural Networks
Code Completion with Neural Attention and Pointer Networks
Long Short Term Memory within Recurrent Neural Networks
Other Classification Models: Recurrent Neural Network (RNN)
Lecture 16: Recurrent Neural Networks (RNNs)
Recurrent Encoder-Decoder Networks for Time-Varying Dense Predictions
Machine Translation(MT)
Graph Neural Networks Amog Kamsetty January 30, 2019.
Attention.
Please enjoy.
LSTM: Long Short Term Memory
Meta Learning (Part 2): Gradient Descent as LSTM
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
-- Ray Mooney, Association for Computational Linguistics (ACL) 2014
Recurrent Neural Networks (RNNs)
A unified extension of lstm to deep network
Rgh
Question Answering System
Recurrent Neural Networks
Deep learning: Recurrent Neural Networks CV192
Cengizhan Can Phoebe de Nooijer
Bidirectional LSTM-CRF Models for Sequence Tagging
ECE 517: Reinforcement Learning in Artificial Intelligence Lecture 17: TRTRL, Implementation Considerations, Apprenticeship Learning November 3, 2010.
ECE – Pattern Recognition Lecture 8 – Performance Evaluation
LHC beam mode classification
ECE – Pattern Recognition Lecture 4 – Parametric Estimation
Neural Machine Translation by Jointly Learning to Align and Translate
Andrew Karl, Ph.D. James Wisnowski, Ph.D. Lambros Petropoulos
Presentation transcript:

ECE599/692 - Deep Learning Lecture 14 – Recurrent Neural Network (RNN) Hairong Qi, Gonzalez Family Professor Electrical Engineering and Computer Science University of Tennessee, Knoxville http://www.eecs.utk.edu/faculty/qi Email: hqi@utk.edu

Outline Introduction LSTM vs. GRU Applications and Implementations References: [1] Luis Serrano, A Friendly Introduction to Recurrent Neural Networks, https://www.youtube.com/watch?v=UNmqTiOnRfg, Aug. 2018 [2] Brandon Rohrer, Recurrent Neural Networks (RNN) and Long, Short-Term Memory (LSTM), https://www.youtube.com/watch?v=WCUNPb-5EYI, Jun. 2017 [3] Denny Britz, Recurrent Neural Networks Tutorial, http://www.wildml.com/2015/09/recurrent-neural-networks-tutorial-part-1-introduction-to-rnns/, Sept. 2015 (Implementation) [4] Colah’s blog, Understanding LSTM Networks, http://colah.github.io/posts/2015-08-Understanding-LSTMs/, Aug. 2015

A friendly introduction to NN [1]

A friendly introduction to RNN

A more complicated case

A more complicated case (cont’d)

A more complicated case (cont’d)

The children’s book example from Brandon Rohrer [2] Doug saw Jane. Jane saw Spot. Spot saw Doug. Dictionary = {Doug, Jane, Spot, saw, .} Potential mistakes: Doug saw Doug. Doug saw Jane saw Spot saw Doug… Doug.

The standard RNN module     Gradient vanishing issue: by the end of the RNN, the data from the first timestep has very little impact on the output of the RNN. An example of word prediction, “I grew up in France… I speak fluent French.”

The long-short term memory (LSTM) module LSTMs are explicitly designed to avoid the long-term dependency problem.

Two keys of LSTM “Cell state” which works like a conveyor belt runs straight down the entire chain, easy for information to flow along without changes. “Gates” which control or decide what kind of information could go or throw away from the cell state. Cell state Gate

LSTM – forget gate Take the example of a language model trying to predict the next word based on all the previous ones. In such a problem, the cell state might include the gender of the present subject, so that the correct pronouns can be used. When we see a new subject, we want to forget the gender of the old subject.

LSTM – input gate

LSTM – cell state update It actually drops the information about the old subject’s gender and add the new information, as we decided in the previous steps.

LSTM – output gate

LSTM - revisit

The gated recurrent units (GRUs) module Similar with LSTM but with only two gates and less parameters. The “update gate” determines how much of previous memory to be kept. The “reset gate” determines how to combine the new input with the previous memory.

Comparison of the gating mechanism

LSTM vs. GRU

Application example: The talking face Goal: Given an arbitrary audio clip and a face image, automatically generate realistic and smooth face video with accurate lip sync. [Suwajanakorn et al., 2017] Application: Face animation, entertainment, video bandwidth reduction, etc.

The proposed framework The proposed method: conditional video generation

The proposed framework