Recent Developments in Deep Learning Quoc V. Le Stanford University and Google.

Slides:



Advertisements
Similar presentations
Advanced topics.
Advertisements

Rajat Raina Honglak Lee, Roger Grosse Alexis Battle, Chaitanya Ekanadham, Helen Kwong, Benjamin Packer, Narut Sereewattanawoot Andrew Y. Ng Stanford University.
Generalizing Backpropagation to Include Sparse Coding David M. Bradley and Drew Bagnell Robotics Institute Carnegie.
ImageNet Classification with Deep Convolutional Neural Networks
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009
Deep Learning.
Supervised and Unsupervised learning and application to Neuroscience Cours CA6b-4.
Unsupervised Learning With Neural Nets Deep Learning and Neural Nets Spring 2015.
AN ANALYSIS OF SINGLE- LAYER NETWORKS IN UNSUPERVISED FEATURE LEARNING [1] Yani Chen 10/14/
Submitted by:Supervised by: Ankit Bhutani Prof. Amitabha Mukerjee (Y )Prof. K S Venkatesh.
Image Classification using Sparse Coding: Advanced Topics
Comp 5013 Deep Learning Architectures Daniel L. Silver March,
Nantes Machine Learning Meet-up 2 February 2015 Stefan Knerr CogniTalk
Kuan-Chuan Peng Tsuhan Chen
CSC2535: Advanced Machine Learning Lecture 6a Convolutional neural networks for hand-written digit recognition Geoffrey Hinton.
Deep Learning for Vision
Deep Learning for Speech and Language Yoshua Bengio, U. Montreal NIPS’2009 Workshop on Deep Learning for Speech Recognition and Related Applications December.
DistBelief: Large Scale Distributed Deep Networks Quoc V. Le
Video Tracking Using Learned Hierarchical Features
A shallow introduction to Deep Learning
Large-scale Deep Unsupervised Learning using Graphics Processors
Presented by: Mingyuan Zhou Duke University, ECE June 17, 2011
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Dr. Z. R. Ghassabi Spring 2015 Deep learning for Human action Recognition 1.
Andrew Ng Machine Learning and AI via Brain simulations Andrew Ng Stanford University Adam Coates Quoc Le Honglak Lee Andrew Saxe Andrew Maas Chris Manning.
Deep Convolutional Nets
Neural networks in modern image processing Petra Budíková DISA seminar,
A shallow look at Deep Learning
Introduction to Deep Learning
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
ImageNet Classification with Deep Convolutional Neural Networks Presenter: Weicong Chen.
Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.
Convolutional Neural Network
Neural Networks William Cohen [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ]
Introduction to Convolutional Neural Networks
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Understanding Convolutional Neural Networks for Object Recognition
Convolutional Neural Networks
Deep Learning and Its Application to Signal and Image Processing and Analysis Class III - Fall 2016 Tammy Riklin Raviv, Electrical and Computer Engineering.
Convolutional Neural Network
The Relationship between Deep Learning and Brain Function
Deep Learning Amin Sobhani.
Data Mining, Neural Network and Genetic Programming
ECE 5424: Introduction to Machine Learning
Convolutional Neural Fabrics by Shreyas Saxena, Jakob Verbeek
Deep Learning Insights and Open-ended Questions
Week III: Deep Tracking
Learning Mid-Level Features For Recognition
Article Review Todd Hricik.
Neural networks (3) Regularization Autoencoder
ECE 6504 Deep Learning for Perception
Deep learning and applications to Natural language processing
Training Techniques for Deep Neural Networks
Deep Belief Networks Psychology 209 February 22, 2013.
Dipartimento di Ingegneria «Enzo Ferrari»
Machine Learning: The Connectionist
Department of Electrical and Computer Engineering
Introduction to Neural Networks
Limitations of Traditional Deep Network Architectures
Towards Understanding the Invertibility of Convolutional Neural Networks Anna C. Gilbert1, Yi Zhang1, Kibok Lee1, Yuting Zhang1, Honglak Lee1,2 1University.
[Figure taken from googleblog
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Visualizing and Understanding Convolutional Networks
Deep Learning Some slides are from Prof. Andrew Ng of Stanford.
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
CS295: Modern Systems: Application Case Study Neural Network Accelerator Sang-Woo Jun Spring 2019 Many slides adapted from Hyoukjun Kwon‘s Gatech “Designing.
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION
Presentation transcript:

Recent Developments in Deep Learning Quoc V. Le Stanford University and Google

Purely supervised Quoc V. Le

Almost abandoned between Overfitting, slow, many local minima, gradient vanishing In 2006, Hinton, et. al. proposed RBMs to pretrain a deep neural network In 2009, Raina, et. al. proposed to use GPUs to train deep neural network Deep Learning Quoc V. Le

Deep Learning In 2010, Dahl, et. al. trained a deep neural network using GPUs to beat the state-of-the-art in speech recognition In 2012, Le, et. al. trained a deep neural network using a cluster of machines to beat the state-of-the-art in ImageNet In 2012, Krizhevsky, et. al. won the ImageNet challenge with NN In 2012, Mikolov, et. al. trained a recurrent neural network to achieve state-of-the-art in language modelling Quoc V. Le

State-of-the-art in Acoustic Modelling Acoustic modelling: -Previous method: Mixture of Gaussians -M.D. Zeiler, M. Ranzato, R. Monga, M. Mao, K. Yang, Q.V. Le, P. Nguyen, A. Senior, V. Vanhoucke, J. Dean, G. Hinton. On Rectified Linear Units for Speech Processing. ICASSP, HMM Language modelling Acoustic modelling Quoc V. Le

Purely supervised Classifying phonemes Quoc V. Le

State-of-the-art in Computer Vision -Previous method: Hand-crafted features -Q.V. Le, M.A. Ranzato, R. Monga, M. Devin, K. Chen, G.S. Corrado, J. Dean, A.Y. Ng. Building high-level features using large scale unsupervised learning. ICML, Krizhevsky, A., Sutskever, I. and Hinton, G. E. ImageNet Classification Using Deep Convolutional Neural Networks. NIPS 2012 Quoc V. Le

-Architecture: -Trained using unsupervised data, layer by layer State-of-the-art in Computer Vision Quoc V. Le

Deep Learning at Google What Google have? -Lots of data -Lots of computations -Problems that require good features What Google don’t have? -Time to invent features for each of the problems Quoc V. Le

Local receptive field networks Machine #1Machine #2 Machine #3Machine #4 Le, et al., Tiled Convolutional Neural Networks. NIPS 2010 RICA features Image

Le, et al., Building high-level features using large-scale unsupervised learning. ICML 2012 Dean, et al., Large scale distributed deep networks. NIPS Asynchronous Parallel Stochastic Gradient Descent Parameter Server Model Workers Data Shards W ’ = W +  ∆W ∆W∆W W’W’

x: Input data m: Number of examples  Trade of between reconstruction and sparsity W: Parameter matrix Number of rows in W: The number of features Feature representation: Le, et al., ICA with Reconstruction Cost for Efficient Overcomplete Feature Learning. NIPS 2011 Sparse Autoencoders (RICA - Le, et al, 2011)

Training Dataset: 10 million 200x200 unlabeled images from YouTube/Web Train on 2000 machines (16000 cores) for 1 week using Google infrastructure 1.15 billion parameters -100x larger than previously reported -Small compared to visual cortex Pooling Size = 5 Number of maps = 8 Image Size = 200 Number of output channels = 8 Number of input channels = 3 One layer RF size = 18 Input to another layer above (image with 8 channels) W H LCN Size = 5 Le, et al., Building high-level features using large-scale unsupervised learning. ICML 2012 Image RICA

x1x1 x2x2 x3x3 x4x4 a3a3 a2a2 a1a1 Visualization

Top stimuli from the test setOptimal stimulus by numerical optimization The face neuron Le, et al., Building high-level features using large-scale unsupervised learning. ICML 2012

Optimal stimulus by numerical optimization Le, et al., Building high-level features using large-scale unsupervised learning. ICML 2012 The cat neuron

Pooling Size = 5 Number of maps = 8 Image Size = 200 Number of output channels = 8 Number of input channels = 3 One layer RF size = 18 Input to another layer above (image with 8 channels) W H LCN Size = 5 Feature 2 Feature 3 Feature 4 Feature Visualization Le, et al., Building high-level features using large-scale unsupervised learning. ICML 2012 Feature 1

Pooling Size = 5 Number of maps = 8 Image Size = 200 Number of output channels = 8 Number of input channels = 3 One layer RF size = 18 Input to another layer above (image with 8 channels) W H LCN Size = 5 Feature 6 Feature 5 Feature 7 Feature 8 Le, et al., Building high-level features using large-scale unsupervised learning. ICML 2012 Feature Visualization

ImageNet classification 22,000 categories 14,000,000 images Hand-engineered features (SIFT, HOG, LBP), Spatial pyramid, SparseCoding/Compression Le, et al., Building high-level features using large-scale unsupervised learning. ICML 2012

x1x1 x2x2 x3x3 x4x4 a3a3 a2a2 a1a1 Input to a 22,000-way classifier

Using only 1000 categories, our method > 60% 0.005% Random guess 9.5% State-of-the-art (Weston, Bengio ‘11) 18.3% Feature learning From raw pixels Le, et al., Building high-level features using large-scale unsupervised learning. ICML 2012

Indian elephantAfrican elephant Cassette playerTape player Malaria mosquitoYellow fever mosquito

People / PlungerSwimming / Person / Swim trunk / Snorkel Person / People / Pingpong / Wheel / … / Ping-pong ball People / Tree / Street / Marching order /… Bearskin

Seat-beltBoston rocker ArcheryShredder Quoc V. Le

Amusement, Park Face Hammock Quoc V. Le

Dean, et al., Large scale distributed deep networks. NIPS 2012.

Theoretical questions -Properties of local minima and generalization -Role of unsupervised pretraining -Better weight initialization -Nonlinearities and invariance properties Quoc V. Le

Q.V. Le, M.A. Ranzato, R. Monga, M. Devin, G. Corrado, K. Chen, J. Dean, A.Y. Ng. Building high-level features using large-scale unsupervised learning. ICML, Q.V. Le, J. Ngiam, Z. Chen, D. Chia, P. Koh, A.Y. Ng. Tiled Convolutional Neural Networks. NIPS, Q.V. Le, W.Y. Zou, S.Y. Yeung, A.Y. Ng. Learning hierarchical spatio-temporal features for action recognition with independent subspace analysis. CVPR, Q.V. Le, T. Sarlos, A. Smola. Fastfood – Approximate nonlinear expansions in loglinear time. ICML, 2013 Q.V. Le, J. Ngiam, A. Coates, A. Lahiri, B. Prochnow, A.Y. Ng. On optimization methods for deep learning. ICML, Q.V. Le, A. Karpenko, J. Ngiam, A.Y. Ng. ICA with Reconstruction Cost for Efficient Overcomplete Feature Learning. NIPS, Q.V. Le, J. Han, J. Gray, P. Spellman, A. Borowsky, B. Parvin. Learning Invariant Features for Tumor Signatures. ISBI, I.J. Goodfellow, Q.V. Le, A.M. Saxe, H. Lee, A.Y. Ng, Measuring invariances in deep networks. NIPS, References