Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009

Slides:



Advertisements
Similar presentations
Deep Belief Nets and Restricted Boltzmann Machines
Advertisements

Deep Learning Bing-Chen Tsai 1/21.
Object recognition and scene “understanding”
CIAR Second Summer School Tutorial Lecture 2a Learning a Deep Belief Net Geoffrey Hinton.
Stochastic Neural Networks Deep Learning and Neural Nets Spring 2015.
CS590M 2008 Fall: Paper Presentation
Advanced topics.
Rajat Raina Honglak Lee, Roger Grosse Alexis Battle, Chaitanya Ekanadham, Helen Kwong, Benjamin Packer, Narut Sereewattanawoot Andrew Y. Ng Stanford University.
Nathan Wiebe, Ashish Kapoor and Krysta Svore Microsoft Research ASCR Workshop Washington DC Quantum Deep Learning.
Stacking RBMs and Auto-encoders for Deep Architectures References:[Bengio, 2009], [Vincent et al., 2008] 2011/03/03 강병곤.
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
POSTER TEMPLATE BY: Multi-Sensor Health Diagnosis Using Deep Belief Network Based State Classification Prasanna Tamilselvan.
Learning Deep Energy Models
Learning Convolutional Feature Hierarchies for Visual Recognition
Deep Learning.
How to do backpropagation in a brain
Deep Belief Networks for Spam Filtering
AN ANALYSIS OF SINGLE- LAYER NETWORKS IN UNSUPERVISED FEATURE LEARNING [1] Yani Chen 10/14/
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Submitted by:Supervised by: Ankit Bhutani Prof. Amitabha Mukerjee (Y )Prof. K S Venkatesh.
Overview of Back Propagation Algorithm
Comp 5013 Deep Learning Architectures Daniel L. Silver March,
Multiclass object recognition
Nantes Machine Learning Meet-up 2 February 2015 Stefan Knerr CogniTalk
Playing with features for learning and prediction Jongmin Kim Seoul National University.
How to do backpropagation in a brain
Decoding Human Face Processing Ankit Awasthi Prof. Harish Karnick.
Boltzmann Machines and their Extensions S. M. Ali Eslami Nicolas Heess John Winn March 2013 Heriott-Watt University.
Cortical Receptive Fields Using Deep Autoencoders Work done as a part of CS397 Ankit Awasthi (Y8084) Supervisor: Prof. H. Karnick.
Using Fast Weights to Improve Persistent Contrastive Divergence Tijmen Tieleman Geoffrey Hinton Department of Computer Science, University of Toronto ICML.
A shallow introduction to Deep Learning
Varieties of Helmholtz Machine Peter Dayan and Geoffrey E. Hinton, Neural Networks, Vol. 9, No. 8, pp , 1996.
Large-scale Deep Unsupervised Learning using Graphics Processors
Presented by: Mingyuan Zhou Duke University, ECE June 17, 2011
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Dr. Z. R. Ghassabi Spring 2015 Deep learning for Human action Recognition 1.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 19: Learning Restricted Boltzmann Machines Geoffrey Hinton.
Towards Total Scene Understanding: Classification, Annotation and Segmentation in an Automatic Framework N 工科所 錢雅馨 2011/01/16 Li-Jia Li, Richard.
Introduction to Deep Learning
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Sum-Product Networks: A New Deep Architecture
Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.
Deep learning Tsai bing-chen 10/22.
CSC2515 Fall 2008 Introduction to Machine Learning Lecture 8 Deep Belief Nets All lecture slides will be available as.ppt,.ps, &.htm at
Neural Networks William Cohen [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ]
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
Learning Deep Generative Models by Ruslan Salakhutdinov
Convolutional Neural Network
Energy models and Deep Belief Networks
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
ECE 5424: Introduction to Machine Learning
Learning Mid-Level Features For Recognition
Article Review Todd Hricik.
Restricted Boltzmann Machines for Classification
Multimodal Learning with Deep Boltzmann Machines
Deep learning and applications to Natural language processing
Deep Learning Qing LU, Siyuan CAO.
Deep Belief Networks Psychology 209 February 22, 2013.
Deep Learning Workshop
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Computer Vision James Hays
Towards Understanding the Invertibility of Convolutional Neural Networks Anna C. Gilbert1, Yi Zhang1, Kibok Lee1, Yuting Zhang1, Honglak Lee1,2 1University.
Convolutional Neural Networks
المشرف د.يــــاســـــــــر فـــــــؤاد By: ahmed badrealldeen
Visualizing and Understanding Convolutional Networks
CSC321 Winter 2007 Lecture 21: Some Demonstrations of Restricted Boltzmann Machines Geoffrey Hinton.
Presentation transcript:

Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009 Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and Andrew Y. Ng ICML 2009 Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009

Outline Motivations Contributions Backgrounds Algorithms Experiment results Deep Vs Shallow Conclusions

Motivations To Learn hierarchical models which simultaneously represent multiple levels, e.g., pixel intensities, edges, object parts, objects, and beyond can be represented by layers from low to high. Combining top-down and bottom-up processing of an image. Limitations of deep belief networks (DBNs) Scaling DBNs to realistic-size images remains challenging: images are high-dimentional and objects can appear at arbitrary locations in images.

Contributions Convolutional RBM: feature detectors are shared among all locations in an image. Probabilistic max-pooling: in a probabilistic sound way allowing higher-layer units to cover larger areas of the input. The first translation invariant hierarchical generative model supporting both top-down and bottom-up probabilistic inference and sales to realistic image sizes.

Backgrounds: Restricted Boltzmann Machine (RBM) (binary v) (real-value v) Giving the visible layer, the hidden units are conditionally independent, and vise versa. Efficient block Gibbs sampling can be performed by alternately sampling each layer’s units. Computing the exact gradient of the log-likelihood is intractable, so the contrastive divergence approximation is commonly used.

Backgrounds: Deep belief network (DBN) In a DBN, two adjacent layers have a full set of connections between them, but no two units in the same layer are connected. A DBN can be formed by stacking RBMs. An efficient algorithm for training DBNs (Hinton et al., 2006): greedily training each layer, from lowest to highest, as an RBM using the previous layer's activations as inputs.

Algorithms: Convolutional RBM (CRBM)

Algorithms: Probabilistic max-pooling

Algorithms: Probabilistic max-pooling Each unit in a pooling layer computes the maximum activation of the units in a small region of the detection layer. Shrinking the representation with max-pooling allows higher-layer representations to be invariant to small translations of the input and reduces the computational burden. Max-pooling was intended only for feed-forward architectures. A generative model of images which supports both top-down and bottom-up inference is of interest.

Algorithms: Sparsity regulations Only a tiny fraction of the units should be active in relation to a given stimulus. Regularizing the objective function to encourage each of the hidden units to have a mean activation close to some small constant .

Algorithms: Convolutional DBN (CDBN) CDBN consists of several max-pooling-CRBMs stacked on top of one another. Once a given layer is trained, its weights are frozen, and its activations are used as input to the next layer.

Hierarchical probabilistic inference

Experimental Results: natural images

Experimental Results: image classification

Experimental Results: unsupervised learning of object parts

Experimental Results: Hierarchical probabilistic inference

Deep Vs Shallow . From Jason Weston’s slides: DEEP LEARNING VIA SEMI-SUPERVISED EMBEDDING, ICML 2009 WORKSHOP ON LEARNING FEATURE HIERARCHIES From Francis Bach’s slides:  Convex sparse methods for feature hierarchies, ICML 2009 WORKSHOP ON LEARNING FEATURE HIERARCHIES

Conclusions Convolutional deep belief network: A scalable generative model for learning hierarchical representations from unlabeled images. Performing well in a variety of visual recognition tasks.