Restricted Boltzmann Machines for Classification

Slides:



Advertisements
Similar presentations
Greedy Layer-Wise Training of Deep Networks
Advertisements

Thomas Trappenberg Autonomous Robotics: Supervised and unsupervised learning.
Deep Learning Bing-Chen Tsai 1/21.
Stochastic Neural Networks Deep Learning and Neural Nets Spring 2015.
CS590M 2008 Fall: Paper Presentation
Advanced topics.
Rajat Raina Honglak Lee, Roger Grosse Alexis Battle, Chaitanya Ekanadham, Helen Kwong, Benjamin Packer, Narut Sereewattanawoot Andrew Y. Ng Stanford University.
Nathan Wiebe, Ashish Kapoor and Krysta Svore Microsoft Research ASCR Workshop Washington DC Quantum Deep Learning.
Stacking RBMs and Auto-encoders for Deep Architectures References:[Bengio, 2009], [Vincent et al., 2008] 2011/03/03 강병곤.
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
What is the Best Multi-Stage Architecture for Object Recognition? Ruiwen Wu [1] Jarrett, Kevin, et al. "What is the best multi-stage architecture for object.
Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009
Deep Learning.
Neural NetworksNN 11 Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Deep Belief Networks for Spam Filtering
Distributed Representations of Sentences and Documents
AN ANALYSIS OF SINGLE- LAYER NETWORKS IN UNSUPERVISED FEATURE LEARNING [1] Yani Chen 10/14/
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Deep Boltzman machines Paper by : R. Salakhutdinov, G. Hinton Presenter : Roozbeh Gholizadeh.
Nantes Machine Learning Meet-up 2 February 2015 Stefan Knerr CogniTalk
Using Fast Weights to Improve Persistent Contrastive Divergence Tijmen Tieleman Geoffrey Hinton Department of Computer Science, University of Toronto ICML.
Varieties of Helmholtz Machine Peter Dayan and Geoffrey E. Hinton, Neural Networks, Vol. 9, No. 8, pp , 1996.
Training Restricted Boltzmann Machines using Approximations to the Likelihood Gradient Tijmen Tieleman University of Toronto (Training MRFs using new algorithm.
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Dr. Z. R. Ghassabi Spring 2015 Deep learning for Human action Recognition 1.
CSC 2535 Lecture 8 Products of Experts Geoffrey Hinton.
Training Restricted Boltzmann Machines using Approximations to the Likelihood Gradient Tijmen Tieleman University of Toronto.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 18 Learning Boltzmann Machines Geoffrey Hinton.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 19: Learning Restricted Boltzmann Machines Geoffrey Hinton.
Introduction to Deep Learning
Cognitive models for emotion recognition: Big Data and Deep Learning
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Deep learning Tsai bing-chen 10/22.
Neural Networks William Cohen [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ]
CSC321 Lecture 24 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
CSC321 Lecture 27 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Deep Learning Primer Swadhin Pradhan Reading Group Presentation 03/30/2016, UT Austin.
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
Learning Deep Generative Models by Ruslan Salakhutdinov
Convolutional Neural Network
Deep Learning Amin Sobhani.
Energy models and Deep Belief Networks
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
Article Review Todd Hricik.
Matt Gormley Lecture 16 October 24, 2016
LECTURE ??: DEEP LEARNING
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Multimodal Learning with Deep Boltzmann Machines
Intelligent Information System Lab
Classification of Hand-Written Digits Using Scattering Convolutional Network Dongmian Zou Advisor: Professor Radu Balan.
Deep learning and applications to Natural language processing
Deep Learning Qing LU, Siyuan CAO.
Structure learning with deep autoencoders
Unsupervised Learning and Autoencoders
Deep Learning Workshop
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Restricted Boltzman Machines
Department of Electrical and Computer Engineering
Deep Architectures for Artificial Intelligence
ECE 599/692 – Deep Learning Lecture 9 – Autoencoder (AE)
Deep Belief Nets and Ising Model-Based Network Construction
[Figure taken from googleblog
龙星计划课程-深度学习 天津大学 7月2日-7月5日.
CSC321 Winter 2007 Lecture 21: Some Demonstrations of Restricted Boltzmann Machines Geoffrey Hinton.
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Autoencoders David Dohan.
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Restricted Boltzmann Machines for Classification Jin Mao Postdoc, School of Information, University of Arizona Nov. 3rd, 2016

Outline Restricted Boltzmann Machines RBM For Classification Semi-supervised Learning Experiments

Restricted Boltzmann Machines General RBM Boltzmann Machines (BMs) are a particular form of log-linear Markov Random Field (MRF), i.e., the energy function is linear in its parameters. Restrict: no visible unit is connected to any other visible unit and no hidden unit is connected to any other hidden unit.

Restricted Boltzmann Machines General RBM The energy function for x: The data negative log-likelihood gradient: log-linear Z

Restricted Boltzmann Machines Learning Weights Set the states of the visible units according to a training example Next, update the states of the hidden units For each hidden unit, compute its activation energy For each edge, compute positive weight for each edge Positive(e_ij)

Restricted Boltzmann Machines Learning Weights Now reconstruct the visible units Update the weight of each edge For each visible unit, compute its activation energy For each edge, compute negative weight for each edge Negative(e_ij) measures the association that the network itself generates (or “daydreams” about) when no units are fixed to training data.

Restricted Boltzmann Machines Learning Weights Test convergence by KL-Divergence

Restricted Boltzmann Machines Application The power lies in the hidden layer In classification problems as feature extractors (Gehler et al., 2006) as a good initial training phase for deep neural network classifiers (Hinton,2007).

RBM For Classification Generative Model Objective: Energy Function: Gradient: Activation Functions:

RBM For Classification Discriminative Model Objective: Energy Function: Gradient:

RBM For Classification Hybrid Model Objective:

Semi-supervised Learning There are few labeled training data but many unlabeled examples of inputs. Semi-supervised learning algorithms (Chapelle et al., 2006) address this situation by using the unlabeled data to introduce constraints on the trained model. In the RBM framework, a natural constraint is the model should be a good generative model of the unlabeled data Unsupervised objectives:

Semi-supervised Learning Combine unsupervised objective with supervised objective

Experiments Model Selection Experiments on two classification problems: character recognition and text classification. In all experiments, we performed model selection on a validation set before testing. For the different RBM models, model selection consisted in finding good values for the learning rate, the size of the hidden layer n and good weights for the different types of learning (generative and semi-supervised weights).

Experiments Model Selection Document Classification Character Recognition Document Classification

Reference Other materials This presentation is from: Larochelle, H., & Bengio, Y. (2008, July). Classification using discriminative restricted Boltzmann machines. In Proceedings of the 25th international conference on Machine learning (pp. 536-543). ACM. Other materials https://deeplearning4j.org/restrictedboltzmannmachine http://deeplearning.net/tutorial/rbm.html http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann-machines/ Hinton, G. (2010). A practical guide to training restricted Boltzmann machines. Momentum, 9(1), 926.

Thank you!