Restricted Boltzman Machines

Slides:



Advertisements
Similar presentations
Greedy Layer-Wise Training of Deep Networks
Advertisements

Deep Learning Bing-Chen Tsai 1/21.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Stochastic Neural Networks Deep Learning and Neural Nets Spring 2015.
CS590M 2008 Fall: Paper Presentation
CS 678 –Boltzmann Machines1 Boltzmann Machine Relaxation net with visible and hidden units Learning algorithm Avoids local minima (and speeds up learning)
Tuomas Sandholm Carnegie Mellon University Computer Science Department
Comp 3503 / 5013 Dynamic Neural Networks Daniel L. Silver March,
Deep Learning.
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
1 Neural networks 3. 2 Hopfield network (HN) model A Hopfield network is a form of recurrent artificial neural network invented by John Hopfield in 1982.
Deep Belief Networks for Spam Filtering
Associative Learning.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Deep Boltzman machines Paper by : R. Salakhutdinov, G. Hinton Presenter : Roozbeh Gholizadeh.
Comp 5013 Deep Learning Architectures Daniel L. Silver March,
Using Fast Weights to Improve Persistent Contrastive Divergence Tijmen Tieleman Geoffrey Hinton Department of Computer Science, University of Toronto ICML.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
Learning Lateral Connections between Hidden Units Geoffrey Hinton University of Toronto in collaboration with Kejie Bao University of Toronto.
Geoffrey Hinton CSC2535: 2013 Lecture 5 Deep Boltzmann Machines.
Training Restricted Boltzmann Machines using Approximations to the Likelihood Gradient Tijmen Tieleman University of Toronto.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 18 Learning Boltzmann Machines Geoffrey Hinton.
CIAR Summer School Tutorial Lecture 1b Sigmoid Belief Nets Geoffrey Hinton.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 19: Learning Restricted Boltzmann Machines Geoffrey Hinton.
Boltzman Machines Stochastic Hopfield Machines Lectures 11e 1.
Deep learning Tsai bing-chen 10/22.
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
1 Restricted Boltzmann Machines and Applications Pattern Recognition (IC6304) [Presentation Date: ] [ Ph.D Candidate,
Today’s Lecture Neural networks Training
Neural networks.
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
Introduction to Neural Networks
Learning Deep Generative Models by Ruslan Salakhutdinov
Deep Learning Amin Sobhani.
Energy models and Deep Belief Networks
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
ECE 471/571 - Lecture 15 Hopfield Network 03/29/17.
Real Neurons Cell structures Cell body Dendrites Axon
Restricted Boltzmann Machines for Classification
LECTURE ??: DEEP LEARNING
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Multimodal Learning with Deep Boltzmann Machines
CSE 473 Introduction to Artificial Intelligence Neural Networks
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Deep Learning Qing LU, Siyuan CAO.
Structure learning with deep autoencoders
Unsupervised Learning and Autoencoders
Deep Learning Workshop
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CSE P573 Applications of Artificial Intelligence Neural Networks
Department of Electrical and Computer Engineering
ECE 471/571 - Lecture 19 Hopfield Network.
Lecture 11. MLP (III): Back-Propagation
ANN Design and Training
Recurrent Neural Networks
Deep Architectures for Artificial Intelligence
Deep Belief Nets and Ising Model-Based Network Construction
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CSE 573 Introduction to Artificial Intelligence Neural Networks
Artificial Neural Networks
Regulation Analysis using Restricted Boltzmann Machines
Neural Networks Geoff Hulten.
Boltzmann Machine (BM) (§6.4)
CSC321 Winter 2007 Lecture 21: Some Demonstrations of Restricted Boltzmann Machines Geoffrey Hinton.
COSC 4335: Part2: Other Classification Techniques
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
CSC 578 Neural Networks and Deep Learning
Artificial Neural Networks / Spring 2002
Presentation transcript:

Restricted Boltzman Machines with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017 12/11/2018 Deep Learning Workshop

Johann von Neumann’s Opinion “All of this will lead to theories [of computation] which are much less rigidly of an all-or-none nature than past and present a formal logic …. There are numerous indications to make us believe that this new system of formal logic will move closer to another discipline which has been little linked in the past with logic. This is thermodynamics, primarily in the form it was received from Boltzmann, and is that part of theoretical physics which comes nearest in some of its aspects to manipulating and measuring information.” John Von Neumann, Collected Works, Vol.5, p.304

Boltzmann Machine (BM) An unsupervised learning system A type of stochastic recurrent neural network (form of Markov Random Field) invented by Geoffrey Hinton and Terry Sejnowski in 1985 Can be seen as the stochastic, generative counterpart of Hopfield nets Based on the mathematics of thermodynamics Complex training algorithm that does not scale well (grows exponential with number of nodes) 12/11/2018 Deep Learning Workshop

Boltzmann Machine (BM) BM is a type of stochastic recurrent neural network All units (neurons) are binary & stochastic Characteristics Unsupervised learning Associative memory Long training time Energy function

Restricted Boltzmann Machine (RBM) A generative stochastic artificial neural network that can learn a probability distribution over its set of inputs. Restrict connectivity of network Undirected bipartite graphical model Restrict number of iterations in + and – phases

Restricted Boltzmann Machine (RBM) Does not allow intra-layer connections (bipartite architecture) Learning both recognition model and generative model

Deep Learning Workshop RBM 12/11/2018 Deep Learning Workshop

Restricted Boltzman Machine hj pj=1/(1-e-Σj) Recall = Relaxation Σj=wijvi j wij vo or ho i Σi=wijhj vi pi=1/(1-e-Σi) SF/Fantasy Oscar Winner Source: http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann-machines/

Restricted Boltzman Machine hj pj=1/(1-e-Σj) Recall = Relaxation Σj=wijvi j wij vo or ho i Σi=wijhj vi pi=1/(1-e-Σi) SF/Fantasy Oscar Winner Source: http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann-machines/

Restricted Boltzman Machine hj pj=1/(1-e-Σj) Recall = Relaxation SF/Fantasy Oscar Winner Σj=wijvi j wij vo or ho i Σi=wijhj vi pi=1/(1-e-Σi) SF/Fantasy Oscar Winner Source: http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann-machines/

Restricted Boltzman Machine hj pj=1/(1-e-Σj) Recall = Relaxation SF/Fantasy Oscar Winner Σj=wijvi j wij vo or ho i Σi=wijhj vi pi=1/(1-e-Σi) SF/Fantasy Oscar Winner Source: http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann-machines/

Restricted Boltzman Machine hj pj=1/(1-e-Σj) Learning = ~ Gradient Descent = Constrastive Divergence Update hidden units Σj=wijvi j P=P+vihj vo or ho i Σi=wijhj vi pi=1/(1-e-Σi) SF/Fantasy Oscar Winner Source: http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann-machines/

Restricted Boltzman Machine hj pj=1/(1-e-Σj) Learning = ~ Gradient Descent = Constrastive Divergence Reconstruct visible units j Σj=wijvi vo or ho i Σi=wijhj vi pi=1/(1-e-Σi) SF/Fantasy Oscar Winner Source: http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann-machines/

Restricted Boltzman Machine hj pj=1/(1-e-Σj) Learning = ~ Gradient Descent = Constrastive Divergence Reupdate hidden units Σj=wijvi j N=N+vihj vo or ho i Σi=wijhj vi pi=1/(1-e-Σi) SF/Fantasy Oscar Winner Source: http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann-machines/

Restricted Boltzman Machine hj pj=1/(1-e-Σj) Learning = ~ Gradient Descent = Constrastive Divergence Update weights Σj=wijvi j wij=wij +ηΔwij vo or ho Δwij=<P>-<N> i Σi=wijhj vi pi=1/(1-e-Σi) SF/Fantasy Oscar Winner Source: http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann-machines/

Why does this weight update agorithm work? In the first phase, vihj measures the association between the ith and jth unit that we want the network to learn from our training examples In the “reconstruction” phase, where the RBM generates the states of visible units based on its hypotheses about the hidden units alone,  vihj  measures the association that the network itself generates (or “daydreams” about) when no units are fixed to training data So by adding <P>-<N> to each weight, we are helping the network’s daydreams better match the reality of our training examples This performs an approximate gradient descent on the derivative of the reconstruction error with respect to the weights 12/11/2018 Deep Learning Workshop

Multi-modal Deep Learning 8 eight

Multi-modal Deep Learning DEMO 8 eight

Restricted Boltzman Machine RBM Overview: http://blog.echen.me/2011/07/18/introduction-to-restricted-boltzmann- machines/ Wikipedia on DLA and RBM: http://en.wikipedia.org/wiki/Deep_learning RBM Details and Code: http://www.deeplearning.net/tutorial/rbm.html

Restricted Boltzman Machine Geoff Hinton on RBMs: RBMs and Constrastive Divergence Algorithm http://www.youtube.com/watch?v=fJjkHAuW0Yk An example of RBM Learning http://www.youtube.com/watch?v=Ivj7jymShN0 RBMs applied to Collaborative Filtering http://www.youtube.com/watch?v=laVC6WFIXjg

Additional References Coursera course – Neural Networks fro Machine Learning: https://class.coursera.org/neuralnets-2012-001/lecture ML: Hottest Tech Trend in next 3-5 Years http://www.youtube.com/watch?v=b4zr9Zx5WiE

TUTORIAL 6 Develop and train a RBM network (Python code) Rbm.py