Cortical Receptive Fields Using Deep Autoencoders Work done as a part of CS397 Ankit Awasthi (Y8084) Supervisor: Prof. H. Karnick.

Slides:



Advertisements
Similar presentations
Greedy Layer-Wise Training of Deep Networks
Advertisements

Deep Learning Bing-Chen Tsai 1/21.
Chapter 2.
Stochastic Neural Networks Deep Learning and Neural Nets Spring 2015.
CS590M 2008 Fall: Paper Presentation
Advanced topics.
Stacking RBMs and Auto-encoders for Deep Architectures References:[Bengio, 2009], [Vincent et al., 2008] 2011/03/03 강병곤.
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
Un Supervised Learning & Self Organizing Maps. Un Supervised Competitive Learning In Hebbian networks, all neurons can fire at the same time Competitive.
Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009
Learning Convolutional Feature Hierarchies for Visual Recognition
Deep Learning.
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Unsupervised Learning With Neural Nets Deep Learning and Neural Nets Spring 2015.
COGNITIVE NEUROSCIENCE
AN ANALYSIS OF SINGLE- LAYER NETWORKS IN UNSUPERVISED FEATURE LEARNING [1] Yani Chen 10/14/
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Submitted by:Supervised by: Ankit Bhutani Prof. Amitabha Mukerjee (Y )Prof. K S Venkatesh.
Overview of Back Propagation Algorithm
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Multiclass object recognition
CIAR Second Summer School Tutorial Lecture 2b Autoencoders & Modeling time series with Boltzmann machines Geoffrey Hinton.
How to do backpropagation in a brain
Decoding Human Face Processing Ankit Awasthi Prof. Harish Karnick.
CSC2535: Computation in Neural Networks Lecture 11: Conditional Random Fields Geoffrey Hinton.
Low Level Visual Processing. Information Maximization in the Retina Hypothesis: ganglion cells try to transmit as much information as possible about the.
A shallow introduction to Deep Learning
2 2  Background  Vision in Human Brain  Efficient Coding Theory  Motivation  Natural Pictures  Methodology  Statistical Characteristics  Models.
Varieties of Helmholtz Machine Peter Dayan and Geoffrey E. Hinton, Neural Networks, Vol. 9, No. 8, pp , 1996.
Learning Lateral Connections between Hidden Units Geoffrey Hinton University of Toronto in collaboration with Kejie Bao University of Toronto.
Presented by: Mingyuan Zhou Duke University, ECE June 17, 2011
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Projects: 1.Predictive coding in balanced spiking networks (Erwan Ledoux). 2.Using Canonical Correlation Analysis (CCA) to analyse neural data (David Schulz).
Geoffrey Hinton CSC2535: 2013 Lecture 5 Deep Boltzmann Machines.
Chapter 3: Neural Processing and Perception. Neural Processing and Perception Neural processing is the interaction of signals in many neurons.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 18 Learning Boltzmann Machines Geoffrey Hinton.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 19: Learning Restricted Boltzmann Machines Geoffrey Hinton.
Boltzman Machines Stochastic Hopfield Machines Lectures 11e 1.
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
Introduction to Deep Learning
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 6: Applying backpropagation to shape recognition Geoffrey Hinton.
Cognitive models for emotion recognition: Big Data and Deep Learning
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.
Deep learning Tsai bing-chen 10/22.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Neural Networks William Cohen [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ]
CSC321 Lecture 24 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
Statistical Modeling and Learning in Vision --- cortex-like generative models Ying Nian Wu UCLA Department of Statistics JSM, August 2010.
CSC321 Lecture 27 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
1 Nonlinear models for Natural Image Statistics Urs Köster & Aapo Hyvärinen University of Helsinki.
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
Learning Deep Generative Models by Ruslan Salakhutdinov
Energy models and Deep Belief Networks
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
Article Review Todd Hricik.
Restricted Boltzmann Machines for Classification
Multimodal Learning with Deep Boltzmann Machines
Deep Learning Qing LU, Siyuan CAO.
Structure learning with deep autoencoders
Deep learning Introduction Classes of Deep Learning Networks
CSC321 Winter 2007 Lecture 21: Some Demonstrations of Restricted Boltzmann Machines Geoffrey Hinton.
Autoencoders Supervised learning uses explicit labels/correct output in order to train a network. E.g., classification of images. Unsupervised learning.
by Khaled Nasr, Pooja Viswanathan, and Andreas Nieder
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Cortical Receptive Fields Using Deep Autoencoders Work done as a part of CS397 Ankit Awasthi (Y8084) Supervisor: Prof. H. Karnick

The visual pathway

Some Terms A cell, neuron, neural unit, unit of the neural network may be used interchangeably and all refer to neuron in the visual cortex. Receptive field of a neuron refers to the region in space in which the presence of a stimulus will change the response of the neuron.

Precortical stages Retinal cells and the cells in the LGN are center on surround off and vice versa

Why focus on cortical perception?? Most cells in the precortical stages are hard- coded to a large extent and may be innate Cortical cells are mostly learned through postnatal visual stimulation Hubel Wiesel showed that irreversible damage was produced in kittens by sufficient visual deprivation during the so-called critical period

What are these???

How did you do that?? Surely you did not use only visual information Processing in the later stages of visual cortex has some top-down influence Much of the visual inference involves input from other modalities (say facial emotion recognition) Thus we focus only on those stages of processing which require/use only visual information

Neurological Findings (with electrodes in cat’s cortex !!) The visual cortex consists of simple and complex cells Simple cells can be characterized by a certain distributions of on and off areas Complex cells could not be explained with a simple distribution of on and off areas Receptive fields for simple cells should look like oriented edge detectors Receptive fields of different cells may be overlapping

Topographic Representation There is a systematic mapping of each structure to the next The optic fibers from a part of the retina are connected to a small part in LGN A part of LGN is similarly connected to a small part in the primary visual cortex This topography continues in other cortical regions Convergence at each stage Larger receptive fields in later stages

How do we learn these layers then??

Why Deep Learning?? Brains have a deep architecture Humans organize their ideas hierarchically, through composition of simpler ideas Insufficiently deep architectures can be exponentially inefficient Deep architectures facilitate feature and sub- feature sharing

Neural Networks (~1985) input vector hidden layers outputs Back-propagate error signal to get derivatives for learning Compare outputs with correct answer to get error signal

Restricted Boltzmann Machines (RBM) We restrict the connectivity to make learning easier. – Only one layer of hidden units. – No connections between hidden units. Energy of a joint configuration is defined as (for binary visible units) (for real visible units) Hidden(h) i j Visible(v)

Restricted Boltzmann Machines (RBM) (contd.) Probability of a configuration is defined as The hidden nodes are conditionally independent given the visible layer and vice versa Using the definition of the energy function and probability, the conditional probabilities come out to be as follows

Maximum likelihood learning for an RBM i j i j i j i j t = 0 t = 1 t = 2 t = infinity Start with a training vector on the visible units. Then alternate between updating all the hidden units in parallel and updating all the visible units in parallel. a fantasy

Training a deep network

Deep Autoencoder(Hinton et.al. 2006)

Sparse DBNs(Lee at. al. 2007) In order to have a sparse hidden layer, the average activation of a hidden unit over the training is constrained to a certain small quantity The optimization problem in the learning algorithm would look like

Related Work In [4], the authors had shown the features learned by independent component analysis were oriented edge detectors Lee et. al. in [10] show that the second layer learned using sparse DBNs match certain properties of cells in V2 area of the visual cortex Bengio et.al. in [3] discuss ways of visualizinng higher layer features Lee et.al. in [4] have come up with convolutional DBNs which incorporates weight sharing across the visual field and probablistic max pooling operation

Our experimental setting We trained sparse DBNs on 100,000 randomly sampled patches of natural images of size 14x14 The image were preprocessed to have same overall contrast and whitened as in [5] The hidden units in the first, second, third layer are all 200 in number

Getting first layer hidden features To maximize the activation of the i th hidden unit, the input v should be Recall what was said about receptive fields of simple cells (oriented edge detectors)

First Hidden Layer Features (with each epoch)

Effect of Sparsity

Higher Layer Features Projecting the a higher layer's weights onto the response of the previous layer…..useless!!! Three different methods of projecting the hidden units onto the input space – Linear Combination of Previous layer filters, Lee et.al. [2] – Sampling from a hidden unit, Hinton et.al. [5], Bengio et.al.[3] – Activation Maximization, Bengio et.al. [3]

Linear Combination of Previous Layer Filters Only few connections to the previous layer have their weights either too high or too low Some of the largest weighted connections are used for linear combination Overlooks the non-linearity in the network from one layer to the other Simple and efficient

Linear Combination of Previous Layer Filters (Results)

Sampling from Hidden Units Deep Autoencoder ( using RBMs ) is a generative model ( top down sampling) and any two adjacent layers form a RBM ( Gibbs sampling) Clamp a particular hidden unit to 1 during Gibbs sampling and then do a top down sampling to the input layer

Sampling from Hidden Units (Results)

Activation Maximization Intuition same as that for first layer features Optimization problem is much more difficult In general a non convex problem Solve for local minima for different random initializations, then take average or the minimum etc.

Activation Maximization (Results)

Analysis of Results As observed, the second layer features are able to capture a combination of edges or angled stimuli The third layer features are very difficult to make sense of in terms of simple geometrical elements No good characterization of these cells is available, thus not to choose between the different methods

Larger Receptive Fields for Higher Layers We offer a simple solution to extend the size of the receptive fields for higher layers Using the RBM trained on natural image patches, compute the response over the entire image with overlapping patches Responses of some neighboring patches are taken as input for the next layer RBM This is repeated for the whole network This has not been investigated exhaustively

Results(linear combination) First Layer Second Layer

Conclusion and Future Work Similarities in the receptive fields Support for Deep Learning Methods as computational model for cortical processing Able to learn more complete parts of objects in the higher layers with bigger receptive fields Future work would be to extend these ideas and establish the cognitive relevance of the computational models

References 1.Georey E. Hinton, Yee-Whye Teh and Simon Osindero, A Fast Learning Algorithm forDeep Belief Nets. Neural Computation, pages , Volume 18, D. H. Hubel & T. N. Wiesel Jiro Gyoba, Recpetive Fields, Binocular Interaction And Functional Architecture In The Cat's Visual Cortex,The Journal of Physiology, Vol. 16;0, No. 1., Michael J. Lyons, Shigeru Akamatsu, Miyuki Kamachi & Jiro Gyoba, Coding Facial Expressions with Gabor Wavelets,Proceedings, Third IEEE International Conference on Automatic Face and Gesture Recognition,pp , 1-19.April Hateren, J. H. van and Schaaf, A. van der, Independent Component Filters of Natural Images Compared with Simple Cells in Primary Visual Cortex,Proceedings: Biological Sciences, vol 265,pages , March Georey E. Hinton (2010). A Practical Guide to Training Restricted Boltzmann Machines,Technical Report,Volume 1 6.Dumitru Erhan, Yoshua Bengio, Aaron Courville, and Pascal Vincent (2010). Visualizing Higher-Layer Features of a Deep Network,Technical Report 1341

References(contd.) 7. Honglak Lee, Roger Grosse,Rajesh Ranganath, Andrew Y. Ng. Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations,ICML Geoffrey E. Hinton Learning multiple layers of representation,Trends in Cognitive Sciences Vol.11 No.10, Andrew Ng. (2010). Sparse Autoencoder(lecture notes). 10.Honglak Lee, Chaitanya Ekanadham, Andrew Y. Ng, Sparse deep belief net model for visual area V2NIPS, Ruslan Salakhutdinov Learning Deep Generative Models Phd thesis, Yoshua Bengio Learning Deep Architectures for AI Foundations and Trends in Machine Learning,Vol. 2, No. 1, 1127, 2009