Cognitive models for emotion recognition: Big Data and Deep Learning

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS Zemel
Greedy Layer-Wise Training of Deep Networks
Deep Learning Bing-Chen Tsai 1/21.
CIAR Second Summer School Tutorial Lecture 2a Learning a Deep Belief Net Geoffrey Hinton.
Stochastic Neural Networks Deep Learning and Neural Nets Spring 2015.
CS590M 2008 Fall: Paper Presentation
Advanced topics.
Stacking RBMs and Auto-encoders for Deep Architectures References:[Bengio, 2009], [Vincent et al., 2008] 2011/03/03 강병곤.
What kind of a Graphical Model is the Brain?
Supervised Learning Recap
Automatic Speech Recognition II  Hidden Markov Models  Neural Network.
Introduction The aim the project is to analyse non real time EEG (Electroencephalogram) signal using different mathematical models in Matlab to predict.
Deep Learning.
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Artificial Intelligence (CS 461D)
Neural NetworksNN 11 Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
How to do backpropagation in a brain
Deep Belief Networks for Spam Filtering
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Deep Boltzman machines Paper by : R. Salakhutdinov, G. Hinton Presenter : Roozbeh Gholizadeh.
Document Classification using Deep Belief Nets Lawrence McAfee 6/9/08 CS224n, Sprint ‘08.
Can computer simulations of the brain allow us to see into the mind? Geoffrey Hinton Canadian Institute for Advanced Research & University of Toronto.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
How to do backpropagation in a brain
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Functional Brain Signal Processing: EEG & fMRI Lesson 8 Kaushik Majumdar Indian Statistical Institute Bangalore Center M.Tech.
Using Fast Weights to Improve Persistent Contrastive Divergence Tijmen Tieleman Geoffrey Hinton Department of Computer Science, University of Toronto ICML.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Varieties of Helmholtz Machine Peter Dayan and Geoffrey E. Hinton, Neural Networks, Vol. 9, No. 8, pp , 1996.
Learning Lateral Connections between Hidden Units Geoffrey Hinton University of Toronto in collaboration with Kejie Bao University of Toronto.
Geoffrey Hinton CSC2535: 2013 Lecture 5 Deep Boltzmann Machines.
Introduction to Neural Networks and Example Applications in HCI Nick Gentile.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 18 Learning Boltzmann Machines Geoffrey Hinton.
CIAR Summer School Tutorial Lecture 1b Sigmoid Belief Nets Geoffrey Hinton.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
How to learn a generative model of images Geoffrey Hinton Canadian Institute for Advanced Research & University of Toronto.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
CSC321: Introduction to Neural Networks and Machine Learning Lecture 19: Learning Restricted Boltzmann Machines Geoffrey Hinton.
Introduction to Deep Learning
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
CSC2515 Fall 2008 Introduction to Machine Learning Lecture 8 Deep Belief Nets All lecture slides will be available as.ppt,.ps, &.htm at
CSC321 Lecture 24 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
CSC321 Lecture 27 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
1 Restricted Boltzmann Machines and Applications Pattern Recognition (IC6304) [Presentation Date: ] [ Ph.D Candidate,
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Big data classification using neural network
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
Learning Deep Generative Models by Ruslan Salakhutdinov
Deep Feedforward Networks
Energy models and Deep Belief Networks
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
Restricted Boltzmann Machines for Classification
LECTURE ??: DEEP LEARNING
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Multimodal Learning with Deep Boltzmann Machines
Deep Learning Qing LU, Siyuan CAO.
Structure learning with deep autoencoders
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Department of Electrical and Computer Engineering
Deep Architectures for Artificial Intelligence
Neuro-Computing Lecture 4 Radial Basis Function Network
Deep Belief Nets and Ising Model-Based Network Construction
Regulation Analysis using Restricted Boltzmann Machines
CSC321 Winter 2007 Lecture 21: Some Demonstrations of Restricted Boltzmann Machines Geoffrey Hinton.
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Cognitive models for emotion recognition: Big Data and Deep Learning Big Data seminar Presentation 10.14.15

Outline Emotiv demo Data Acquisition Cognitive models for emotions recognition Big Data Deep Learning

Human Brain: the Big Data model The most recent publication “Reconstruction and Simulation of Neocortical Microcircuitry”, published in Cell on Oct. 8, 2015 It’s a piece of rat brain containing about 30,000 neurons and 40 million synaptic connections. There’s nothing remarkable about it, except that it isn’t real ! It’s a digital reconstruction—a representation of a one-third cubic millimeter of rat neocortex—and it seems to work like the real thing.

Cerebral cortex The newer portion of the cerebral cortex serves as the center of higher mental functions for humans. The neocortex contains some 100 billion cells.

Emotional brain

The spectrum of emotions

Stages of emotion regulation

MindDriver project: objectives Acquisition and processing of EEG data corresponding to different cognitive states and emotional responses of the human subjects. Used to feed the unsupervised and supervised learning models. Relies on the Deep Learning models with different levels of depth. Implemented using Big Data technologies. The task: Identify the most effective parameters of the learning procedure that would allow to reliably recognize such emotions, considered to be basic, as anger, disgust, fear, happiness, sadness, and surprise, as well as their derivatives: amusement, contempt, contentment, embarrassment, excitement, guilt, pride in achievement, relief, satisfaction, sensory pleasure, and shame.

Signal processing FFT and filter banks Walsh spectrum Fractals Wavelets Resulted in spectrograms with a number of temporal windows The research objective: identify the method most adequate for emotion recognition and mapping onto the brain cortex

Deep Learning as a DBN A fragment of a Deep Belief Network Unsupervised learning Spectrogram as an input (visual nodes) Stacked layers Supervised layer as a final one Parallel solutions at each stage (Map-Reduce) Backpropagation at the last stage

Deep Learning classification Deep networks for unsupervised or generative learning, which are intended to capture high-order correlation of the observed or visible data for pattern analysis or synthesis purposes when no information about target class labels is available. Deep networks for supervised learning, which are intended to directly provide discriminative power for pattern classification purposes, often by characterizing the posterior distributions of classes conditioned on the visible data. Target label data are always available in direct or indirect forms for such supervised learning. They are also called discriminative deep networks. Hybrid deep networks, where the goal is discrimination which is assisted, often in a significant way, with the outcomes of generative or unsupervised deep networks.

Restricted Boltzmann Machine The most widely used hybrid deep architecture A special type of Markov random field that has one layer of (typically Bernoulli) stochastic hidden units and one layer of (typically Bernoulli or Gaussian) stochastic visible or observable units

RBM: theory and method The joint distribution p(v,h;θ) over the visible units v and hidden units h, given the model parameters θ, is defined in terms of an energy function E(v,h;θ) of The marginal probability that the model assigns to a visible vector v is The energy function is defined as The conditional probabilities can be efficiently calculated as

Deep Learning: Stacking the RBM’s Stacking a number of the RBMs learned layer by layer from bottom up gives rise to a DBN. The stacking procedure is as follows. After learning a RBM with binary features such as spectral bins indexed for different electrodes, we treat the activation probabilities of its hidden units as the data for training the Bernoulli Bernoulli RBM one layer up. The activation probabilities of the second layer Bernoulli-Bernoulli RBM are then used as the visible data input for the third-layer Bernoulli-Bernoulli RBM, and so on. It has been shown that the stacking procedure improves a variational lower bound on the likelihood of the training data under the composite model. That is, this greedy procedure achieves approximate maximum likelihood learning. This learning procedure is unsupervised and requires no class label.

The Deep Belief Network – Deep Neural Network (DBN-DNN) architecture

Training example (fictiouos) Six visible nodes (V1..V6): theta rhythm presence (1) and otherwise (0) at different electrode locations Two hidden nodes to differentiate unknown emotional states Training samples: A(1,1,1,0,0,0), B(1,0,1,0,0,0), C(1,1,1,0,0,0), D(0,1,1,1,0,0), E(0,0,1,1,1,0), F(0,0,1,1,1,0) The network learned the following weights: Bias Unit Hidden 1 Hidden 2 Bias Unit -0.08257658 -0.19041546 1.57007782 V1 -0.82602559 -7.08986885 4.96606654 V2 -1.84023877 -5.18354129 2.27197472 V3 3.92321075 2.51720193 4.11061383 V4 0.10316995 6.74833901 -4.00505343 V5 -0.97646029 3.25474524 -5.59606865 V6 -4.44685751 -2.81563804 -2.91540988 In the testing mode, a sample (0,0,0,1,1,) is tested. It turns Hidden 1 on and Hidden 2 off. Interpretation?