Regulation Analysis using Restricted Boltzmann Machines

Slides:



Advertisements
Similar presentations
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
Advertisements

Markov Networks Alan Ritter.
Deep Learning Bing-Chen Tsai 1/21.
CS590M 2008 Fall: Paper Presentation
Nathan Wiebe, Ashish Kapoor and Krysta Svore Microsoft Research ASCR Workshop Washington DC Quantum Deep Learning.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
CS 678 –Boltzmann Machines1 Boltzmann Machine Relaxation net with visible and hidden units Learning algorithm Avoids local minima (and speeds up learning)
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Contextual Classification by Melanie Ganz Lecture 6, Medical Image Analysis 2008.
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
Bayesian Networks Alan Ritter.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Deep Boltzman machines Paper by : R. Salakhutdinov, G. Hinton Presenter : Roozbeh Gholizadeh.
Cristina Manfredotti D.I.S.Co. Università di Milano - Bicocca An Introduction to the Use of Bayesian Network to Analyze Gene Expression Data Cristina Manfredotti.
Learning Energy-Based Models of High-Dimensional Data Geoffrey Hinton Max Welling Yee-Whye Teh Simon Osindero
A Brief Introduction to Graphical Models
6. Experimental Analysis Visible Boltzmann machine with higher-order potentials: Conditional random field (CRF): Exponential random graph model (ERGM):
Detecting robust time-delayed regulation in Mycobacterium tuberculosis Iti Chaturvedi and Jagath C Rajapakse INCOB 2009.
CS774. Markov Random Field : Theory and Application Lecture 08 Kyomin Jung KAIST Sep
Boltzmann Machines and their Extensions S. M. Ali Eslami Nicolas Heess John Winn March 2013 Heriott-Watt University.
Using Fast Weights to Improve Persistent Contrastive Divergence Tijmen Tieleman Geoffrey Hinton Department of Computer Science, University of Toronto ICML.
CSC2535: Computation in Neural Networks Lecture 11: Conditional Random Fields Geoffrey Hinton.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
Learning Lateral Connections between Hidden Units Geoffrey Hinton University of Toronto in collaboration with Kejie Bao University of Toronto.
Markov Random Fields Probabilistic Models for Images
CIAR Second Summer School Tutorial Lecture 1a Sigmoid Belief Nets and Boltzmann Machines Geoffrey Hinton.
CSC 2535 Lecture 8 Products of Experts Geoffrey Hinton.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Fields of Experts: A Framework for Learning Image Priors (Mon) Young Ki Baik, Computer Vision Lab.
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
CSC2535 Lecture 4 Boltzmann Machines, Sigmoid Belief Nets and Gibbs sampling Geoffrey Hinton.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 18 Learning Boltzmann Machines Geoffrey Hinton.
Lecture 2: Statistical learning primer for biologists
CIAR Summer School Tutorial Lecture 1b Sigmoid Belief Nets Geoffrey Hinton.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 19: Learning Restricted Boltzmann Machines Geoffrey Hinton.
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
Cognitive models for emotion recognition: Big Data and Deep Learning
Pattern Recognition and Machine Learning
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Asymptotic Behavior of Stochastic Complexity of Complete Bipartite Graph-Type Boltzmann Machines Yu Nishiyama and Sumio Watanabe Tokyo Institute of Technology,
Gaussian Process Networks Nir Friedman and Iftach Nachman UAI-2K.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
CSC2535 Lecture 5 Sigmoid Belief Nets
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Graphical Models for Segmenting and Labeling Sequence Data Manoj Kumar Chinnakotla NLP-AI Seminar.
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
Learning Deep Generative Models by Ruslan Salakhutdinov
Deep Feedforward Networks
Multi-task learning approaches to modeling context-specific networks
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
Artificial Intelligence (CS 370D)
Restricted Boltzmann Machines for Classification
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Multimodal Learning with Deep Boltzmann Machines
Deep Learning Qing LU, Siyuan CAO.
Structure learning with deep autoencoders
Prof. Adriana Kovashka University of Pittsburgh April 4, 2017
Restricted Boltzman Machines
Estimating Networks With Jumps
Deep Belief Nets and Ising Model-Based Network Construction
Markov Random Fields Presented by: Vladan Radosavljevic.
Network Inference Chris Holmes Oxford Centre for Gene Function, &,
Boltzmann Machine (BM) (§6.4)
CSC321 Winter 2007 Lecture 21: Some Demonstrations of Restricted Boltzmann Machines Geoffrey Hinton.
CSC 578 Neural Networks and Deep Learning
Random Neural Network Texture Model
Presentation transcript:

Regulation Analysis using Restricted Boltzmann Machines Network Modeling Seminar, 10/1/2013 Patrick Michl

Agenda Biological Problem Analysing the regulation of metabolism Modeling Implementation & Results

Biological Problem Analysing the regulation of metabolism Signal Regulation Metabolism A linear metabolic pathway of enzymes (E) …

Biological Problem Analysing the regulation of metabolism Signal Regulation Metabolism … is regulated by transcription factors (TF) …

Biological Problem Analysing the regulation of metabolism Signal Regulation Metabolism … which respond to signals (S)

Biological Problem Analysing the regulation of metabolism Upregulated linear pathways …

Biological Problem Analysing the regulation of metabolism … can appear in different patterns

Biological Problem Analysing the regulation of metabolism ? ? E Which transcription factors and signals cause this patterns …

Biological Problem Analysing the regulation of metabolism ? ? ? ? E … and how do they interact? (topological structure)

Agenda Biological Problem Analysing the regulation of metabolism Network Modeling Restricted Boltzmann Machines (RBM) Validation & Implementation

Network Modeling Restricted Boltzmann Machines (RBM) TF E Lets start with some pathway of our interest …

Network Modeling Restricted Boltzmann Machines (RBM) TF E … and lists of interesting TFs and interesting SigMols

Network Modeling Restricted Boltzmann Machines (RBM) TF E How to model the topological structure?

Network Modeling Restricted Boltzmann Machines (RBM) Graphical Models Graphical Models can preserve topological structures …

Network Modeling Restricted Boltzmann Machines (RBM) Graphical Models Directed Graph Undirected Graph … … but there are many types of graphical models

Network Modeling Restricted Boltzmann Machines (RBM) Graphical Models Directed Graph Bayesian Networks Undirected Graph … The most common type is the Bayesian Network (BN) …

Network Modeling Restricted Boltzmann Machines (RBM) Bayesian Networks a b c P[a,b,c] 0.1 1 0.9 0.5 … a b ? c Bayesian Networks use joint probabilities …

Network Modeling Restricted Boltzmann Machines (RBM) Bayesian Networks a b c P[a,b,c] 0.1 1 0.9 0.5 … a b c … to represents conditional dependencies in an acyclic graph …

Network Modeling Restricted Boltzmann Machines (RBM) Bayesian Networks b a d c … but the regulation mechanism of a cell can be more complicated

Network Modeling Restricted Boltzmann Machines (RBM) Graphical Models Directed Graph Bayesian Networks Undirected Graph Markov Random Fields … Another type of graphical models are Markov Random Fields (MRF)…

Network Modeling Restricted Boltzmann Machines (RBM) Markov Random Fields Motivation (Ising Model) A set of magnetic dipoles (spins) is arranged in a graph (lattice) where neighbors are coupled with a given strengt ... which emerged with the Ising Model from statistical Physics …

Network Modeling Restricted Boltzmann Machines (RBM) Markov Random Fields Motivation (Ising Model) A set of magnetic dipoles (spins) is arranged in a graph (lattice) where neighbors are coupled with a given strengt ... which uses local energies to calculate new states …

Network Modeling Restricted Boltzmann Machines (RBM) Markov Random Fields Drawback By allowing cyclic dependencies the computational costs explode b a d c … the drawback are high computational costs …

Network Modeling Restricted Boltzmann Machines (RBM) Graphical Models Directed Graph Bayesian Networks Undirected Graph Markov Random Fields Restricted Boltzmann Machines (RBM) … … which can be avoided by using Restricted Boltzmann Machines

Network Modeling Restricted Boltzmann Machines (RBM) Neuron like units RBMs are Artificial Neuronal Networks …

Network Modeling Restricted Boltzmann Machines (RBM) v1 v2 v3 v4 h1 h2 h3 … with two layers: visible units (v) and hidden units (h)

Network Modeling Restricted Boltzmann Machines (RBM) v1 v2 v3 v4 h1 h2 h3 Visible units are strictly connected with hidden units

Network Modeling Restricted Boltzmann Machines (RBM) 𝑉 ≔set of visible units 𝑥 𝑣 ≔value of unit 𝑣,∀𝑣∈𝑉 𝑥 𝑣 ∈𝑅, ∀𝑣∈𝑉 In our model the visible units have continuous values …

Network Modeling Restricted Boltzmann Machines (RBM) 𝑉 ≔set of visible units 𝑥 𝑣 ≔value of unit 𝑣,∀𝑣∈𝑉 𝑥 𝑣 ∈𝑅, ∀𝑣∈𝑉 𝐻 ≔set of hidden units 𝑥 ℎ ≔value of unit ℎ, ∀ℎ∈𝐻 𝑥 ℎ ∈{0, 1}, ∀ℎ∈𝐻 … and the hidden units binary values

Network Modeling Restricted Boltzmann Machines (RBM) 𝑥 𝑣 ~𝑁 𝑏 𝑣 + ℎ 𝑤 𝑣ℎ 𝑥 ℎ , 𝜎 𝑣 ,∀𝑣∈𝑉 𝜎 𝑣 ≔ std. dev. of unit 𝑣 𝑏 𝑣 ≔bias of unit 𝑣 𝑤 𝑣ℎ ≔weight of edge (𝑣,ℎ) Visible units are modeled with gaussians to encode data …

Network Modeling Restricted Boltzmann Machines (RBM) 𝑥 𝑣 ~𝑁 𝑏 𝑣 + ℎ 𝑤 𝑣ℎ 𝑥 ℎ , 𝜎 𝑣 ,∀𝑣∈𝑉 𝜎 𝑣 ≔ std. dev. of unit 𝑣 𝑏 𝑣 ≔bias of unit 𝑣 𝑤 𝑣ℎ ≔weight of edge (𝑣,ℎ) 𝑥 ℎ ~sigmoid 𝑏 ℎ + 𝑣 𝑤 𝑣ℎ 𝑥 𝑣 𝜎 𝑣 ,∀ℎ∈𝐻 𝑏 ℎ ≔bias of unit ℎ … and hidden units with simoids to encode dependencies

Network Modeling Restricted Boltzmann Machines (RBM) Learning in Restricted Boltzmann Machines Task: Find dependencies in data ↔ Find configuration of parameters with maximum likelihood (to data) The challenge is to find the configuration of the parameters …

Network Modeling Restricted Boltzmann Machines (RBM) Learning in Restricted Boltzmann Machines Task: Find dependencies in data ↔ Find configuration of parameters with maximum likelihood (to data) In RBMs configurations of parameters have probabilities, that can be defined by local energies Local Energy 1 2 𝐸 𝑣 ≔− ℎ 𝑤 𝑣ℎ 𝑥 𝑣 𝜎 𝑣 𝑥 ℎ + ( 𝑥 𝑣 − 𝑏 𝑣 ) 2 2 𝜎 𝑣 2 𝐸 ℎ ≔− 𝑣 𝑤 𝑣ℎ 𝑥 𝑣 𝜎 𝑣 𝑥 ℎ + 𝑥 ℎ 𝑏 ℎ Like in the Ising model the units states correspond to local energies …

Network Modeling Restricted Boltzmann Machines (RBM) Learning in Restricted Boltzmann Machines Task: Find dependencies in data ↔ Find configuration of parameters with maximum likelihood (to data) ↔ Minimize global energy (to data) Global Energy 𝐸≔ 𝑣 𝐸 𝑣 + ℎ 𝐸 ℎ =− 𝑣 ℎ 𝑤 𝑣ℎ 𝑥 𝑣 𝜎 𝑣 𝑥 ℎ + 𝑣 ( 𝑥 𝑣 − 𝑏 𝑣 ) 2 2 𝜎 𝑣 2 + ℎ 𝑤 𝑣ℎ 𝑥 𝑣 𝜎 𝑣 𝑥 ℎ … which sum to a global energy, which is our objective function

Network Modeling Restricted Boltzmann Machines (RBM) Learning in Restricted Boltzmann Machines Task: Find dependencies in data ↔ Find configuration of parameters with maximum likelihood (to data) ↔ Minimize global energy (to data) ↔ Perform stochastic gradient descent on 𝜎 𝑣 , 𝑏 𝑣 , 𝑏 ℎ , 𝑤 𝑣ℎ (to data) The optimization can be done using stochastic gradient descent …

Network Modeling Restricted Boltzmann Machines (RBM) Learning in Restricted Boltzmann Machines Task: Find dependencies in data ↔ Find configuration of parameters with maximum likelihood (to data) ↔ Minimize global energy (to data) ↔ Perform stochastic gradient descent on 𝜎 𝑣 , 𝑏 𝑣 , 𝑏 ℎ , 𝑤 𝑣ℎ (to data) Gradient Descent on RBMs The bipartite graph structure allows constrastive divergency learning, using Gibbs-sampling … which has an efficient learning algorithmus

Network Modeling Restricted Boltzmann Machines (RBM) TF E How to model our initial structure as an RBM?

Network Modeling Restricted Boltzmann Machines (RBM) TF E We define S and E as visible Layer …

Network Modeling Restricted Boltzmann Machines (RBM) TF We define S and E as visible Layer …

Network Modeling Restricted Boltzmann Machines (RBM) TF … and TF as hidden Layer

Agenda Biological Problem Analysing the regulation of metabolism Network Modeling Restricted Boltzmann Machines (RBM) Implementation & Results python::metapath

Results Validation of the results Information about the true regulation Information about the descriptive power of the data

Results Validation of the results Information about the true regulation Information about the descriptive power of the data Without this infomation validation can only be done, using simulated data!

Results Simulation 1 First of all we need to understand how the modell handles dependencies and noise To demonstrate this we create very simple data with a simple structure

What can we expect from this model? Simulation 1 S TF E What can we expect from this model?

… as RBM we get 8 visible and 2 hidden units, fully connected Simulation 1 S E TF … as RBM we get 8 visible and 2 hidden units, fully connected

Let‘s feed the machine with samples … Simulation 1 Data S E 1,0,0,1 1,0,0,0 1,1,0,0 1,0,1,0 1,0,1,1 0,0,0,0 0,1,0,0 0,0,1,0 0,0,0,1 Let‘s feed the machine with samples …

.. to get the calculated parameters (especially the weight matrix) Simulation 1 Weight matrix TF1 TF2 S1 0,3 0,8 S2 0,5 0,6 S3 1,0 0,1 S4 E1 0,0 E2 E3 E4 0,2 .. to get the calculated parameters (especially the weight matrix)

The weights are visualized by the intensity of the edges Simulation 1 Weight matrix S TF1 TF2 S1 0,3 0,8 S2 0,5 0,6 S3 1,0 0,1 S4 E1 0,0 E2 E3 E4 0,2 TF E The weights are visualized by the intensity of the edges

Now we can compare the results with the samples Simulation 1 Learning samples S S E 1,0,0,1 1,0,0,0 1,1,0,0 1,0,1,0 1,0,1,1 0,0,0,0 0,1,0,0 0,0,1,0 0,0,0,1 TF E Now we can compare the results with the samples

There‘s a strong dependency between S3 an E1 Simulation 1 Learning samples S S E 1,0,0,1 1,0,0,0 1,1,0,0 1,0,1,0 1,0,1,1 0,0,0,0 0,1,0,0 0,0,1,0 0,0,0,1 TF E There‘s a strong dependency between S3 an E1

S1, S2 and S4 do almost not affect the metabolism … Simulation 1 Learning samples S S E 1,0,0,1 1,0,0,0 1,1,0,0 1,0,1,0 1,0,1,1 0,0,0,0 0,1,0,0 0,0,1,0 0,0,0,1 TF E S1, S2 and S4 do almost not affect the metabolism …

… so we can forget them and get S1,TF1 for our regulation model Simulation 1 S TF E … so we can forget them and get S1,TF1 for our regulation model

We can also take a look at the causal mechanism … Simulation 1 Weight matrix TF1 TF2 S1 0,3 0,8 S2 0,5 0,6 S3 1,0 0,1 S4 E1 0,0 E2 E3 E4 0,2 We can also take a look at the causal mechanism …

The edge (S3, TF1) dominates TF1 … Simulation 1 Weight matrix TF1 TF2 S1 0,3 0,8 S2 0,5 0,6 S3 1,0 0,1 S4 E1 0,0 E2 E3 E4 0,2 The edge (S3, TF1) dominates TF1 …

Also E1 seems to have an effect on S3 (fewer than S3 on E1) Simulation 1 TF1 e1 TF1 e4 s3 TF1 TF1 TF1 e3 e2 Also E1 seems to have an effect on S3 (fewer than S3 on E1)

Of course we want to compare the method with Bayesian Networks Results Comparing to Bayesian Networks For this purpose we simulate data in three steps Of course we want to compare the method with Bayesian Networks

Of course we want to compare the method with Bayesian Networks Results Comparing to Bayesian Networks Step 1 Choose number of Genes (E+S) and create random bimodal distributed data Of course we want to compare the method with Bayesian Networks

Of course we want to compare the method with Bayesian Networks Results Comparing to Bayesian Networks Step 1 Choose number of Genes (E+S) and create random bimodal distributed data Step 2 Manipulate data in a fixed order Of course we want to compare the method with Bayesian Networks

Of course we want to compare the method with Bayesian Networks Results Comparing to Bayesian Networks Step 1 Choose number of Genes (E+S) and create random bimodal distributed data Step 2 Manipulate data in a fixed order Step 3 Add noise to manipulated data and normalize data Of course we want to compare the method with Bayesian Networks

Of course we want to compare the method with Bayesian Networks Results Comparing to Bayesian Networks Idea ‚melt down‘ the bimodal distribution from very sharp to very noisy Try to find the original causal structure with BN and RBM Measure Accuracy by counting the right and wrong dependencies Of course we want to compare the method with Bayesian Networks

Of course we want to compare the method with Bayesian Networks Results Simulation 2 Step 1: Number of visible nodes 8 (4E, 4S) Create intergradient datasets from sharp to noisy bimodal distribution 𝜎 1 =0.0, 𝜎 1 =0.3, 𝜎 3 =0.9, 𝜎 4 =1.2, 𝜎 4 =1.5 Step 2 + 3: Data Manipulation + add noise 𝑒 1 =0.5 𝑠 1 +0.5 𝑠 2 +𝑁(𝜇=0, 𝜎) 𝑒 2 =0.5 𝑠 2 +0.5 𝑠 3 +𝑁(𝜇=0, 𝜎) 𝑒 3 =0.5 𝑠 3 +0.5 𝑠 4 +𝑁(𝜇=0, 𝜎) 𝑒 4 =0.5 𝑠 4 +0.5 𝑠 1 +𝑁(𝜇=0, 𝜎) Of course we want to compare the method with Bayesian Networks

Results Simulation 2 RBM Model (𝜎=0.0)

Results Simulation 2 Causal Mechanism (𝜎=0.0)

Results Simulation 2 Comparison BN / RBM

Conclusion Conclusion RBMs are more stable against noise compared to BNs. It has to be assumed that RBMs have high predictive power regarding the regulation mechanisms of cells The drawback are high computational costs Since RBMs are getting more popular (Face recognition / Voice recognition, Image transformation). Many new improvements in facing the computational costs have been made.

Acknowledgement eilsLABS PD Dr. Rainer König Prof. Dr Roland Eils Network Modeling Group