IMA, October 29, 2007 Slide 1 T H E B I O I N F O R M A T I C S C E N T R E A continuous probabilistic model of local RNA 3-D structure Jes Frellsen The.

Slides:



Advertisements
Similar presentations
Systems biology SAMSI Opening Workshop Algebraic Methods in Systems Biology and Statistics September 14, 2008 Reinhard Laubenbacher Virginia Bioinformatics.
Advertisements

Bayesian network for gene regulatory network construction
Trustworthy Service Selection and Composition CHUNG-WEI HANG MUNINDAR P. Singh A. Moini.
Autonomic Scaling of Cloud Computing Resources
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
Dynamic Bayesian Networks (DBNs)
Simulation Metamodeling using Dynamic Bayesian Networks in Continuous Time Jirka Poropudas (M.Sc.) Aalto University School of Science and Technology Systems.
Introduction of Probabilistic Reasoning and Bayesian Networks
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Geometric Algorithms for Conformational Analysis of Long Protein Loops J. Cortess, T. Simeon, M. Remaud- Simeon, V. Tran.
Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –One exception: games with multiple moves In particular, the Bayesian.
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
Non-coding RNA William Liu CS374: Algorithms in Biology November 23, 2004.
Hidden Markov Model Special case of Dynamic Bayesian network Single (hidden) state variable Single (observed) observation variable Transition probability.
Bayesian Classification of Protein Data Thomas Huber Computational Biology and Bioinformatics Environment ComBinE Department of Mathematics.
Extending Evolutionary Programming to the Learning of Dynamic Bayesian Networks Allan Tucker Xiaohui Liu Birkbeck College University of London.
Genome evolution: a sequence-centric approach Lecture 3: From Trees to HMMs.
Overview of STAT 270 Ch 1-9 of Devore + Various Applications.
Who am I and what am I doing here? Allan Tucker A brief introduction to my research
Chapter 14 Simulation. Monte Carlo Process Statistical Analysis of Simulation Results Verification of the Simulation Model Computer Simulation with Excel.
Learning Bayesian Networks
Introduction to molecular networks Sushmita Roy BMI/CS 576 Nov 6 th, 2014.
Dynamic Bayesian Networks CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanningLearning.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Cristina Manfredotti D.I.S.Co. Università di Milano - Bicocca An Introduction to the Use of Bayesian Network to Analyze Gene Expression Data Cristina Manfredotti.
Lecture II-2: Probability Review
Real-Time Odor Classification Through Sequential Bayesian Filtering Javier G. Monroy Javier Gonzalez-Jimenez
Computer Simulation A Laboratory to Evaluate “What-if” Questions.
Introduction to Monte Carlo Methods D.J.C. Mackay.
L7.1b Continuous Random Variables CONTINUOUS RANDOM VARIABLES NORMAL DISTRIBUTIONS AD PROBABILITY DISTRIBUTIONS.
1 Performance Evaluation of Computer Networks: Part II Objectives r Simulation Modeling r Classification of Simulation Modeling r Discrete-Event Simulation.
Using Bayesian Networks to Analyze Expression Data N. Friedman, M. Linial, I. Nachman, D. Hebrew University.
Monte Carlo Simulation CWR 6536 Stochastic Subsurface Hydrology.
Probabilistic Mechanism Analysis. Outline Uncertainty in mechanisms Why consider uncertainty Basics of uncertainty Probabilistic mechanism analysis Examples.
Probabilistic graphical models. Graphical models are a marriage between probability theory and graph theory (Michael Jordan, 1998) A compact representation.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Using Bayesian Networks to Analyze Whole-Genome Expression Data Nir Friedman Iftach Nachman Dana Pe’er Institute of Computer Science, The Hebrew University.
Markov Random Fields Probabilistic Models for Images
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Randomized Algorithms for Bayesian Hierarchical Clustering
Processing Sequential Sensor Data The “John Krumm perspective” Thomas Plötz November 29 th, 2011.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Comp. Genomics Recitation 9 11/3/06 Gene finding using HMMs & Conservation.
Dependency Networks for Collaborative Filtering and Data Visualization UAI-2000 발표 : 황규백.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Flat clustering approaches
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Chapter 8. Learning of Gestures by Imitation in a Humanoid Robot in Imitation and Social Learning in Robots, Calinon and Billard. Course: Robots Learning.
MINRMS: an efficient algorithm for determining protein structure similarity using root-mean-squared-distance Andrew I. Jewett, Conrad C. Huang and Thomas.
CSE 474 Simulation Modeling | MUSHFIQUR ROUF CSE474:
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
CS-ROSETTA Yang Shen et al. Presented by Jonathan Jou.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Computational methods for inferring cellular networks II Stat 877 Apr 17 th, 2014 Sushmita Roy.
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
Constrained Hidden Markov Models for Population-based Haplotyping
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
CSCI 5822 Probabilistic Models of Human and Machine Learning
Hidden Markov Models Part 2: Algorithms
DataMining, Morgan Kaufmann, p Mining Lab. 김완섭 2004년 10월 27일
Large Time Scale Molecular Paths Using Least Action.
Input Output HMMs for modeling network dynamics
Markov Random Fields Presented by: Vladan Radosavljevic.
Learning Probabilistic Graphical Models Overview Learning Problems.
LECTURE 15: REESTIMATION, EM AND MIXTURES
Experimental Overview
Presentation transcript:

IMA, October 29, 2007 Slide 1 T H E B I O I N F O R M A T I C S C E N T R E A continuous probabilistic model of local RNA 3-D structure Jes Frellsen The Bioinformatics Centre Department of Molecular Biology University of Copenhagen

IMA, October 29, 2007 Slide 2 T H E B I O I N F O R M A T I C S C E N T R E Background 3D structure is important for understanding the function of non-coding RNA molecules Experimental methods for determining 3D structure are time consuming and sometimes difficult Local structure is typically modeled by using discretization E.g. fragment libraries are used in current methods for structure prediction Our group has recently made a continuous probabilistic model of local protein structure with great success [PLoS Comput Biol 2006, 2: ] Dynamic Bayesian Networks Directional statistics We have used a similar approach to model local structure of RNA

IMA, October 29, 2007 Slide 3 T H E B I O I N F O R M A T I C S C E N T R E Representation of RNA Each nucleotide in an RNA molecule can be represented by the base type and 7 dihedrals angles Allows for accurate conversion into coordinates of all atoms in the structure using standard values

IMA, October 29, 2007 Slide 4 T H E B I O I N F O R M A T I C S C E N T R E Angle distributions Each variable lies on a circle Requires directional statistics Each variable is multi-modal Can be described by a mixture of simple distributions Von Mises distribution The angles co-vary both within nucleotides and between consecutive nucleotides We model this by a sequential model

IMA, October 29, 2007 Slide 5 T H E B I O I N F O R M A T I C S C E N T R E Our model An DBN with 3 random variables per angle: Discrete input variable indicating angle type (7 states) Hidden variable with 20 states Output variable representation the angle value and the CPDs given the hidden state is modelled by Von Mises distributions Structure of an IOHMM with continuous output (except bookkeeping) Does not impose a groping of the angles Parameters are estimated by stochastic EM from experimental data

IMA, October 29, 2007 Slide 6 T H E B I O I N F O R M A T I C S C E N T R E Evaluating the model Individual angle distributions The model captures the distribution of the individual angles E.g. the -angle and the -angle:

IMA, October 29, 2007 Slide 7 T H E B I O I N F O R M A T I C S C E N T R E Evaluating the model Pairwise distribution The model captures the pairwise dependencies between the angles E.g. the pairwise distribution of  and  (inter-nucleotide)

IMA, October 29, 2007 Slide 8 T H E B I O I N F O R M A T I C S C E N T R E Proof of concept: generating decoys for a target structure A simple simulated annealing scheme: 1.Sample a whole structure, S, without clashes 2.Make new structure, S’, by resampling four consecutive angles in S (randomly picked) 3.Evaluate S’ a.If it has clashed it is rejected b.If it has a better energy than S then S’ is set to be the new S c.If it has a worse energy then with probability, p, S’ is set to be the new S (otherwise it is rejected) d.Go to step 2 In the scheme we used p = e (E-E’) /T, where T decreases with time a simple “energy function” that promotes structure with the same Watson-Crick base pair as are found in the target structure

IMA, October 29, 2007 Slide 9 T H E B I O I N F O R M A T I C S C E N T R E Results of generating 1,500 decoys for 5 different structures Target Structure Length (Bases) Decoys < 4Å Decoys < 3Å Lowest RMSD 1ZIH1258.8%21.3%1.55Å 1RNG1255.1%3.5%2.48Å 1XWP1328.3%5.8%2.03Å 1I4B1334.6%0.1%2.91Å 1PJY2210.0%1.9%1.89Å Target structureBest decoy 1ZIH

IMA, October 29, 2007 Slide 10 T H E B I O I N F O R M A T I C S C E N T R E Perspectives The model assigns a probability distribution to the conformational space and describes many aspects of local RNA structure well It has numerous applications! It allows for fast probabilistic sampling of locally RNA-like structures Can thus be used in RNA 3D structure prediction The model can be used to calculate the probabilities of seeing different local structures Can thus be used for quality validation of experimentally determined structures

IMA, October 29, 2007 Slide 11 T H E B I O I N F O R M A T I C S C E N T R E Acknowledgements The research was conducted in the structural bioinformatics group, lead by Thomas Hamelryck, by Jes Frellsen, Ida Moltke, Martin Thiim and Thomas Hamelryck We would like to thank Our collaborator Senior Research Professor Kanti V. Mardia from The University of Leeds for his contributions on directional statistics. The Richardsons Lab at Duke University for making their RNA dataset available JF thanks IMA for the invitation to the conference JF is funded by The Danish Council for Strategic Research TH is funded by The Danish Council for Technology and Innovation

IMA, October 29, 2007 Slide 12 T H E B I O I N F O R M A T I C S C E N T R E Bayesian Networks and Dynamic Bayesian Networks A BN is a DAG where Nodes are random variables Edges represent conditional dependencies in the factorization of the joint probability The graph encodes conditional indepencies E.g. A and D is conditional independent give C DBNs are the time series expansion of BNs E.g. an HMM: