Download presentation
Presentation is loading. Please wait.
Published byPhebe Butler Modified over 9 years ago
1
Using Bayesian Networks to Analyze Expression Data By Friedman Nir, Linial Michal, Nachman Iftach, Pe'er Dana (2000) Presented by Nikolaos Aravanis Lysimachos Zografos Alexis Ioannou
2
Outline Introduction Bayesian Networks Application to expression data Application to cell cycle expression patterns Discussion and future work
3
The Road to Microarray Data Analysis Development of microarrays –Measure all the genes of an organism Enormous amount of data Challenge: Analyze datasets and infer biological interactions
4
Most Common Analysis Tool Clustering Algorithms Allocate groups of genes with similar expression patterns over a set of experiments Discover genes that are co-regulated
5
Problems Data give only a partial picture –Key events are not reflected (translation and protein (in) activation) Amount of samples give few information for constructing full detailed models Using current technologies even few samples have high noise to signal ratio
6
Possible Solution Analyze gene expression patterns that uncover properties of the transcriptional program Examine dependence and conditional independence of the data Bayesian Networks
7
Represent dependence structure between multiple interacting quantities Capable of handling noise and estimating the confidence in the different features of the network Focus on interactions whose signal is strong Useful for describing processes composed of locally interacting components Statistical foundations for learning Bayesian networks and the statistics to do so are well understood and have been successfully applied Provide models of causal influence
8
Informal Introduction to Bayesian Networks Let P(X,Y) be a joint distribution over variables X and Y X and Y independent if P(X,Y) = P(X)P(Y) for all values X and Y Gene A is transcriptor factor of gene B We expect their expression level to be dependent A parent of B B trascription factor of C Expression levels of each pair are dependent If A does not directly affect C, if we fix the expression level of B, we will observe A and C are independent P(A|B,C) = P(A|B) (A and C conditionally independent of B) I(A;C|B)
9
Informal Introduction to Bayesian Networks (contd’) Component of Bayesian Networks is that each variable is a stochastic function of its parents Stochastic models are natural in gene expression domain The biological models we want to process are stochastic Measurements are noisy
10
Representing Distributions with Bayesian Networks Representation of joint probability distribution consisting of 2 components Directed acyclic graph (G) Conditional distribution for each variable given its parents in G G encodes Markov Assumption By applying chain rule this decomposes in product form
11
Equivalence Classes of BNs A BN implies further independence assumptions => Ind(G) >1 graphs can imply the same assumptions => Equivalent networks if Ind(G)=Ind(G')
12
Equivalence Classes of BNs A BN implies further independence statements => Ind(G) >1 graphs can imply the same statements => Equivalent networks if Ind(G)=Ind(G') Ind(G)=Ind(G')=Ø
13
Equivalence Classes of BNs For equivalent networks: DAGs have the same underlying undirected graph. PDAGs are used to represent them.
14
Equivalence Classes of BNs For equivalent networks: DAGs have the same underlying undirected graph. PDAGs are used to represent them. Disagreeing edge
15
Question: Given dataset D, what BN, B= best matches D? Answer: Statistically motivated scoring function to evaluate each BN: e.g. Bayesian Score S(G:D)=logP(G|D)=logP(D|G)+logP(G)+C, where C is a constant independent of G and P(D|G)=∫P(D|G,Θ)P(Θ|G)dΘ is the marginal likelihood over all parameters for G. Learning BNs
16
Question: Given dataset D, what BN, B= best matches D? Answer: Statistically motivated scoring function to evaluate each BN: e.g. Bayesian Score S(G:D)=logP(G|D)=logP(D|G)+logP(G)+C, where C is a constant independent of G and P(D|G)=∫P(D|G,Θ)P(Θ|G)dΘ is the marginal likelihood over all parameters for G. Learning BNs
17
Learning BNs (contd) Steps: – Decide priors (P(Θ|D), P(G)) => Use of BDe priors (structure equivalent, decomposable) – Find G to maximize S(G:D) NP hard problem =>local search using local permutations of candidate G (Heckerman et al. 1995)
18
Learning Causal Patterns – Bayesian Network is model of dependencies – Interest in modelling the process that generated them. => model the flow of causality in the system of interest and create a Causal Network (CN). A Causal Network models the probability distribution as well as the effect of causality.
19
CNs VS BNs: - CNs interpret parents as immediate causes (c.f. BNs) - CNs and BNs relate when using the Causal Markov Assumption : “given the values of a variable's immediate causes, it is independent of its earlier causes”, if this holds, then BN==CN Learning Causal Patterns
20
CNs VS BNs: - CNs interpret parents as immediate causes (c.f. BNs) - CNs and BNs relate when using the Causal Markov Assumption : “given the values of a variable's immediate causes, it is independent of its earlier causes”, if this holds, then BN==CN Learning Causal Patterns X YX Y equivalent BNs but not CNs
21
Applying BNs to Expression Data Expression level of each gene as a random variable Other attributes (e.g temperature, exp. conditions) that affect the system can be modelled as random variables Bayesian Net/ Dependency structure can answer queries CON: problems in computational complexity and the statistical significance of the resulting networks. PRO: genetic regulation networks are sparse
22
Representing Partial Models – Gene networks: many variables => >1 plausible models (not enough data) – we can learn up to equivalence class. Focus on feature learning in order to have a causal network:
23
Representing Partial Models Features: - Markov relations (e.g. Markov Blanket) - Order relations (e.g. X is an ancestor of Y in all networks)
24
Representing Partial Models Features: - Markov relations (e.g. Markov Blanket) - Order relations (e.g. X is an ancestor of Y in all networks) Feature learning leads to a Causal Network
25
Statistical Confidence of Features – Likelihood that a given feature is actually true. – Can't calculate posterior (P(G|D)) => Bootstrap method for i=1...n resample D with replacement -> D'; learn G' from D'; end
26
Statistical Confidence of Features Individual feature confidence (IFC) IFC = (1/n)∑{f(G')} where f(G') = 1 if the feature exists in G'
27
Efficient Learning Algorithms – Vast search space => need efficient algorithms – Attention on relevant regions of the search space => Sparse Candidate Algorithm
28
Efficient Learning Algorithms Sparse Candidate Algorithm Identify a small number of candidate parents for each gene based on simple local statistics (e.g. correlation). – Restrict our search to networks with the candidate parents – Potential pitfall: early choice => Solution: adaptive algorithm
29
Discretization The practical side: Need to define the local probability model for each variable. => discretize experimental data into -1,0,1 (expression level lower, similar, higher than control) Set control by averaging. Set a threshold ratio for significantly higher/lower.
30
Application to Cell Cycle Expression Patterns 76 gene expression measurements of the mRNA levels of 6177 Saccharomyces cerevisiae ORFs. Six time series under different cell cycle synchronization methods(Spellman 1998). 800 differentially expressed, 250 clustered in 8 distinct clusters. Variables for the networks represent the expression level of the 800 genes. Introduced an additional variable that denoted the cell cycle phase to deal with the temporal nature of the cell cycle process and forced it as a root in the network Applied Sparse Candidate Algorithm to 200- fold bootstrap of the original data. Used no prior biological knowledge in the learning algorithm
31
Network with all edges
32
Network with edges that represent relations with confidence level above 0.3
33
YNL058C Local Map Edges Markov Ancestors Descendants SGD entry YPD entry
34
Robustness analysis Use 250 gene data for robustness analysis Create random data set by permuting the order of experiments independently for each gene No “real” features are expected to be found
35
Robustness analysis (contd’) Lower confidence for order and Markov relations in the random data set Longer and heavier tail in the high confidence region in the original data set Sparser networks learned from real data Features learned in original data with high confidence level are not an artifact of the bootstrap estimation
36
Robustness analysis (contd’) Compared confidence level of learned features between 250 and 800 gene data set Strong linear correlation Compared confidence level of learned features between different discretization thresholds Definite linear tendency
37
Biological Analysis Order relations Dominant genes indicate potential causal sources of the cell cycle process Dominance score of X where is the confidence in X being ancestor of Y, k is used to reward high confidence features and t is a threshold to discard low confidence ones
38
Biological Analysis (contd’) Dominant genes are key genes in basic cell functions
39
Biological Analysis (contd’) Top Markov relations reveal functional relations between genes 1. Both genes known: The relations make sense biologically 2. One unknown gene: Firm homologies to proteins functionally related to the other gene 3. Two unknown genes: Physically adjacent to the chromosome, presumably regulated by the same mechanism FAR1- ASH1, low correlation, different clusters, known though to participate in a mating type switch CLN2 is likely to be a parent to RNR3, SVS1, SRO4 and RAD41. Appeared in same cluster. No links between the 4 genes. CLN2 is known to be a central cycle control and there is no clear biological relationship between the others Markov relations
40
Discussion and Future Work Applied Sparse Candidate Algorithm and Bootstrap resampling to extract a Bayesian Network for the 800 genes data set of Spellman Used no prior biological knowledge Derived biologically plausible conclusions Capability of discovering causal relationships, interactions between genes and rich structure between clusters. Developing hybrid algorithms with clustering algorithms to learn models over clustered genes Extensions: –Learn local probability models dealing with continuous data –Improve theory and algorithms –Include biological knowledge as prior knowledge –Improve search heuristics –Apply Dynamic Bayesian Networks to temporal data –Discover causal patterns (using interventional data)
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.