Download presentation
Presentation is loading. Please wait.
Published byTracy Elliott Modified over 8 years ago
1
A SENSITIVITY ANALYSIS OF A BIOLOGICAL MODULE DISCOVERY PIPELINE James Long International Arctic Research Center University of Alaska Fairbanks March 25, 2015
2
A Sensitivity Analysis of a Biological Module Discovery Pipeline
3
Gene Expression A Sensitivity Analysis of a Biological Module Discovery Pipeline
4
Gene Expression Synthetic Gene Expression Data A Sensitivity Analysis of a Biological Module Discovery Pipeline
5
Gene Expression Synthetic Gene Expression Data CODENSE A Sensitivity Analysis of a Biological Module Discovery Pipeline
6
Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis A Sensitivity Analysis of a Biological Module Discovery Pipeline
7
Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline
8
Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline
9
Gene Expression
17
Hill Function
18
Gene Expression Hill Function
19
Gene Expression Hill Function,
20
Gene Expression
21
Hill Function
22
Generalized Hill Function
24
for activators
25
Generalized Hill Function for activators for repressors
26
Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline
27
Synthetic Gene Expression Data
28
NEMO
29
Synthetic Gene Expression Data NEMO – Network Motif Language
30
Synthetic Gene Expression Data NEMO – Network Motif Language COPASI
31
Synthetic Gene Expression Data NEMO – Network Motif Language COPASI – Complex Pathway Simulator
32
Synthetic Gene Expression Data NEMO – Network Motif Language COPASI – Complex Pathway Simulator
33
NEMO
34
G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
35
G0 G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
36
NEMO G0( G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
37
NEMO G0(P1 G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
38
NEMO G0(P1+ G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
39
NEMO G0(P1+, G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
40
NEMO G0(P1+,P2+, G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
41
NEMO G0(P1+,P2+,P3-, G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
42
NEMO G0(P1+,P2+,P3-,P4-, G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
43
NEMO G0(P1+,P2+,P3-,P4-,P5+) G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
44
NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
45
NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G2(…) G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
46
NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G2(…) etc. G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
47
NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G2(…) etc.) GLIST( G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
48
NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G2(…) etc.)] GLIST( [ G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5
49
Dense Overlapping Regulon (DOR) NEMO G0G1 G3 G4G5 G2
50
G3(P0+,P1+,P2-) NEMO Dense Overlapping Regulon (DOR) G0G1 G3 G4G5 G2
51
G3(P0+,P1+,P2-), G4(P0+,P1+) NEMO Dense Overlapping Regulon (DOR) G0G1 G3 G4G5 G2
52
G3(P0+,P1+,P2-), G4(P0+,P1+), G5(P1-,P2+) NEMO Dense Overlapping Regulon (DOR) G0G1 G3 G4G5 G2
53
G0G1 G3 G4G5 G2 DOR(G3(P0+,P1+,P2-), G4(P0+,P1+), G5(P1-,P2+)) NEMO Dense Overlapping Regulon (DOR)
54
Negative auto-regulation NEMO G0
55
Negative auto-regulation G0 G0(P0-) NEMO
56
Feed-forward loop (FFL) G0G1 G2 NEMO
57
G0G1 G2 P0 NEMO Feed-forward loop (FFL)
58
G0G1 G2 P0(+G1 NEMO Feed-forward loop (FFL)
59
G0G1 G2 P0(+G1+G2 NEMO Feed-forward loop (FFL)
60
G0G1 G2 P0(+G1+G2+ NEMO Feed-forward loop (FFL)
61
G0G1 G2 P0(+G1+G2+) NEMO Feed-forward loop (FFL)
62
Multi-output FFL G0G1 G2 NEMO
63
G0G1 G2 G3 NEMO Multi-output FFL
64
G0G1 G2 G3G4 NEMO Multi-output FFL
65
P0(+G1+G2+) G0G1 G2 G3G4 NEMO Multi-output FFL
66
P0(+G1+(G2,G3,G4)+) G0G1 G2 G3G4 NEMO Multi-output FFL
67
P0(+G1+(G2,G3,G4)+)) G0G1 G2 G3G4 NEMO Multi-output FFL TMLIST(
68
Single-input module (SIM) G0 G1 G2G3 NEMO
69
P0(+G1,G2,G3) G0 G1 G2G3 NEMO Single-input module (SIM)
70
G0G1G2 G3G4G5G6 G7G8G9G10 NEMO
71
G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-) NEMO
72
G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-) NEMO
73
G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6) NEMO
74
G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), G7(P3+,P4-,P5+) NEMO
75
G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), G7(P3+,P4-,P5+), G8(P4+) NEMO
76
G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+) NEMO
77
G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-) NEMO
78
G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-)) NEMO
79
G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), TMLIST(P0(+G1-G2-), P1(+G3,G4,G5,G6)), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-)) NEMO
80
G0G1G2 G3G4G5G6 G7G8G9G10 GLIST(G0(P5-)), TMLIST(P0(+G1-G2-), P1(+G3,G4,G5,G6)), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-)) NEMO
81
G0G1G2 G3G4G5G6 G7G8G9G10 [ GLIST(G0(P5-)), TMLIST(P0(+G1-G2-), P1(+G3,G4,G5,G6)), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-)) ] NEMO
82
G0G1G2 G3G4G5G6 G7G8G9G10 [ GLIST(G0(P5-)), TMLIST(P0(+G1-G2-), P1(+G3,G4,G5,G6)), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5- :F(power(sin(P5),2)))) ] NEMO
83
NEMO compiler emits SBML
84
NEMO NEMO compiler emits SBML – Systems Biology Markup Language
85
NEMO NEMO compiler emits SBML – Systems Biology Markup Language – Uses libsbml from http://sbml.orghttp://sbml.org
86
NEMO NEMO compiler emits SBML – Systems Biology Markup Language – Uses libsbml from http://sbml.orghttp://sbml.org More accurate to call it a “language translator”
87
NEMO NEMO compiler emits SBML – Systems Biology Markup Language – Uses libsbml from http://sbml.orghttp://sbml.org More accurate to call it a “language translator”, that adds random generalized Hill functions!
88
NEMO NEMO compiler emits SBML – Systems Biology Markup Language – Uses libsbml from http://sbml.orghttp://sbml.org More accurate to call it a “language translator”, that adds random generalized Hill functions! Bioinformatics 2008 Jan 1;24(1):132-4
89
NEMO G0 G0 G1 G1 [GLIST(G0(P1+), G1(P0+)) ]
90
NEMO
91
NEMO B_0 P1 K_0 n_0 1 P1 K_0 n_0 tau
92
NEMO dc_0 P0 tau
93
NEMO P0 tau B_1 P0 K_1 n_1 1
94
NEMO P0 K_1 n_1 tau dc_1 P1 tau
95
NEMO P1 tau
96
Synthetic Gene Expression Data NEMO – Network Motif Language COPASI – Complex Pathway Simulator
101
Also outputs a table of data for each time step.
102
COPASI – Complex Pathway Simulator Also outputs a table of data for each time step, the last column of which is our synthetic data!
103
COPASI – Complex Pathway Simulator
104
Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline
105
The CODENSE algorithm
106
Input – a series of expression correlation graphs, each representing a different state for an organism.
107
The CODENSE algorithm Input – a series of expression correlation graphs, each representing a different state for an organism. Output – groups of genes (modules) whose expression is correlated across the series of expression correlation graphs.
109
Expression Correlation
110
Pearson’s Correlation
111
Expression Correlation Pearson’s Correlation – Linear dependence between two variables
112
Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score
113
Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score – Number of standard deviations above the mean
114
Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score – Number of standard deviations above the mean Mutual Information
115
Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score – Number of standard deviations above the mean Mutual Information – A measure of mutual dependence between variables
116
Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score – Number of standard deviations above the mean Mutual Information – A measure of mutual dependence between variables – Non-linear dependence OK
135
The CODENSE algorithm
136
Manuscript in preparation
137
ODES – Overlapping Dense Subgraph Algorithm
139
vertex
140
edge
141
vertex edge connected
142
vertex edge connected
143
vertex edge connected cut vertex
144
vertex edge cut vertex
145
vertex edge cut vertex disconnected
146
vertex edge disconnected
149
ODES Density of a graph
150
ODES Number of actual edges / Number of possible edges
151
ODES Density of a graph Number of actual edges / Number of possible edges Number of possible edges
152
ODES Density of a graph Number of actual edges / Number of possible edges Number of possible edges Density
153
ODES Density of a graph Number of actual edges / Number of possible edges Number of possible edges Density Degree of a vertex
154
ODES Density of a graph Number of actual edges / Number of possible edges Number of possible edges Density Degree of a vertex Number of edges incident to the vertex
155
ODES Theorem
156
ODES A connected graph, with density, and number of vertices, has at least one non-cut vertex where degree, the average degree of vertices in.
157
ODES Theorem A connected graph, with density, and number of vertices, has at least one non-cut vertex where degree, the average degree of vertices in. Removal of from does not decrease the density of.
158
ODES Theorem A connected graph, with density, and number of vertices, has at least one non-cut vertex where degree, the average degree of vertices in. Removal of from does not decrease the density of. Bioinformatics (2010) 26 (21): 2788-2789.
160
8 vertices
161
22 edges
162
8 vertices 22 edges average degree = 44/8 = 5.5
163
8 vertices 22 edges average degree = 44/8 = 5.5 density = 2*22/(8(8-1)) ~ 0.78
164
8 vertices 22 edges average degree = 44/8 = 5.5 density = 2*22/(8(8-1)) ~ 0.78
165
8 vertices 22 edges average degree = 5.5 density ~ 0.78
166
7 vertices 17 edges average degree ~ 4.86 density ~ 0.81
167
7 vertices 17 edges average degree ~ 4.86 density ~ 0.81
168
6 vertices 13 edges average degree ~ 4.33 density ~ 0.87
169
6 vertices 13 edges average degree ~ 4.33 density ~ 0.87
170
5 vertices 9 edges average degree = 3.6 density ~ 0.9
171
5 vertices 9 edges average degree = 3.6 density ~ 0.9
172
4 vertices 6 edges average degree = 3.0 density = 1.0
173
4 vertices 6 edges average degree = 3.0 density = 1.0
174
3 vertices 3 edges average degree = 2.0 density = 1.0
175
3 vertices 3 edges average degree = 2.0 density = 1.0
176
2 vertices 1 edge average degree = 1.0 density = 1.0
177
2 vertices 1 edge average degree = 1.0 density = 1.0
178
2 vertices 1 edge average degree = 1.0 density = 1.0
179
3 vertices 3 edges average degree = 2.0 density = 1.0
180
3 vertices 3 edges average degree = 2.0 density = 1.0
181
4 vertices 6 edges average degree = 3.0 density = 1.0
182
4 vertices 6 edges average degree = 3.0 density = 1.0
183
5 vertices 9 edges average degree = 3.6 density ~ 0.9
184
5 vertices 9 edges average degree = 3.6 density ~ 0.9
185
6 vertices 13 edges average degree ~ 4.33 density ~ 0.87
186
6 vertices 13 edges average degree ~ 4.33 density ~ 0.87
187
7 vertices 17 edges average degree ~ 4.86 density ~ 0.81
188
7 vertices 17 edges average degree ~ 4.86 density ~ 0.81
189
8 vertices 22 edges average degree = 5.5 density ~ 0.78
190
8 vertices 22 edges average degree = 5.5 density ~ 0.78
191
9 vertices 24 edges average degree ~ 5.3 density ~ 0.67
192
8 vertices 22 edges average degree = 5.5 density ~ 0.78
193
8 vertices 22 edges average degree = 5.5 density ~ 0.78 Note: brute-force search is confined to actual dense subgraphs.
194
Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline
195
Regionalized Sensitivity Analysis
196
Monte Carlo model runs Regionalized Sensitivity Analysis
197
Monte Carlo model runs Evaluate one or more binary Objective Functions Regionalized Sensitivity Analysis
198
Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? Regionalized Sensitivity Analysis
199
Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? – Exact modules returned w/ limited false positives? Regionalized Sensitivity Analysis
200
Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? – Exact modules returned w/ limited false positives? – Approximate modules returned w/ limited false positives? Regionalized Sensitivity Analysis
201
Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? – Exact modules returned w/ limited false positives? – Approximate modules returned w/ limited false positives? – Half of known modules returned approximately w/ limited false positives? Regionalized Sensitivity Analysis
202
Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? – Exact modules returned w/ limited false positives? – Approximate modules returned w/ limited false positives? – Half of known modules returned approximately w/ limited false positives? Increment parameter bins based on Objective Function conformance or non-conformance Regionalized Sensitivity Analysis
211
Compile the NEMO representation of the canonical network Regionalized Sensitivity Analysis
212
Compile the NEMO representation of the canonical network Import into COPASI Regionalized Sensitivity Analysis
213
Compile the NEMO representation of the canonical network Import into COPASI Set up model, and save in COPASI format Regionalized Sensitivity Analysis
214
Compile the NEMO representation of the canonical network Import into COPASI Set up model, and save in COPASI format Create template from COPASI format file where all genes are turned off (B = 0) Regionalized Sensitivity Analysis
215
Compile the NEMO representation of the canonical network Import into COPASI Set up model, and save in COPASI format Create template from COPASI format file where all genes are turned off (B = 0) Create synthetic data by turning some genes and SIMs on, taking last column of COPASI output. Regionalized Sensitivity Analysis
216
Compile the NEMO representation of the canonical network Import into COPASI Set up model, and save in COPASI format Create template from COPASI format file where all genes are turned off (B = 0) Create synthetic data by turning some genes and SIMs on, taking last column of COPASI output. Add noise to output. Regionalized Sensitivity Analysis
217
Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline
218
Results, One Module
219
First RSA runs used a static transcription network Results, One Module
220
First RSA runs used a static transcription network – For Objective Function 1, no noise added Results, One Module
221
First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation Results, One Module
222
First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 Results, One Module
223
First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio Results, One Module
224
First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio For Z-scores, sensitivity to PC cutoff score greatly attenuated Results, One Module
225
First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio For Z-scores, sensitivity to PC cutoff score greatly attenuated – Lower PC scores allow more false positives to enter pipeline Results, One Module
226
First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio For Z-scores, sensitivity to PC cutoff score greatly attenuated – Lower PC scores allow more false positives to enter pipeline – Significant sensitivities to similarity cutoff score and minimum density for a dense subgraph in coherent dense subgraphs Results, One Module
227
First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio For Z-scores, sensitivity to PC cutoff score greatly attenuated – Lower PC scores allow more false positives to enter pipeline – Significant sensitivities to similarity cutoff score and minimum density for a dense subgraph in coherent dense subgraphs – Still a very low conformance/non-conformance ratio Results, One Module
228
First RSA runs used a static transcription network – For other Objective Functions Results, One Module
229
First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Results, One Module
230
First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Results, One Module
231
First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Observations Results, One Module
232
First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Observations – No good at finding the exact module Results, One Module
233
First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Observations – No good at finding the exact module – No noise is unrealistic Results, One Module
234
First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Observations – No good at finding the exact module – No noise is unrealistic – Static network parameters are unrealistic Results, One Module
235
RSA with noise and dynamic network parameters Results, One Module
236
RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum Results, One Module
237
RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% Results, One Module
238
RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% – Jumps to 30-45% with Z-score methods Results, One Module
239
RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% – Jumps to 30-45% with Z-score methods – Sensitivities mostly disappear in the presence of noise Results, One Module
240
RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% – Jumps to 30-45% with Z-score methods – Sensitivities mostly disappear in the presence of noise sensitive to how fast module genes reach equilibrium Results, One Module
241
RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% – Jumps to 30-45% with Z-score methods – Sensitivities mostly disappear in the presence of noise sensitive to how fast module genes reach equilibrium sensitive to percentage of time a module is ‘on’ in the data Results, One Module
242
RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods Results, One Module
243
RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified Results, One Module
244
RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% Results, One Module
245
RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified Results, One Module
246
RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified – conformance/non-conformance drops to ~ 15-25% Results, One Module
247
RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified – conformance/non-conformance drops to ~ 15-25% sigma = 33% of datum to be modified Results, One Module
248
RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified – conformance/non-conformance drops to ~ 15-25% sigma = 33% of datum to be modified – conformance/non-conformance drops to ~ 5-10% Results, One Module
249
RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified – conformance/non-conformance drops to ~ 15-25% sigma = 33% of datum to be modified – conformance/non-conformance drops to ~ 5-10% for 25% & 33% cases, sensitivities appear to non-module maximum expression coefficient distribution mu and sigma Results, One Module
250
RSA with noise and dynamic network parameters Results, Relaxed Module
251
RSA with noise and dynamic network parameters – Pipeline parameters fixed, MI calculation performed Results, Relaxed Module
252
RSA with noise and dynamic network parameters – Pipeline parameters fixed, MI calculation performed – Module parameters picked from a wider distribution Results, Relaxed Module
253
RSA with noise and dynamic network parameters – Pipeline parameters fixed, MI calculation performed – Module parameters picked from a wider distribution – Sensitive to rate at which module parameters reach equilibrium, and percentage of time module is ‘on’ in the data Results, Relaxed Module
254
RSA with noise and dynamic network parameters – Pipeline parameters fixed, MI calculation performed – Module parameters picked from a wider distribution – Sensitive to rate at which module parameters reach equilibrium, and percentage of time module is ‘on’ in the data – Conformance/non-conformance ratio is ~ 35-50% Results, Relaxed Module
255
MI calculation Results, Relaxed Module
256
MI calculation – Invoked only when PC + Z-score fails to infer an edge Results, Relaxed Module
257
MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Results, Relaxed Module
258
MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Hypothesis is that module members are expressed in quantities that are not vastly different from each other Results, Relaxed Module
259
MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Hypothesis is that module members are expressed in quantities that are not vastly different from each other – Invoked only if expression levels are not small Results, Relaxed Module
260
MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Hypothesis is that module members are expressed in quantities that are not vastly different from each other – Invoked only if expression levels are not small At levels considered ‘noise’ Results, Relaxed Module
261
MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Hypothesis is that module members are expressed in quantities that are not vastly different from each other – Invoked only if expression levels are not small At levels considered ‘noise’ – Typically increases conformance/non-conformance ratios by ~ 0-5% Results, Relaxed Module
262
Results, Two Modules
263
Most realistic case yet Results, Two Modules
264
Most realistic case yet With PC, Z-score, and noise, pipeline is sensitive to rate at which module genes reach equilibrium, and sensitive to Z-score cutoff value, as well as percentage of time module is ‘on’ in the data Results, Two Modules
265
Most realistic case yet With PC, Z-score, and noise, pipeline is sensitive to rate at which module genes reach equilibrium, and sensitive to Z-score cutoff value, as well as percentage of time module is ‘on’ in the data – conforming/non-conforming ratio is ~ 25-40% Results, Two Modules
266
Most realistic case yet With PC, Z-score, and noise, pipeline is sensitive to rate at which module genes reach equilibrium, and sensitive to Z-score cutoff value, as well as percentage of time module is on in the data – conforming/non-conforming ratio is ~ 25-40% With the addition of an MI calculation, the only sensitivity is percentage of time module is ‘on’ in the data Results, Two Modules
267
Most realistic case yet With PC, Z-score, and noise, pipeline is sensitive to rate at which module genes reach equilibrium, and sensitive to Z-score cutoff value, as well as percentage of time module is on in the data – conforming/non-conforming ratio is ~ 25-40% With the addition of an MI calculation, the only sensitivity is percentage of time module is ‘on’ in the data – conforming/non-conforming ratio is ~ 30-45% Results, Two Modules
268
With only a sensitivity to the percentage of time the module is ‘on’ in the data Results, Two Modules
269
With only a sensitivity to the percentage of time the module is ‘on’ in the data – Pipeline is robust in the face of network variation Results, Two Modules
270
With only a sensitivity to the percentage of time the module is ‘on’ in the data – Pipeline is robust in the face of network variation – Can start the pipeline ‘support’ parameter at a high value, Results, Two Modules
271
With only a sensitivity to the percentage of time the module is ‘on’ in the data – Pipeline is robust in the face of network variation – Can start the pipeline ‘support’ parameter at a high value, and turn it down until modules are detected! Results, Two Modules
274
Conclusions
275
NEMO – Network Motif language developed Conclusions
276
– Language translator for a qualitative transcription network description to a quantitative SBML model Conclusions
277
NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator Conclusions
278
NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator To generate synthetic gene expression data Conclusions
279
NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator To generate synthetic gene expression data – Microarray data Conclusions
280
NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator To generate synthetic gene expression data – Microarray data – NGS data Conclusions
281
NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator To generate synthetic gene expression data – Microarray data – NGS data – Bioinformatics (2008) 24 (1): 132-134. Conclusions
282
Conclusions ODES – Overlapping Dense Subgraph Algorithm
283
Conclusions – In the class of exponentially exact algorithms
284
Conclusions ODES – Overlapping Dense Subgraph Algorithm – In the class of exponentially exact algorithms – Confines brute-force search domain to actual dense subgraphs
285
Conclusions ODES – Overlapping Dense Subgraph Algorithm – In the class of exponentially exact algorithms – Confines brute-force search domain to actual dense subgraphs – Bioinformatics (2010) 26 (21): 2788-2789.
286
Open source CODENSE algorithm developed Conclusions
287
– Improved expression correlation algorithms Conclusions
288
Open source CODENSE algorithm developed – Improved expression correlation algorithms – Uses ODES dense subgraph algorithm Conclusions
289
Open source CODENSE algorithm developed – Improved expression correlation algorithms – Uses ODES dense subgraph algorithm – Successful identification of modules from synthetic data Conclusions
290
Open source CODENSE algorithm developed – Improved expression correlation algorithms – Uses ODES dense subgraph algorithm – Successful identification of modules from synthetic data – Manuscript in preparation Conclusions
291
Regionalized Sensitivity Analysis Performed Conclusions
292
– Pipeline is insensitive to reasonably chosen parameters Conclusions
293
Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise Conclusions
294
Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise, except for the ‘support’ parameter Conclusions
295
Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise, except for the ‘support’ parameter – Pipeline is insensitive to transcription network variability Conclusions
296
Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise, except for the ‘support’ parameter – Pipeline is insensitive to transcription network variability – Pipeline is robust in the face of noise Conclusions
297
Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise, except for the ‘support’ parameter – Pipeline is insensitive to transcription network variability – Pipeline is robust in the face of noise – Up to a 50% conformance/non-conformance ratio Conclusions
298
Future Work
299
Package the pipeline code for use by researchers. Future Work
300
Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Future Work
301
Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Future Work
302
Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Test on larger networks. Future Work
303
Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Test on larger networks. Test on real data! Future Work
304
Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Test on larger networks. Test on real data! Investigate spanning tree initialization of ODES Future Work
305
Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Test on larger networks. Test on real data! Investigate spanning tree initialization of ODES – Changes exact algorithm into high-probability heuristic Future Work
306
Thank You
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.