Presentation is loading. Please wait.

Presentation is loading. Please wait.

A SENSITIVITY ANALYSIS OF A BIOLOGICAL MODULE DISCOVERY PIPELINE James Long International Arctic Research Center University of Alaska Fairbanks March 25,

Similar presentations


Presentation on theme: "A SENSITIVITY ANALYSIS OF A BIOLOGICAL MODULE DISCOVERY PIPELINE James Long International Arctic Research Center University of Alaska Fairbanks March 25,"— Presentation transcript:

1 A SENSITIVITY ANALYSIS OF A BIOLOGICAL MODULE DISCOVERY PIPELINE James Long International Arctic Research Center University of Alaska Fairbanks March 25, 2015

2 A Sensitivity Analysis of a Biological Module Discovery Pipeline

3 Gene Expression A Sensitivity Analysis of a Biological Module Discovery Pipeline

4 Gene Expression Synthetic Gene Expression Data A Sensitivity Analysis of a Biological Module Discovery Pipeline

5 Gene Expression Synthetic Gene Expression Data CODENSE A Sensitivity Analysis of a Biological Module Discovery Pipeline

6 Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis A Sensitivity Analysis of a Biological Module Discovery Pipeline

7 Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline

8 Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline

9 Gene Expression

10

11

12

13

14

15

16

17 Hill Function

18 Gene Expression Hill Function

19 Gene Expression Hill Function,

20 Gene Expression

21 Hill Function

22 Generalized Hill Function

23

24 for activators

25 Generalized Hill Function for activators for repressors

26 Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline

27 Synthetic Gene Expression Data

28 NEMO

29 Synthetic Gene Expression Data NEMO – Network Motif Language

30 Synthetic Gene Expression Data NEMO – Network Motif Language COPASI

31 Synthetic Gene Expression Data NEMO – Network Motif Language COPASI – Complex Pathway Simulator

32 Synthetic Gene Expression Data NEMO – Network Motif Language COPASI – Complex Pathway Simulator

33 NEMO

34 G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

35 G0 G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

36 NEMO G0( G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

37 NEMO G0(P1 G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

38 NEMO G0(P1+ G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

39 NEMO G0(P1+, G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

40 NEMO G0(P1+,P2+, G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

41 NEMO G0(P1+,P2+,P3-, G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

42 NEMO G0(P1+,P2+,P3-,P4-, G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

43 NEMO G0(P1+,P2+,P3-,P4-,P5+) G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

44 NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

45 NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G2(…) G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

46 NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G2(…) etc. G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

47 NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G2(…) etc.) GLIST( G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

48 NEMO G0(P1+,P2+,P3-,P4-,P5+) G1(…) G2(…) etc.)] GLIST( [ G0 G0 G1 G1 G2G2 G3 G3 G4 G4 G5 G5

49 Dense Overlapping Regulon (DOR) NEMO G0G1 G3 G4G5 G2

50 G3(P0+,P1+,P2-) NEMO Dense Overlapping Regulon (DOR) G0G1 G3 G4G5 G2

51 G3(P0+,P1+,P2-), G4(P0+,P1+) NEMO Dense Overlapping Regulon (DOR) G0G1 G3 G4G5 G2

52 G3(P0+,P1+,P2-), G4(P0+,P1+), G5(P1-,P2+) NEMO Dense Overlapping Regulon (DOR) G0G1 G3 G4G5 G2

53 G0G1 G3 G4G5 G2 DOR(G3(P0+,P1+,P2-), G4(P0+,P1+), G5(P1-,P2+)) NEMO Dense Overlapping Regulon (DOR)

54 Negative auto-regulation NEMO G0

55 Negative auto-regulation G0 G0(P0-) NEMO

56 Feed-forward loop (FFL) G0G1 G2 NEMO

57 G0G1 G2 P0 NEMO Feed-forward loop (FFL)

58 G0G1 G2 P0(+G1 NEMO Feed-forward loop (FFL)

59 G0G1 G2 P0(+G1+G2 NEMO Feed-forward loop (FFL)

60 G0G1 G2 P0(+G1+G2+ NEMO Feed-forward loop (FFL)

61 G0G1 G2 P0(+G1+G2+) NEMO Feed-forward loop (FFL)

62 Multi-output FFL G0G1 G2 NEMO

63 G0G1 G2 G3 NEMO Multi-output FFL

64 G0G1 G2 G3G4 NEMO Multi-output FFL

65 P0(+G1+G2+) G0G1 G2 G3G4 NEMO Multi-output FFL

66 P0(+G1+(G2,G3,G4)+) G0G1 G2 G3G4 NEMO Multi-output FFL

67 P0(+G1+(G2,G3,G4)+)) G0G1 G2 G3G4 NEMO Multi-output FFL TMLIST(

68 Single-input module (SIM) G0 G1 G2G3 NEMO

69 P0(+G1,G2,G3) G0 G1 G2G3 NEMO Single-input module (SIM)

70 G0G1G2 G3G4G5G6 G7G8G9G10 NEMO

71 G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-) NEMO

72 G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-) NEMO

73 G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6) NEMO

74 G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), G7(P3+,P4-,P5+) NEMO

75 G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), G7(P3+,P4-,P5+), G8(P4+) NEMO

76 G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+) NEMO

77 G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-) NEMO

78 G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), P0(+G1-G2-), P1(+G3,G4,G5,G6), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-)) NEMO

79 G0G1G2 G3G4G5G6 G7G8G9G10 G0(P5-), TMLIST(P0(+G1-G2-), P1(+G3,G4,G5,G6)), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-)) NEMO

80 G0G1G2 G3G4G5G6 G7G8G9G10 GLIST(G0(P5-)), TMLIST(P0(+G1-G2-), P1(+G3,G4,G5,G6)), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-)) NEMO

81 G0G1G2 G3G4G5G6 G7G8G9G10 [ GLIST(G0(P5-)), TMLIST(P0(+G1-G2-), P1(+G3,G4,G5,G6)), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5-)) ] NEMO

82 G0G1G2 G3G4G5G6 G7G8G9G10 [ GLIST(G0(P5-)), TMLIST(P0(+G1-G2-), P1(+G3,G4,G5,G6)), DOR(G7(P3+,P4-,P5+), G8(P4+), G9(P3+,P5+,P6+), G10(P5- :F(power(sin(P5),2)))) ] NEMO

83 NEMO compiler emits SBML

84 NEMO NEMO compiler emits SBML – Systems Biology Markup Language

85 NEMO NEMO compiler emits SBML – Systems Biology Markup Language – Uses libsbml from http://sbml.orghttp://sbml.org

86 NEMO NEMO compiler emits SBML – Systems Biology Markup Language – Uses libsbml from http://sbml.orghttp://sbml.org More accurate to call it a “language translator”

87 NEMO NEMO compiler emits SBML – Systems Biology Markup Language – Uses libsbml from http://sbml.orghttp://sbml.org More accurate to call it a “language translator”, that adds random generalized Hill functions!

88 NEMO NEMO compiler emits SBML – Systems Biology Markup Language – Uses libsbml from http://sbml.orghttp://sbml.org More accurate to call it a “language translator”, that adds random generalized Hill functions! Bioinformatics 2008 Jan 1;24(1):132-4

89 NEMO G0 G0 G1 G1 [GLIST(G0(P1+), G1(P0+)) ]

90 NEMO

91 NEMO B_0 P1 K_0 n_0 1 P1 K_0 n_0 tau

92 NEMO dc_0 P0 tau

93 NEMO P0 tau B_1 P0 K_1 n_1 1

94 NEMO P0 K_1 n_1 tau dc_1 P1 tau

95 NEMO P1 tau

96 Synthetic Gene Expression Data NEMO – Network Motif Language COPASI – Complex Pathway Simulator

97

98

99

100

101 Also outputs a table of data for each time step.

102 COPASI – Complex Pathway Simulator Also outputs a table of data for each time step, the last column of which is our synthetic data!

103 COPASI – Complex Pathway Simulator

104 Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline

105 The CODENSE algorithm

106 Input – a series of expression correlation graphs, each representing a different state for an organism.

107 The CODENSE algorithm Input – a series of expression correlation graphs, each representing a different state for an organism. Output – groups of genes (modules) whose expression is correlated across the series of expression correlation graphs.

108

109 Expression Correlation

110 Pearson’s Correlation

111 Expression Correlation Pearson’s Correlation – Linear dependence between two variables

112 Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score

113 Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score – Number of standard deviations above the mean

114 Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score – Number of standard deviations above the mean Mutual Information

115 Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score – Number of standard deviations above the mean Mutual Information – A measure of mutual dependence between variables

116 Expression Correlation Pearson’s Correlation – Linear dependence between two variables Pearson’s Correlation with Z-score – Number of standard deviations above the mean Mutual Information – A measure of mutual dependence between variables – Non-linear dependence OK

117

118

119

120

121

122

123

124

125

126

127

128

129

130

131

132

133

134

135 The CODENSE algorithm

136 Manuscript in preparation

137 ODES – Overlapping Dense Subgraph Algorithm

138

139 vertex

140 edge

141 vertex edge connected

142 vertex edge connected

143 vertex edge connected cut vertex

144 vertex edge cut vertex

145 vertex edge cut vertex disconnected

146 vertex edge disconnected

147

148

149 ODES Density of a graph

150 ODES Number of actual edges / Number of possible edges

151 ODES Density of a graph Number of actual edges / Number of possible edges Number of possible edges

152 ODES Density of a graph Number of actual edges / Number of possible edges Number of possible edges Density

153 ODES Density of a graph Number of actual edges / Number of possible edges Number of possible edges Density Degree of a vertex

154 ODES Density of a graph Number of actual edges / Number of possible edges Number of possible edges Density Degree of a vertex Number of edges incident to the vertex

155 ODES Theorem

156 ODES A connected graph, with density, and number of vertices, has at least one non-cut vertex where degree, the average degree of vertices in.

157 ODES Theorem A connected graph, with density, and number of vertices, has at least one non-cut vertex where degree, the average degree of vertices in. Removal of from does not decrease the density of.

158 ODES Theorem A connected graph, with density, and number of vertices, has at least one non-cut vertex where degree, the average degree of vertices in. Removal of from does not decrease the density of. Bioinformatics (2010) 26 (21): 2788-2789.

159

160 8 vertices

161 22 edges

162 8 vertices 22 edges average degree = 44/8 = 5.5

163 8 vertices 22 edges average degree = 44/8 = 5.5 density = 2*22/(8(8-1)) ~ 0.78

164 8 vertices 22 edges average degree = 44/8 = 5.5 density = 2*22/(8(8-1)) ~ 0.78

165 8 vertices 22 edges average degree = 5.5 density ~ 0.78

166 7 vertices 17 edges average degree ~ 4.86 density ~ 0.81

167 7 vertices 17 edges average degree ~ 4.86 density ~ 0.81

168 6 vertices 13 edges average degree ~ 4.33 density ~ 0.87

169 6 vertices 13 edges average degree ~ 4.33 density ~ 0.87

170 5 vertices 9 edges average degree = 3.6 density ~ 0.9

171 5 vertices 9 edges average degree = 3.6 density ~ 0.9

172 4 vertices 6 edges average degree = 3.0 density = 1.0

173 4 vertices 6 edges average degree = 3.0 density = 1.0

174 3 vertices 3 edges average degree = 2.0 density = 1.0

175 3 vertices 3 edges average degree = 2.0 density = 1.0

176 2 vertices 1 edge average degree = 1.0 density = 1.0

177 2 vertices 1 edge average degree = 1.0 density = 1.0

178 2 vertices 1 edge average degree = 1.0 density = 1.0

179 3 vertices 3 edges average degree = 2.0 density = 1.0

180 3 vertices 3 edges average degree = 2.0 density = 1.0

181 4 vertices 6 edges average degree = 3.0 density = 1.0

182 4 vertices 6 edges average degree = 3.0 density = 1.0

183 5 vertices 9 edges average degree = 3.6 density ~ 0.9

184 5 vertices 9 edges average degree = 3.6 density ~ 0.9

185 6 vertices 13 edges average degree ~ 4.33 density ~ 0.87

186 6 vertices 13 edges average degree ~ 4.33 density ~ 0.87

187 7 vertices 17 edges average degree ~ 4.86 density ~ 0.81

188 7 vertices 17 edges average degree ~ 4.86 density ~ 0.81

189 8 vertices 22 edges average degree = 5.5 density ~ 0.78

190 8 vertices 22 edges average degree = 5.5 density ~ 0.78

191 9 vertices 24 edges average degree ~ 5.3 density ~ 0.67

192 8 vertices 22 edges average degree = 5.5 density ~ 0.78

193 8 vertices 22 edges average degree = 5.5 density ~ 0.78 Note: brute-force search is confined to actual dense subgraphs.

194 Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline

195 Regionalized Sensitivity Analysis

196 Monte Carlo model runs Regionalized Sensitivity Analysis

197 Monte Carlo model runs Evaluate one or more binary Objective Functions Regionalized Sensitivity Analysis

198 Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? Regionalized Sensitivity Analysis

199 Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? – Exact modules returned w/ limited false positives? Regionalized Sensitivity Analysis

200 Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? – Exact modules returned w/ limited false positives? – Approximate modules returned w/ limited false positives? Regionalized Sensitivity Analysis

201 Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? – Exact modules returned w/ limited false positives? – Approximate modules returned w/ limited false positives? – Half of known modules returned approximately w/ limited false positives? Regionalized Sensitivity Analysis

202 Monte Carlo model runs Evaluate one or more binary Objective Functions – Are only exact known modules returned? – Exact modules returned w/ limited false positives? – Approximate modules returned w/ limited false positives? – Half of known modules returned approximately w/ limited false positives? Increment parameter bins based on Objective Function conformance or non-conformance Regionalized Sensitivity Analysis

203

204

205

206

207

208

209

210

211 Compile the NEMO representation of the canonical network Regionalized Sensitivity Analysis

212 Compile the NEMO representation of the canonical network Import into COPASI Regionalized Sensitivity Analysis

213 Compile the NEMO representation of the canonical network Import into COPASI Set up model, and save in COPASI format Regionalized Sensitivity Analysis

214 Compile the NEMO representation of the canonical network Import into COPASI Set up model, and save in COPASI format Create template from COPASI format file where all genes are turned off (B = 0) Regionalized Sensitivity Analysis

215 Compile the NEMO representation of the canonical network Import into COPASI Set up model, and save in COPASI format Create template from COPASI format file where all genes are turned off (B = 0) Create synthetic data by turning some genes and SIMs on, taking last column of COPASI output. Regionalized Sensitivity Analysis

216 Compile the NEMO representation of the canonical network Import into COPASI Set up model, and save in COPASI format Create template from COPASI format file where all genes are turned off (B = 0) Create synthetic data by turning some genes and SIMs on, taking last column of COPASI output. Add noise to output. Regionalized Sensitivity Analysis

217 Gene Expression Synthetic Gene Expression Data CODENSE Regionalized Sensitivity Analysis Results A Sensitivity Analysis of a Biological Module Discovery Pipeline

218 Results, One Module

219 First RSA runs used a static transcription network Results, One Module

220 First RSA runs used a static transcription network – For Objective Function 1, no noise added Results, One Module

221 First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation Results, One Module

222 First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 Results, One Module

223 First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio Results, One Module

224 First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio For Z-scores, sensitivity to PC cutoff score greatly attenuated Results, One Module

225 First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio For Z-scores, sensitivity to PC cutoff score greatly attenuated – Lower PC scores allow more false positives to enter pipeline Results, One Module

226 First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio For Z-scores, sensitivity to PC cutoff score greatly attenuated – Lower PC scores allow more false positives to enter pipeline – Significant sensitivities to similarity cutoff score and minimum density for a dense subgraph in coherent dense subgraphs Results, One Module

227 First RSA runs used a static transcription network – For Objective Function 1, no noise added Highly sensitive to PC cutoff score for PC only correlation – 0.999 to 1.0 – Very low conformance/non-conformance ratio For Z-scores, sensitivity to PC cutoff score greatly attenuated – Lower PC scores allow more false positives to enter pipeline – Significant sensitivities to similarity cutoff score and minimum density for a dense subgraph in coherent dense subgraphs – Still a very low conformance/non-conformance ratio Results, One Module

228 First RSA runs used a static transcription network – For other Objective Functions Results, One Module

229 First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Results, One Module

230 First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Results, One Module

231 First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Observations Results, One Module

232 First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Observations – No good at finding the exact module Results, One Module

233 First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Observations – No good at finding the exact module – No noise is unrealistic Results, One Module

234 First RSA runs used a static transcription network – For other Objective Functions Conformance/non-conformance ratio rises to ~ 50% Only sensitive to the parameter that declares the fraction of data sets that an edge must exist in to be included in the summary graph Observations – No good at finding the exact module – No noise is unrealistic – Static network parameters are unrealistic Results, One Module

235 RSA with noise and dynamic network parameters Results, One Module

236 RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum Results, One Module

237 RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% Results, One Module

238 RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% – Jumps to 30-45% with Z-score methods Results, One Module

239 RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% – Jumps to 30-45% with Z-score methods – Sensitivities mostly disappear in the presence of noise Results, One Module

240 RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% – Jumps to 30-45% with Z-score methods – Sensitivities mostly disappear in the presence of noise sensitive to how fast module genes reach equilibrium Results, One Module

241 RSA with noise and dynamic network parameters – normally distributed noise, mu=0, sigma=10% of datum – Still low conformance/non-conformance ratios for PC only, ~ 10% – Jumps to 30-45% with Z-score methods – Sensitivities mostly disappear in the presence of noise sensitive to how fast module genes reach equilibrium sensitive to percentage of time a module is ‘on’ in the data Results, One Module

242 RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods Results, One Module

243 RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified Results, One Module

244 RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% Results, One Module

245 RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified Results, One Module

246 RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified – conformance/non-conformance drops to ~ 15-25% Results, One Module

247 RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified – conformance/non-conformance drops to ~ 15-25% sigma = 33% of datum to be modified Results, One Module

248 RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified – conformance/non-conformance drops to ~ 15-25% sigma = 33% of datum to be modified – conformance/non-conformance drops to ~ 5-10% Results, One Module

249 RSA with noise and dynamic network parameters – Different amounts of noise, Z-score methods sigma = 20% of datum to be modified – conformance/non-conformance drops to ~ 25-35% sigma = 25% of datum to be modified – conformance/non-conformance drops to ~ 15-25% sigma = 33% of datum to be modified – conformance/non-conformance drops to ~ 5-10% for 25% & 33% cases, sensitivities appear to non-module maximum expression coefficient distribution mu and sigma Results, One Module

250 RSA with noise and dynamic network parameters Results, Relaxed Module

251 RSA with noise and dynamic network parameters – Pipeline parameters fixed, MI calculation performed Results, Relaxed Module

252 RSA with noise and dynamic network parameters – Pipeline parameters fixed, MI calculation performed – Module parameters picked from a wider distribution Results, Relaxed Module

253 RSA with noise and dynamic network parameters – Pipeline parameters fixed, MI calculation performed – Module parameters picked from a wider distribution – Sensitive to rate at which module parameters reach equilibrium, and percentage of time module is ‘on’ in the data Results, Relaxed Module

254 RSA with noise and dynamic network parameters – Pipeline parameters fixed, MI calculation performed – Module parameters picked from a wider distribution – Sensitive to rate at which module parameters reach equilibrium, and percentage of time module is ‘on’ in the data – Conformance/non-conformance ratio is ~ 35-50% Results, Relaxed Module

255 MI calculation Results, Relaxed Module

256 MI calculation – Invoked only when PC + Z-score fails to infer an edge Results, Relaxed Module

257 MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Results, Relaxed Module

258 MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Hypothesis is that module members are expressed in quantities that are not vastly different from each other Results, Relaxed Module

259 MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Hypothesis is that module members are expressed in quantities that are not vastly different from each other – Invoked only if expression levels are not small Results, Relaxed Module

260 MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Hypothesis is that module members are expressed in quantities that are not vastly different from each other – Invoked only if expression levels are not small At levels considered ‘noise’ Results, Relaxed Module

261 MI calculation – Invoked only when PC + Z-score fails to infer an edge – Invoked only if expression levels are comparable Hypothesis is that module members are expressed in quantities that are not vastly different from each other – Invoked only if expression levels are not small At levels considered ‘noise’ – Typically increases conformance/non-conformance ratios by ~ 0-5% Results, Relaxed Module

262 Results, Two Modules

263 Most realistic case yet Results, Two Modules

264 Most realistic case yet With PC, Z-score, and noise, pipeline is sensitive to rate at which module genes reach equilibrium, and sensitive to Z-score cutoff value, as well as percentage of time module is ‘on’ in the data Results, Two Modules

265 Most realistic case yet With PC, Z-score, and noise, pipeline is sensitive to rate at which module genes reach equilibrium, and sensitive to Z-score cutoff value, as well as percentage of time module is ‘on’ in the data – conforming/non-conforming ratio is ~ 25-40% Results, Two Modules

266 Most realistic case yet With PC, Z-score, and noise, pipeline is sensitive to rate at which module genes reach equilibrium, and sensitive to Z-score cutoff value, as well as percentage of time module is on in the data – conforming/non-conforming ratio is ~ 25-40% With the addition of an MI calculation, the only sensitivity is percentage of time module is ‘on’ in the data Results, Two Modules

267 Most realistic case yet With PC, Z-score, and noise, pipeline is sensitive to rate at which module genes reach equilibrium, and sensitive to Z-score cutoff value, as well as percentage of time module is on in the data – conforming/non-conforming ratio is ~ 25-40% With the addition of an MI calculation, the only sensitivity is percentage of time module is ‘on’ in the data – conforming/non-conforming ratio is ~ 30-45% Results, Two Modules

268 With only a sensitivity to the percentage of time the module is ‘on’ in the data Results, Two Modules

269 With only a sensitivity to the percentage of time the module is ‘on’ in the data – Pipeline is robust in the face of network variation Results, Two Modules

270 With only a sensitivity to the percentage of time the module is ‘on’ in the data – Pipeline is robust in the face of network variation – Can start the pipeline ‘support’ parameter at a high value, Results, Two Modules

271 With only a sensitivity to the percentage of time the module is ‘on’ in the data – Pipeline is robust in the face of network variation – Can start the pipeline ‘support’ parameter at a high value, and turn it down until modules are detected! Results, Two Modules

272

273

274 Conclusions

275 NEMO – Network Motif language developed Conclusions

276 – Language translator for a qualitative transcription network description to a quantitative SBML model Conclusions

277 NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator Conclusions

278 NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator To generate synthetic gene expression data Conclusions

279 NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator To generate synthetic gene expression data – Microarray data Conclusions

280 NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator To generate synthetic gene expression data – Microarray data – NGS data Conclusions

281 NEMO – Network Motif language developed – Language translator for a qualitative transcription network description to a quantitative SBML model – Used as input to the COPASI biochemical simulator To generate synthetic gene expression data – Microarray data – NGS data – Bioinformatics (2008) 24 (1): 132-134. Conclusions

282 Conclusions ODES – Overlapping Dense Subgraph Algorithm

283 Conclusions – In the class of exponentially exact algorithms

284 Conclusions ODES – Overlapping Dense Subgraph Algorithm – In the class of exponentially exact algorithms – Confines brute-force search domain to actual dense subgraphs

285 Conclusions ODES – Overlapping Dense Subgraph Algorithm – In the class of exponentially exact algorithms – Confines brute-force search domain to actual dense subgraphs – Bioinformatics (2010) 26 (21): 2788-2789.

286 Open source CODENSE algorithm developed Conclusions

287 – Improved expression correlation algorithms Conclusions

288 Open source CODENSE algorithm developed – Improved expression correlation algorithms – Uses ODES dense subgraph algorithm Conclusions

289 Open source CODENSE algorithm developed – Improved expression correlation algorithms – Uses ODES dense subgraph algorithm – Successful identification of modules from synthetic data Conclusions

290 Open source CODENSE algorithm developed – Improved expression correlation algorithms – Uses ODES dense subgraph algorithm – Successful identification of modules from synthetic data – Manuscript in preparation Conclusions

291 Regionalized Sensitivity Analysis Performed Conclusions

292 – Pipeline is insensitive to reasonably chosen parameters Conclusions

293 Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise Conclusions

294 Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise, except for the ‘support’ parameter Conclusions

295 Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise, except for the ‘support’ parameter – Pipeline is insensitive to transcription network variability Conclusions

296 Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise, except for the ‘support’ parameter – Pipeline is insensitive to transcription network variability – Pipeline is robust in the face of noise Conclusions

297 Regionalized Sensitivity Analysis Performed – Pipeline is insensitive to reasonably chosen parameters In the presence of noise, except for the ‘support’ parameter – Pipeline is insensitive to transcription network variability – Pipeline is robust in the face of noise – Up to a 50% conformance/non-conformance ratio Conclusions

298 Future Work

299 Package the pipeline code for use by researchers. Future Work

300 Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Future Work

301 Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Future Work

302 Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Test on larger networks. Future Work

303 Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Test on larger networks. Test on real data! Future Work

304 Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Test on larger networks. Test on real data! Investigate spanning tree initialization of ODES Future Work

305 Package the pipeline code for use by researchers. Make sensitivity runs on the canonical network, where more than two modules have the opportunity of being turned on. Sensitivity runs on different network topologies. Test on larger networks. Test on real data! Investigate spanning tree initialization of ODES – Changes exact algorithm into high-probability heuristic Future Work

306 Thank You


Download ppt "A SENSITIVITY ANALYSIS OF A BIOLOGICAL MODULE DISCOVERY PIPELINE James Long International Arctic Research Center University of Alaska Fairbanks March 25,"

Similar presentations


Ads by Google