Download presentation
Presentation is loading. Please wait.
Published byHomer Thompson Modified over 9 years ago
1
A Cooperative Coevolutionary Genetic Algorithm for Learning Bayesian Network Structures Arthur Carvalho a3carval@cs.uwaterloo.ca
2
Outline Bayesian Networks CCGA Experiments Conclusion
3
Bayesian Networks AI technique –Diagnosis, predictions, modelling knowledge Graphical model –Represents a joint distribution over a set of random variables –Exploits conditional independence –Concise, natural representation
4
Bayesian Networks X1X1 X1X1 X2X2 X2X2 X3X3 X3X3 TT TF FT FF
5
X1X1 X1X1 X2X2 X2X2 X3X3 X3X3 T F
6
Directed acyclic graph (DAG) –Nodes: random variables –Edges: direct influence of one variable on another Each node is associated with a conditional probability distribution (CPD)
7
Bayesian Networks Learning the structure of the network (DAG) –Structure learning problem Learning parameters that define the CPDs –Parameter estimation –Maximum Likelihood estimation
8
Bayesian Networks Structure learning problem in fully observable datasets –Find a DAG that maximizes P(DAG | Data) –[Cooper & Herskovits, 92]
9
Bayesian Networks NP-Hard [Chickering et al, 1994] –The number of possible structures is superexponential in the number of nodes [Robinson, 1977] –For a network with n nodes, the number of different structures is:
10
Outline Bayesian Networks CCGA Experiments Conclusion
11
CCGA Structure learning task can be decomposed into two dependent subtasks –To find an optimal ordering of the nodes –To find an optimal connectivity matrix
12
CCGA D D A A B B C C DABC 110 D
13
D D A A B B C C DABC 110 D
14
D D A A B B C C DABC 110 D
15
D D A A B B C C DABC 110 D
16
D D A A B B C C DABC 11001 A
17
D D A A B B C C DABC 11001 A
18
D D A A B B C C DABC 11001 A
19
D D A A B B C C DABC 11001 B 1
20
D D A A B B C C D110A01B1C DABC 11001 1
21
Two subpopulations –Binary (edges) –Permutation (nodes) Cooperative Coevolutionary Genetic Algorithm (CCGA) –Each subpopulation is coevolve using a canonical GA
22
CCGA Evaluating individual species –Each subpopulation member is combined with both the best known individual and a random individual from the other subpopulation –The fitness function is applied to the two resulting solutions The highest value is the fitness CCGA-2 [Potter & De Jong, 1994]
23
CCGA OperatorBinaryPermutation SelectionTournament selection CrossoverTwo-point crossoverCycle crossover MutationBit-flip mutationSwap mutation ReplacementPreserve the best solution
24
Outline Bayesian Networks CCGA Experiments Conclusion
25
Experiments Setup –K2 algorithm [Cooper & Herskovits, 1992] –Alarm network 37 nodes and 46 edges –Insurance network 27 nodes and 52 edges –Three datasets 1000, 3000, and 5000 instances –100 executions
26
Experiments Parameters: ParameterValue Generations250 Population size100 Probability crossover0.6 Probability mutation: binary population 1 / max # of edges Probability mutation: permutation population0.5
27
Experiments Alarm network DatasetAlgorithmAverage Standard Deviation p-value Alarm 1000 −11,569.02 CCGA−12,166.21178.23 0.0067 K2−12,226.24161.33 Alarm 3000 −33,759.28 CCGA−35,020.31396.06 0.0125 K2−35,138.41341.31 Alarm 5000 −55,575.11 CCGA−57,282.66548.55 < 0.0001 K2−57,574.12492.94
28
Experiments Insurance network DatasetAlgorithmAverage Standard Deviation p-value Insurance 1000 −15,397.00 CCGA−15,787.28279.19 < 0.0001 K2−16,107.61293.06 Insurance 3000 −43,508.63 CCGA−45,142.37722.67 < 0.0001 K2−45,778.39721.08 Insurance 5000 −72,183.51 CCGA−74,624.88995.01 < 0.0001 K2−75,837.821,249.83
29
Outline Bayesian Networks CCGA Experiments Conclusion
30
New algorithm to solve the structure learning problem –Novel representation –Good performance Future work –Incomplete datasets –Graph-related problems
31
Thank you! Source code and datasets available at: www.cs.uwaterloo.ca/~a3carval/softwares/CCGA_code.rar Arthur Carvalho a3carval@cs.uwaterloo.ca
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.