Extending Evolutionary Programming to the Learning of Dynamic Bayesian Networks Allan Tucker Xiaohui Liu Birkbeck College University of London.

Slides:



Advertisements
Similar presentations
Esri International User Conference | San Diego, CA Technical Workshops | Kevin M. Johnston Shitij Mehta ****************** An Introduction to Dynamic Simulation.
Advertisements

Asset Management Optimization using model based decision support Speaker: Francesco Verre SPE Dinner Meeting – 25 th October 2011 – London.
Dynamic Bayesian Networks (DBNs)
Introduction of Probabilistic Reasoning and Bayesian Networks
Bayesian networks and how they can help us to explore fish species interaction in the Northern gulf of St Lawrence Dr Allan Tucker Centre for Intelligent.
1 A Framework for Modelling Short, High-Dimensional Multivariate Time Series: Preliminary Results in Virus Gene Expression Data Analysis Paul Kellam 1,
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
1 Lecture 8: Genetic Algorithms Contents : Miming nature The steps of the algorithm –Coosing parents –Reproduction –Mutation Deeper in GA –Stochastic Universal.
1 Grouping Multivariate Time Series Variables: Applications to Chemical Process and Visual Field Data Allan Tucker- Birkbeck College Stephen Swift- Brunel.
Spatial Operators for Evolving Dynamic Bayesian Networks from Spatio-Temporal Data Allan Tucker Xiaohui Liu David Garway-Heath Moorfields Eye Hospital.
Temporal Processes Eran Segal Weizmann Institute.
Learning Dynamic Bayesian Networks with Changing Dependencies Allan Tucker Xiaohui Liu IDA 2003.
IMA, October 29, 2007 Slide 1 T H E B I O I N F O R M A T I C S C E N T R E A continuous probabilistic model of local RNA 3-D structure Jes Frellsen The.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Who am I and what am I doing here? Allan Tucker A brief introduction to my research
The Automatic Explanation of Multivariate Time Series (MTS) Allan Tucker.
Bridging the Gap between Applications and Tools: Modeling Multivariate Time Series X Liu, S Swift & A Tucker Department of Computer Science Birkbeck College.
Tutorial 1 Temi avanzati di Intelligenza Artificiale - Lecture 3 Prof. Vincenzo Cutello Department of Mathematics and Computer Science University of Catania.
Local Search and Stochastic Algorithms
Explaining Multivariate Time Series to Detect Early Problem Signs Architectures and Efficient Learning Algorithms for Dynamic Bayesian Networks Allan Tucker,
Making the Most of Small Sample High Dimensional Micro-Array Data Allan Tucker, Veronica Vinciotti, Xiaohui Liu; Brunel University Paul Kellam; Windeyer.
Bayesian Classification and Forecasting of Visual Field Deterioration Allan Tucker, Xiaohui Liu; Brunel University David Garway-Heath; Moorfield’s Eye.
Dynamic Bayesian Networks CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanningLearning.
Tutorial 4 (Lecture 12) Remainder from lecture 10: The implicit fitness sharing method Exercises – any solutions? Questions?
Gaussian process modelling
Genetic Algorithm.
1 Performance Evaluation of Computer Networks: Part II Objectives r Simulation Modeling r Classification of Simulation Modeling r Discrete-Event Simulation.
Cristian Urs and Ben Riveira. Introduction The article we chose focuses on improving the performance of Genetic Algorithms by: Use of predictive models.
1 Local search and optimization Local search= use single current state and move to neighboring states. Advantages: –Use very little memory –Find often.
Estimation of Distribution Algorithms (EDA)
Two Approaches to Calculating Correlated Reserve Indications Across Multiple Lines of Business Gerald Kirschner Classic Solutions Casualty Loss Reserve.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Time Series Data Analysis - I Yaji Sripada. Dept. of Computing Science, University of Aberdeen2 In this lecture you learn What are Time Series? How to.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
What is Genetic Programming? Genetic programming is a model of programming which uses the ideas (and some of the terminology) of biological evolution to.
L ECTURE 3 Notes are modified from EvoNet Flying Circus slides. Found at: www2.cs.uh.edu/~ceick/ai/EC1.ppt.
Design of an Evolutionary Algorithm M&F, ch. 7 why I like this textbook and what I don’t like about it!
Introduction to Evolutionary Algorithms Session 4 Jim Smith University of the West of England, UK May/June 2012.
1 Computing in High Energy and Nuclear Physics, February 2006, Mumbai, India.
FINAL EXAM SCHEDULER (FES) Department of Computer Engineering Faculty of Engineering & Architecture Yeditepe University By Ersan ERSOY (Engineering Project)
Genetic Algorithms Siddhartha K. Shakya School of Computing. The Robert Gordon University Aberdeen, UK
1 Machine Learning 1.Where does machine learning fit in computer science? 2.What is machine learning? 3.Where can machine learning be applied? 4.Should.
UIUC CS 498: Section EA Lecture #21 Reasoning in Artificial Intelligence Professor: Eyal Amir Fall Semester 2011 (Some slides from Kevin Murphy (UBC))
Genetic Algorithms. Evolutionary Methods Methods inspired by the process of biological evolution. Main ideas: Population of solutions Assign a score or.
1 CMSC 671 Fall 2001 Class #25-26 – Tuesday, November 27 / Thursday, November 29.
Algorithms and their Applications CS2004 ( ) 13.1 Further Evolutionary Computation.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Making Time: Pseudo Time-Series for the Temporal Analysis of Cross-Section Data Emma Peeling, Allan Tucker Centre for Intelligent Data Analysis Brunel.
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #12 2/20/02 Evolutionary Algorithms.
STATISTICS AND OPTIMIZATION Dr. Asawer A. Alwasiti.
CITS7212: Computational Intelligence An Overview of Core CI Technologies Lyndon While.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Automated discovery in math Machine learning techniques (GP, ILP, etc.) have been successfully applied in science Machine learning techniques (GP, ILP,
Rotem Golan Department of Computer Science Ben-Gurion University of the Negev, Israel.
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Local Search. Systematic versus local search u Systematic search  Breadth-first, depth-first, IDDFS, A*, IDA*, etc  Keep one or more paths in memory.
1 Chapter 17 2 nd Part Making Complex Decisions --- Decision-theoretic Agent Design Xin Lu 11/04/2002.
Using Bayesian Networks to Predict Plankton Production from Satellite Data By: Rob Curtis, Richard Fenn, Damon Oberholster Supervisors: Anet Potgieter,
Artificial Intelligence By Mr. Ejaz CIIT Sahiwal Evolutionary Computation.
Modeling of Core Protection Calculator System Software February 28, 2005 Kim, Sung Ho Kim, Sung Ho.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
Estimation of Distribution Algorithm and Genetic Programming Structure Complexity Lab,Seoul National University KIM KANGIL.
Genetic Algorithm(GA)
OPERATING SYSTEMS CS 3502 Fall 2017
CSC 380: Design and Analysis of Algorithms
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Boltzmann Machine (BM) (§6.4)
Stepping on Earth: A Roadmap for Data-driven Agent-Based Modelling
CSC 380: Design and Analysis of Algorithms
Coevolutionary Automated Software Correction
Presentation transcript:

Extending Evolutionary Programming to the Learning of Dynamic Bayesian Networks Allan Tucker Xiaohui Liu Birkbeck College University of London

Diagnosis in MTS Useful to know causes for a given set of observations within a time series E.g. Oil Refinery: ‘Why a temperature has become high whilst a pressure has fallen below a certain value?’ Possible paradigm to perform Diagnosis is the Bayesian Network Evolutionary Methods to learn BNs Extend work to Dynamic Bayesian Networks

Dynamic Bayesian Networks Static BNs repeated over t time slices Contemporaneous / Non-Contemporaneous Links Used for Prediction / Diagnosis within dynamic systems

Assumptions - 1 Assume all variables take at least one time slice to impose an effect on another. The more frequently a system generates data, the more likely this will be true (e.g. every minute, second etc.) Contemporaneous Links are excluded from the DBN.

Representation N variables at time slice, t P Triples of the form (x,y,lag) Each triple represents a link from a node at a previous time slice to a node at time t. Example: { (0,0,1); (1,1,1); (2,2,1); (3,3,1); (4,4,1); (0,1,3); (2,1,2); (2,4,5); (4,3,4) }

Search Space Given the first assumption and proposed representation Search Space will be: E.g. 10 variables, MaxLag = 30 Make further assumptions to reduce this and speed up the search

Assumptions - 2 Cross-Description-Length Function (CDLF) Exhibits smoothness of Cross Correlation Function (CCF) cousin Exploit this smoothness using Swap

Assumptions - 3 Auto-Description-Length Function (ADLF) exhibits the lowest DL with time lag=1 Automatically insert these links before evaluation

Evolutionary Programming to find Links with Low Description Length Evolutionary Program (Swap) to find Dynamic Bayesian Network with Low Description Length Dynamic Bayesian Network Multivariate Time Series Explanation Algorithm (using Stochastic Simulation) User

EP to Find low DL links Using an EP approach with self adapting parameters to find a good selection of links with low DL (High Mutual Information) Representation: Each individual is a triple (x,y,lag) Fitness is DL of triple Solution is the resultant population

Swap Operator Select a triple from one parent with equal probability Mutate the current lag with a uniform distribution: Current Lag + U(-MaxLag/10, MaxLag/10) (x,y,lag) Lag: 1 MaxLag X [Lag - MaxLag/10] [Lag + MaxLag/10] (x,y,lag)

Generating Synthetic Data t-3 t-2 t-1 t t-3 t-2 t-1 t t+1 (1) (2)

Oil Refinery Data Data recorded every minute Hundreds of variables Selected 11 interrelated variables Discretized each variable into 4 states Large Time Lags (up to 120 minutes between some variables)

Parameters % 10% Population Size Generations OpRate (KGM/Swap) Slide Mutation Rate EP - DBN StructureSynthetic % 10% Oil Data 5 30 Number of Var MaxLag DataSynthetic Oil Data 1% Mutation Rate Limit (% of all links) ListGenerations EP - DL LinksSynthetic 1% 5 50 Oil Data

Results 1 - Synthetic Data

Results 2 - Oil Refinery Data

Results 3

Explanations - using stochastic simulation

Explanations Input: t - 0: Tail Gas Flow in_state 0 Reboiler Temperature in_state 3 Output: t - 7: Top Temperature in_state 0 t - 54: Feed Rate in_state 1 t - 75: Reactor Inlet Temperature in_state 0

Future Work Exploring the use of different metrics Improving accuracy (e.g. different discretization policies, continuous DBNs) Learning a library of DBNs in order to classify the current state of a system