Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.

Slides:



Advertisements
Similar presentations
Using Parallel Genetic Algorithm in a Predictive Job Scheduling
Advertisements

Genetic Algorithms Contents 1. Basic Concepts 2. Algorithm
Graduate Center/City University of New York University of Helsinki FINDING OPTIMAL BAYESIAN NETWORK STRUCTURES WITH CONSTRAINTS LEARNED FROM DATA Xiannian.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Linkage Problem, Distribution Estimation, and Bayesian Networks Evolutionary Computation 8(3) Martin Pelikan, David E. Goldberg, and Erick Cantu-Paz.
An Approach to Evaluate Data Trustworthiness Based on Data Provenance Department of Computer Science Purdue University.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
Nir Friedman, Iftach Nachman, and Dana Peer Announcer: Kyu-Baek Hwang
Spie98-1 Evolutionary Algorithms, Simulated Annealing, and Tabu Search: A Comparative Study H. Youssef, S. M. Sait, H. Adiche
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
1 Lecture 8: Genetic Algorithms Contents : Miming nature The steps of the algorithm –Coosing parents –Reproduction –Mutation Deeper in GA –Stochastic Universal.
Estimation of Distribution Algorithms Ata Kaban School of Computer Science The University of Birmingham.
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
Feature Selection for Regression Problems
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
Introduction to Genetic Algorithms Yonatan Shichel.
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
. PGM: Tirgul 10 Parameter Learning and Priors. 2 Why learning? Knowledge acquisition bottleneck u Knowledge acquisition is an expensive process u Often.
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
Artificial Intelligence Genetic Algorithms and Applications of Genetic Algorithms in Compilers Prasad A. Kulkarni.
Genetic Algorithms Nehaya Tayseer 1.Introduction What is a Genetic algorithm? A search technique used in computer science to find approximate solutions.
Genetic Algorithm What is a genetic algorithm? “Genetic Algorithms are defined as global optimization procedures that use an analogy of genetic evolution.
Cristina Manfredotti D.I.S.Co. Università di Milano - Bicocca An Introduction to the Use of Bayesian Network to Analyze Gene Expression Data Cristina Manfredotti.
Genetic Algorithm.
FDA- A scalable evolutionary algorithm for the optimization of ADFs By Hossein Momeni.
1 Local search and optimization Local search= use single current state and move to neighboring states. Advantages: –Use very little memory –Find often.
Intro. ANN & Fuzzy Systems Lecture 36 GENETIC ALGORITHM (1)
Estimation of Distribution Algorithms (EDA)
Zorica Stanimirović Faculty of Mathematics, University of Belgrade
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
The Generational Control Model This is the control model that is traditionally used by GP systems. There are a distinct number of generations performed.
Major objective of this course is: Design and analysis of modern algorithms Different variants Accuracy Efficiency Comparing efficiencies Motivation thinking.
Genetic Algorithms Introduction Advanced. Simple Genetic Algorithms: Introduction What is it? In a Nutshell References The Pseudo Code Illustrations Applications.
GENETIC ALGORITHM A biologically inspired model of intelligence and the principles of biological evolution are applied to find solutions to difficult problems.
Derivative Free Optimization G.Anuradha. Contents Genetic Algorithm Simulated Annealing Random search method Downhill simplex method.
GENETIC ALGORITHMS.  Genetic algorithms are a form of local search that use methods based on evolution to make small changes to a popula- tion of chromosomes.
Siddhartha Shakya1 Estimation Of Distribution Algorithm based on Markov Random Fields Siddhartha Shakya School Of Computing The Robert Gordon.
A genetic approach to the automatic clustering problem Author : Lin Yu Tseng Shiueng Bien Yang Graduate : Chien-Ming Hsiao.
2005MEE Software Engineering Lecture 11 – Optimisation Techniques.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Chapter 9 Genetic Algorithms.  Based upon biological evolution  Generate successor hypothesis based upon repeated mutations  Acts as a randomized parallel.
Edge Assembly Crossover
K2 Algorithm Presentation KDD Lab, CIS Department, KSU
Genetic Algorithms Genetic algorithms provide an approach to learning that is based loosely on simulated evolution. Hypotheses are often described by bit.
EE749 I ntroduction to Artificial I ntelligence Genetic Algorithms The Simple GA.
Dependency Networks for Collaborative Filtering and Data Visualization UAI-2000 발표 : 황규백.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Foundation of Computing Systems
Waqas Haider Bangyal 1. Evolutionary computing algorithms are very common and used by many researchers in their research to solve the optimization problems.
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
GENETIC ALGORITHM Basic Algorithm begin set time t = 0;
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
Bayesian Optimization Algorithm, Decision Graphs, and Occam’s Razor Martin Pelikan, David E. Goldberg, and Kumara Sastry IlliGAL Report No May.
Sporadic model building for efficiency enhancement of the hierarchical BOA Genetic Programming and Evolvable Machines (2008) 9: Martin Pelikan, Kumara.
Genetic Algorithm Dr. Md. Al-amin Bhuiyan Professor, Dept. of CSE Jahangirnagar University.
Metaheuristics for the New Millennium Bruce L. Golden RH Smith School of Business University of Maryland by Presented at the University of Iowa, March.
Artificial Intelligence By Mr. Ejaz CIIT Sahiwal Evolutionary Computation.
Computacion Inteligente Least-Square Methods for System Identification.
Breeding Swarms: A GA/PSO Hybrid 簡明昌 Author and Source Author: Matthew Settles and Terence Soule Source: GECCO 2005, p How to get: (\\nclab.csie.nctu.edu.tw\Repository\Journals-
Genetic Algorithm. Outline Motivation Genetic algorithms An illustrative example Hypothesis space search.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Intelligent Exploration for Genetic Algorithms Using Self-Organizing.
Genetic Algorithms.
Bayesian Models in Machine Learning
Artificial Intelligence Chapter 20 Learning and Acting with Bayes Nets
○ Hisashi Shimosaka (Doshisha University)
EE368 Soft Computing Genetic Algorithms.
Boltzmann Machine (BM) (§6.4)
Chapter 20. Learning and Acting with Bayes Nets
Presentation transcript:

Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm

Abstract In this paper, an algorithm based on the concepts of genetic algorithms that uses an estimation of a probability distribution of promising solutions in order to generate new candidate solutions is proposed. To estimate the distribution, techniques for modeling multivariate data by Bayesian networks are used.

1. INTRODUCTION The purpose of this paper is to introduce an algorithm that uses techniques for estimating the joint distribution of multinomial data by Bayesian networks in order to generate new solutions. The combination of information from the set of good solutions and the prior information about a problem is used to estimate the distribution.

2.BAYESIAN OPTIMIZATION ALGORITHM The Bayesian Optimization Algorithm (BOA) (1) set t 0 randomly generate initial population P(0) the first population of strings is generated at random. (2) select a set of promising strings S(t) from P(t) From the current population, the better strings are selected.

(3) construct the network B using a chosen metric and constraints Any metric as a measure for quality of networks and any search algorithm can be used to search over the networks in order to maximize the value of the used metric. (4) generate a set of new strings O(t) according to the joint distribution encoded by B New strings are generated using the joint distribution encoded by the constructed network.

(5) create a new population P(t+1) by replacing some strings from P(t) with O(t) set t t + 1 The new strings are added into the old population, replacing some of the old ones. (6) if the termination criteria are not met, go to (2)

3. BAYESIAN NETWORKS A Bayesian network encodes the relationships between the variables contained in the modeled data. It represents the structure of a problem. Bayesian networks can be used to describe the data as well as to generate new instances of the variables with similar properties as those of given data.

Mathematically, an acyclic Bayesian network with directed edges encodes a joint probability distribution. This can be written as

3.1 CONSTRUCTING THE NETWORK There are two basic components of the algorithms for learning the network structure scoring metric search procedure The space of networks can be reduced by constraint operators.

3.1.1 Bayesian Dirichlet metric This metric combines the prior knowledge about the problem and the statistical data from a given data set. The BD metric for a network B given a data set D of size N, and the background information, denoted by p(D,B| ), is defined as

where p(B| ) is the prior probability of the network B By numbers m’(xi, π Xi ) and p(B| ), prior information about the problem is incorporated into the metric. In our implementation, we set p(B| ) = 1 for all networks

3.1.2 Searching for a Good Network the basic principles of algorithms that can be used for searching over the networks in order to maximize the value of a scoring metric are described. a) k = 0 This case is trivial. An empty network is the best one b) k = 1 For k = 1, there exists a polynomial algorithm for the network construction c) k > 1 For k > 1 the problem gets much more complicated.

Various algorithms can be used in order to find a good network (A simple greedy algorithm, local hill-climbing, or simulated annealing can be used.) In our implementation, we have used a simple greedy algorithm with only edge additions allowed. The algorithm starts with an empty network.

3.2 GENERATING NEW INSTANCES New instances are generated using the joint distribution encoded by the network (see Equation 1). First, the conditional probabilities of each possible instance of each variable given all possible instances of its parents in a given data set are computed. Each iteration, the nodes whose parents are already fixed are generated using the corresponding conditional probabilities.

4.DECOMPOSABLE FUNCTIONS The problems defined by this class of functions can be decomposed into smaller subproblems. There is no straightforward way to relate general decomposable problems and what are the necessary interactions to be taken into account BOA has been confirmed by a number of experiments with various test functions,BOA is better than the others.

5.EXPERIMENTS The experiments were designed in order to show the behavior of the proposed algorithm only on non-overlapping decomposable problems with uniformly scaled deceptive building blocks.

5.1 FUNCTIONS OF UNITATION A function of unitation is a function whose value depends only on the number of ones in a binary input string. Let us have a function of unitation f k defined for strings of length k. Then, the function additively composed of l functions f k is defined as where X is the set of n variables and Si for I {0,..., l − 1} are subsets of k variables from X.

A deceptive function of order 3, denoted by 3-deceptive, is defined as where u is the number of one’s in an input string. A trap function of order 5, denoted by trap-5, is defined as A bipolar deceptive function of order 6, denoted by 6-bipolar,

5.2 RESULTS OF THE EXPERIMENTS For all problems, the average number of fitness evaluations until convergence in 30 independent runs is shown.