Dynamic Optimisation So far have looked at the use of EA’s as function optimisers for stationary functions, i.e. where the fitness landscape remains static.

Slides:



Advertisements
Similar presentations
Genetic Algorithms (Evolutionary Computing) Genetic Algorithms are used to try to “evolve” the solution to a problem Generate prototype solutions called.
Advertisements

Allocation of Frames Each process needs minimum number of pages
On the Genetic Evolution of a Perfect Tic-Tac-Toe Strategy
Maynard Smith Revisited: Spatial Mobility and Limited Resources Shaping Population Dynamics and Evolutionary Stable Strategies Pedro Ribeiro de Andrade.
Tuesday, May 14 Genetic Algorithms Handouts: Lecture Notes Question: when should there be an additional review session?
The Dominance Tournament Method of Monitoring Progress in Coevolution Speaker: Lin, Wei-Kai (2009/04/30) 1.
Genetic Algorithms, Part 2. Evolving (and co-evolving) one-dimensional cellular automata to perform a computation.
Coevolutionary Learning with Genetic Algorithms. Problem for learning algorithms: How to select “training environments” appropriate to different stages.
Theory Chapter 11. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Overview (reduced w.r.t. book) Motivations and problems Holland’s.
6/2/2001 Cooperative Agent Systems: Artificial Agents Play the Ultimatum Game Steven O. Kimbrough Presented at FMEC 2001, Oslo Joint work with Fang Zhong.
Evolutionary Computational Intelligence
Optimization via Search CPSC 315 – Programming Studio Spring 2009 Project 2, Lecture 4 Adapted from slides of Yoonsuck Choe.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
Evolutionary Computation and Co-evolution Alan Blair October 2005.
D Nagesh Kumar, IIScOptimization Methods: M1L4 1 Introduction and Basic Concepts Classical and Advanced Techniques for Optimization.
Genetic Algorithms Overview Genetic Algorithms: a gentle introduction –What are GAs –How do they work/ Why? –Critical issues Use in Data Mining –GAs.
IE 607 Constrained Design: Using Constraints to Advantage in Adaptive Optimization in Manufacturing.
Evolving Multi-modal Behavior in NPCs Jacob Schrum – Risto Miikkulainen –
Parallel Genetic Algorithms with Distributed-Environment Multiple Population Scheme M.Miki T.Hiroyasu K.Hatanaka Doshisha University,Kyoto,Japan.
Prepared by Barış GÖKÇE 1.  Search Methods  Evolutionary Algorithms (EA)  Characteristics of EAs  Genetic Programming (GP)  Evolutionary Programming.
A Genetic Algorithms Approach to Feature Subset Selection Problem by Hasan Doğu TAŞKIRAN CS 550 – Machine Learning Workshop Department of Computer Engineering.
Evolutionary Algorithms BIOL/CMSC 361: Emergence Lecture 4/03/08.
Learning in Multiagent systems
林偉楷 Taiwan Evolutionary Intelligence Laboratory.
Multimodal Optimization (Niching) A/Prof. Xiaodong Li School of Computer Science and IT, RMIT University Melbourne, Australia
SOFT COMPUTING (Optimization Techniques using GA) Dr. N.Uma Maheswari Professor/CSE PSNA CET.
Genetic algorithms Prof Kang Li
More on coevolution and learning Jing Xiao April, 2008.
Coevolutionary Models A/Prof. Xiaodong Li School of Computer Science and IT, RMIT University Melbourne, Australia April.
Zorica Stanimirović Faculty of Mathematics, University of Belgrade
Genetic Algorithms Michael J. Watts
Presenter: Chih-Yuan Chou GA-BASED ALGORITHMS FOR FINDING EQUILIBRIUM 1.
Optimization Problems - Optimization: In the real world, there are many problems (e.g. Traveling Salesman Problem, Playing Chess ) that have numerous possible.
Coevolution Chapter 6, Essentials of Metaheuristics, 2013 Spring, 2014 Metaheuristics Byung-Hyun Ha R2R3.
GENETIC ALGORITHMS.  Genetic algorithms are a form of local search that use methods based on evolution to make small changes to a popula- tion of chromosomes.
Heuristic Optimization Methods Greedy algorithms, Approximation algorithms, and GRASP.
15/06/2003NORPIE 2004, Trondheim1 Genetic Optimization of Electric Machines, a State of the Art Study S. E. Skaar, R. Nilssen.
2005MEE Software Engineering Lecture 11 – Optimisation Techniques.
Evolution strategies Luis Martí DEE/PUC-Rio. ES quick overview Developed: Germany in the 1970’s Early names: I. Rechenberg, H.-P. Schwefel Typically applied.
Algorithms and their Applications CS2004 ( ) 13.1 Further Evolutionary Computation.
Learning by Simulating Evolution Artificial Intelligence CSMC February 21, 2002.
Artificial Intelligence for Games Online and local search
Niching Genetic Algorithms Motivation The Idea Ecological Meaning Niching Techniques.
Evolving cooperation in one-time interactions with strangers Tags produce cooperation in the single round prisoner’s dilemma and it’s.
Chapter 9 Genetic Algorithms.  Based upon biological evolution  Generate successor hypothesis based upon repeated mutations  Acts as a randomized parallel.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Evolving Reactive NPCs for the Real-Time Simulation Game.
Genetic Algorithms Genetic algorithms provide an approach to learning that is based loosely on simulated evolution. Hypotheses are often described by bit.
Evolution strategies Chapter 4. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolution Strategies ES quick overview Developed: Germany.
1. Genetic Algorithms: An Overview  Objectives - Studying basic principle of GA - Understanding applications in prisoner’s dilemma & sorting network.
The Evolution of Specialisation in Groups – Tags (again!) David Hales Centre for Policy Modelling, Manchester Metropolitan University, UK.
1 Danny Hillis and Co-evolution Between Hosts and Parasites I 590 4/11/2005 Pu-Wen(Bruce) Chang.
Iterated Prisoner’s Dilemma Game in Evolutionary Computation Seung-Ryong Yang.
Competitive Coevolution (Predator-Prey Coevolution)
Genetic Algorithms Chapter Description of Presentations
Evolutionary Computing Chapter 11. / 7 Chapter 11: Non-stationary and Noisy Function Optimisation What is a non-stationary problem? Effect of uncertainty.
Chapter 15: Co-Evolutionary Systems
Ch 20. Parameter Control Ch 21. Self-adaptation Evolutionary Computation vol. 2: Advanced Algorithms and Operators Summarized and presented by Jung-Woo.
Evolving Specialisation, Altruism & Group-Level Optimisation Using Tags – The emergence of a group identity? David Hales Centre for Policy Modelling, Manchester.
Evolving Specialisation, Altruism & Group-Level Optimisation Using Tags David Hales Centre for Policy Modelling, Manchester Metropolitan University, UK.
Multi-cellular paradigm The molecular level can support self- replication (and self- repair). But we also need cells that can be designed to fit the specific.
CEng 713, Evolutionary Computation, Lecture Notes parallel Evolutionary Computation.
Game Theory and Cooperation
C.-S. Shieh, EC, KUAS, Taiwan
Evolution strategies Can programs learn?
CSC 380: Design and Analysis of Algorithms
Breeding Bunnies Lab Observe the graph and discuss with your lab mate.
Advanced Artificial Intelligence Evolutionary Search Algorithm
Genetic Algorithms Chapter 3.
CSC 380: Design and Analysis of Algorithms
Coevolutionary Automated Software Correction
Presentation transcript:

Dynamic Optimisation So far have looked at the use of EA’s as function optimisers for stationary functions, i.e. where the fitness landscape remains static w.r.t. time. Many interesting applications are non- stationary: given a potential solution x and a fitness function f, then f(x, t1) and f(x,t2) may differ.

Types of non-stationary function Cobb (1990) defines two ways of classifying non-stationary functions –Switching vs. Continuous based on time-scale of change w.r.t. rate of evaluation –Markovian vs. State Dependant f(x,t) comes from asymptotic distribution or is strictly a function of time.

Strategies Used Cobb(1990) identifies two different classes of strategies for NS environments 1: Expand the memory of the EA. e.g. Goldberg & Smith (87), Smith (88), Dasgupta & MacGregor(1992). good solution if there are a fixed number of states in MSE 2: (Adaptively) expand variation in the population

Experiments with Std GA Cobb (1990) observed the behaviour of a standard GA on a parabolic function with a the optima moving sinusoidally in space. Offline performance decreases as rate of change increases for all p_m As rate of change increases the value of p_m which gives optimal offline performance increases As problem difficulty increased, rate of change that GA could track decreased

Triggered Hyper Mutation(HM) Cobb proposed adaptive mutation mechanism with two mutation rates 0.5/0.001 Higher rate triggered by drop in fitness of best in generation Tracked optima far better than SGA, especially when the time-optima were spatially close

Pitfalls with HM Grefenstette (1992) noted that under certain conditions HM might never get triggered. Function composed of 14 sinusoidal hills initial optima 5.0, subsequent optima with height 10.0, appears after 20 generations, and moves every 20 generations SGA goes to 5.0, rises to 10.0 when optima close but does not track HM, goes to 5, only tracks 10.0 after optima moves close to original peak, then triggering works

Random Immigrants (RI) RI strategy proposed, replace proportion of population with randomly created members at every generation –30% replacement gives best off-line tracking –too high and it can’t converge between changes –however off-line performance decreases with proportion replaced

Comparison of techniques Cobb & Grefenstette (1993) compared HM vs. RI vs SGA (with high mutation rate) Noted difference in nature of mutation –SGA uniform in pop and time –HM uniform in pop, not in time –RI uniform in time, not in pop. NB this (later) version of HM triggered by decrease in running mean over 5 generations of best in pop.

Comparison Experiments 2 landscapes, A as per picture, B composed of various shaped hills. 32 bits each 3 types of change: –linear motion in A, moving 1 step along an axis every 2/5 generations –randomly shifting the optima in A every 20 generations –swap between A and B every 2/20 generations

Comparison Conclusions ran lots of experiments to find good settings for each algorithm with each problem SGA: – p_m 0.1 was reasonably good at the translation tasks, but gave v. poor online performance. –couldn’t track moving optimum or oscillation – p_m needs to be matched to degree of change

Comparison Conclusions cont.. HM –High variance in performance noted –HM rate needed tuning a lot to problem –Much better at tracking sudden changes than SGA, and gave better online performance than SGA or RI`when rate of change was slow enough to allow lower rate of mutation

Comparison Conclusions cont. RI –not very good at tracking linear movement –best at the oscillating task –they hypothesise ( but don’t show) that this is because it allows the preservation of niches –poor performance on stationary problems and slow movements

Vavak 1993-> Concentrated on on-line performance, for control of industrial systems Compared generational vs. steady state GA’s on a 1 dimensional moving problem –SSGA better from random or converged population. –entropy reacts faster and reaches lower value for SSGA

Design choices for SSEAs Smith & Vavak showed choice of replacement strategy was critical Best algorithms –Use elitism, but –Systematically re-evaluate all points in population –Overall a “conservative” delete-oldest policy was found to work best,

Self-Adaptive Approaches Back et al showed that it is not necessary to use a trigger if you have a self-adaptive system. ES was able to track a moving optima reliably

Co-evolution Sometimes the fitness of a solution may not be externally defined, e.g.: –Individuals compete against others from same population E.g. evolving game playing strategies –Individuals must be evaluated in context of solutions from other populations representing Other parts of decomposed problem Competing species Obviously in biology the fitness in our “adaptive landscape” metaphor is a function of the other competing species rather than being fixed.

Some terminology Mutualism or Symbiosis: –The beneficial co-adaptation of different species, may become co- located (e.g. gut flora) –In EC known as Co-operative Coevolution: e.g. lots of work by Bull, Potter and DeJong on function decomposition Predation or Parasitism: –Antagonistic co-adaption e.g. fox/rabbit –Also known as competitive coevolution: e.g. lots of work on test- soluton problems, game-playing etc. Note that both types of coevolution can occur in single population or multi-population forms

How to evaluate fitness? Single population: –Test each solution against every other (or a randomly selected sample if pop. is large) Multiple populations –Need to decide a pairing strategy e.g.: Evolve populations in turn, use best of other pops. Pair with randomly chosen members of other pops –Lots of work by Bull, Husbands verdict still out –Neat solution by Husbands uses cellular model with one of each pop on each grid point –Paredis uses “Life-Time-Fitness-Evalaution” – mean of last twenty encounters with another species

Some examples Hillis 1993: –Co-evolved sorting networks with “difficult” test sequences - found network of smallest size known at the time Axelrod: –Organised competitions for Iterated Prisoners Dilemma Showed he could evolve winning strategies Paredis: –Coevolved solutions and their representations to remove operator bias –Also used coevolution for constraint satisfaction problems

This sounds great, why don’t we all do it …. Desirable behaviour in competitive evolution is known as “arms races” –Red queen effect: run faster and faster just to keep up But problems arise: –Cycling –Mediocre stable states –disengagement

What does the future hold Several groups actively researching solutions to these problems LOTS of potential applications: –Agent based systems for e-commerce, trading, game-playing etc. –Investigating difficult problems in abstract fields eg maths –Evolving team behaviour –Optimisation with (automatic? ) problem decomposition