Empirical Algorithmics Reading Group Oct 11, 2007 Tuning Search Algorithms for Real-World Applications: A Regression Tree Based Approach by Thomas Bartz-Beielstein.

Slides:



Advertisements
Similar presentations
Detecting Statistical Interactions with Additive Groves of Trees
Advertisements

MINERVA: an automated resource provisioning tool for large-scale storage systems G. Alvarez, E. Borowsky, S. Go, T. Romer, R. Becker-Szendy, R. Golding,
Chapter 7 Classification and Regression Trees
Ali Husseinzadeh Kashan Spring 2010
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Experimental Design, Response Surface Analysis, and Optimization
Biologically Inspired AI (mostly GAs). Some Examples of Biologically Inspired Computation Neural networks Evolutionary computation (e.g., genetic algorithms)
CPSC 322, Lecture 16Slide 1 Stochastic Local Search Variants Computer Science cpsc322, Lecture 16 (Textbook Chpt 4.8) February, 9, 2009.
Age-Based Population Dynamics in Evolutionary Algorithms Lisa Guntly.
1 An Asymptotically Optimal Algorithm for the Max k-Armed Bandit Problem Matthew Streeter & Stephen Smith Carnegie Mellon University NESCAI, April
Optimization via Search CPSC 315 – Programming Studio Spring 2009 Project 2, Lecture 4 Adapted from slides of Yoonsuck Choe.
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
02 -1 Lecture 02 Heuristic Search Topics –Basics –Hill Climbing –Simulated Annealing –Best First Search –Genetic Algorithms –Game Search.
Fast Evolutionary Optimisation Temi avanzati di Intelligenza Artificiale - Lecture 6 Prof. Vincenzo Cutello Department of Mathematics and Computer Science.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Using Simulated Annealing and Evolution Strategy scheduling capital products with complex product structure By: Dongping SONG Supervisors: Dr. Chris Hicks.
Optimization via Search CPSC 315 – Programming Studio Spring 2008 Project 2, Lecture 4 Adapted from slides of Yoonsuck Choe.
Ensemble Learning (2), Tree and Forest
Brandon Andrews.  What are genetic algorithms?  3 steps  Applications to Bioinformatics.
Evolutionary algorithms
Ranga Rodrigo April 6, 2014 Most of the sides are from the Matlab tutorial. 1.
Genetic Algorithm.
Genetic Algorithms and Ant Colony Optimisation
Lecture Notes 4 Pruning Zhangxi Lin ISQS
Efficient Model Selection for Support Vector Machines
FDA- A scalable evolutionary algorithm for the optimization of ADFs By Hossein Momeni.
Zorica Stanimirović Faculty of Mathematics, University of Belgrade
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
What is Genetic Programming? Genetic programming is a model of programming which uses the ideas (and some of the terminology) of biological evolution to.
Optimization Problems - Optimization: In the real world, there are many problems (e.g. Traveling Salesman Problem, Playing Chess ) that have numerous possible.
Design of an Evolutionary Algorithm M&F, ch. 7 why I like this textbook and what I don’t like about it!
Introduction to Evolutionary Algorithms Session 4 Jim Smith University of the West of England, UK May/June 2012.
Neural and Evolutionary Computing - Lecture 6
The Generational Control Model This is the control model that is traditionally used by GP systems. There are a distinct number of generations performed.
Genetic Algorithms Genetic Algorithms – What are they? And how they are inspired from evolution. Operators and Definitions in Genetic Algorithms paradigm.
Derivative Free Optimization G.Anuradha. Contents Genetic Algorithm Simulated Annealing Random search method Downhill simplex method.
Simulation in Wind Turbine Vibrations: A Data Driven Analysis Graduate Students: Zijun Zhang PI: Andrew Kusiak Intelligent Systems Laboratory The University.
1 Genetic Algorithms and Ant Colony Optimisation.
Kanpur Genetic Algorithms Laboratory IIT Kanpur 25, July 2006 (11:00 AM) Multi-Objective Dynamic Optimization using Evolutionary Algorithms by Udaya Bhaskara.
Evolutionary Programming
Experimental Algorithmics Reading Group, UBC, CS Presented paper: Fine-tuning of Algorithms Using Fractional Experimental Designs and Local Search by Belarmino.
Genetic Algorithms What is a GA Terms and definitions Basic algorithm.
Probabilistic Algorithms Evolutionary Algorithms Simulated Annealing.
Optimization Problems
Alice E. Smith and Mehmet Gulsen Department of Industrial Engineering
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
Classification and Regression Trees
CPSC 322, Lecture 16Slide 1 Stochastic Local Search Variants Computer Science cpsc322, Lecture 16 (Textbook Chpt 4.8) Oct, 11, 2013.
An Introduction to Genetic Algorithms Lecture 2 November, 2010 Ivan Garibay
Selection and Recombination Temi avanzati di Intelligenza Artificiale - Lecture 4 Prof. Vincenzo Cutello Department of Mathematics and Computer Science.
Metaheuristics for the New Millennium Bruce L. Golden RH Smith School of Business University of Maryland by Presented at the University of Iowa, March.
Eco 6380 Predictive Analytics For Economists Spring 2016 Professor Tom Fomby Department of Economics SMU.
Evolutionary Programming A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Chapter 5.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Genetic Algorithms And other approaches for similar applications Optimization Techniques.
Genetic Algorithm. Outline Motivation Genetic algorithms An illustrative example Hypothesis space search.
Hirophysics.com The Genetic Algorithm vs. Simulated Annealing Charles Barnes PHY 327.
Optimization via Search
Evolutionary Programming
Heuristic Optimization Methods
Evolution strategies Can programs learn?
A Distributed Genetic Algorithm for Learning Evaluation Parameters in a Strategy Game Gary Matthias.
A Study of Genetic Algorithms for Parameter Optimization
Subject Name: Operation Research Subject Code: 10CS661 Prepared By:Mrs
Evolutionary Computation,
More on Search: A* and Optimization
Evolutionary Programming
Boltzmann Machine (BM) (§6.4)
Stochastic Methods.
Presentation transcript:

Empirical Algorithmics Reading Group Oct 11, 2007 Tuning Search Algorithms for Real-World Applications: A Regression Tree Based Approach by Thomas Bartz-Beielstein & Sandor Markon Presenter: Frank Hutter

Motivation How to find a set of working parameters for direct search algorithms when the number of allowed epxeriments is low –i.e. find good parameters with few evaluations Taking a users perspective: –Adopt standard params from the literature –But NFL theorem: cant do good everywhere –Tune for instance class / for optimization instances even on a single instance

Considered approaches Regression analysis ANOVA DACE CART

Elevator Group Control Multi-objective problem –Overall service quality –Traffic throughput –Energy consumption –Transport capacity –Many more … Here: only one objective –Minimize time customers have to wait until they can enter the elevator car

Optimization via Simulation Goal: Optimize expected performance E[y(x 1,…, x n )] (x 1,…, x n controllable) Black box function y

Direct search algorithms Do not construct a model of the fitness function Interesting aside: same nomenclature as I use, but independent Here –Evolution strategy (special class of evolutionary algorithm) –Simulated annealing

Evolution strategies (ES) Start out with parental population at t=0 For each new generation: –Create offsprings Select parent family of size \rho at random Apply recombination to object variables (?) and strategy parameters (?) –Mutation of each offspring –Selection

Many parameters in ES Number of parent individuals Number of offspring individuals Initial mean step sizes ( i ) –Can choose problem-specific, different i for each dimension (not done here) Number of standard deviations (??) Mutation strength (global/individual, extended log-normal rule ??) Mixing number (size of each parent family) Recombination operator –For object variables –For strategy variables Selection mechanims, maximum life span Plus-strategies ( + ) and comma-strategies (, ) Can be generalized by (maximum age of individual)

Simulated Annealing Proposal: Gaussian Markov kernel with scale proportional to the temperature Decrease temperature on a logarithmic cooling schedule Two parameters –Starting temperature –Number of function evaluations at each temperature

Experimental Analysis of Search Heuristics Which parameters have the greatest effect? –Screening Which parameter setting might lead to an improved performance –Modelling –Optimization

Design of experiments (DOE) Choose two factors for each parameter –Both qualitative and quantitative 2 k-p fractional factorial design –2: number of levels for each factor –K parameters –Only 2 k-p experiments –Can be generated from a full factorial design on k-p params –Resolution = (k-p) +1 (is this always the case?) Resolution 2: not useful – main effects are confounded with each other Resolution 3: often used, main effects are unconfounded with each other Resolution 4: all main effects are unconfounded with all 2-factor interactions Resolution 5: all 2-factor interactions are unconfounded with each other Here: 2 III 9-5 fractional factorial design

Regression analysis Using stepAIC function built into R –Akaikes information criterion to penalize many parameters in the model –Line search to improve algorithms performance (?)

Tree based regression Used for screening Based on the fractional factorial design Forward growing –Splitting criterion: minimal variance within the two children –Backward pruning: snipping away branches to maximize penalized cost Using rpart implementation from R –10-fold cross validation –1-SE rule: mean + 1stddev as pessimistic estimate –Threshold complexity parameter: visually chosen based on 1-SE rule

Experimental results 5000 fitness evaluations as termination criterion Initialization already finds good parameters ! only small improvements possible Actual results not too important, but methods! Questions –Is strategy useful? –Improve parameters –Which analysis strategy works?

Two splits (, ): Regression analysis:only first split significant Tuned algorithm found solution with quality y= –Which parameter settings? –What does mean? –How about multiple runs? strategy useful? regression tree analysis

New Gupta vs. classical + selection Tune old and new variants Report new results and runtime for tuning –Just that they do not report the runtime for tuning

Comparison of approaches on Simulated Annealing Only two (continuous) parameters Classical regression fails –No significant effects Regression tree –Best around 10,10 –Based on a full-factorial design with 2 levels each this is pretty shaky

Comparison of approaches E.g. regression trees for screening, then DACE if only a few continuous parameters remain (why the restriction to few?)