Xin-She Yang, Nature-Inspired Optimization Algorithms, Elsevier, 2014.

Slides:



Advertisements
Similar presentations
Particle Swarm Optimization (PSO)
Advertisements

Particle Swarm Optimization
Neural and Evolutionary Computing - Lecture 4 1 Random Search Algorithms. Simulated Annealing Motivation Simple Random Search Algorithms Simulated Annealing.
Application of Heuristics and Meta-Heuristics Scheduling: Job Shop Scheduling Parallel Dedicated Machine (PDS1) Single machine minimize total tardiness.
Firefly Algorithm by Mr Zamani & Hosseini.
PARTICLE SWARM OPTIMISATION (PSO) Perry Brown Alexander Mathews Image:
Firefly Algorithm By Rasool Tavakoli.
Effective gradient-free methods for inverse problems Jyri Leskinen FiDiPro DESIGN project.
Optimization Methods One-Dimensional Unconstrained Optimization
Optimization Methods One-Dimensional Unconstrained Optimization
Metaheuristics The idea: search the solution space directly. No math models, only a set of algorithmic steps, iterative method. Find a feasible solution.
Elements of the Heuristic Approach
Genetic Algorithms Overview Genetic Algorithms: a gentle introduction –What are GAs –How do they work/ Why? –Critical issues Use in Data Mining –GAs.
Optimization of thermal processes2007/2008 Optimization of thermal processes Maciej Marek Czestochowa University of Technology Institute of Thermal Machinery.
July 11, 2001Daniel Whiteson Support Vector Machines: Get more Higgs out of your data Daniel Whiteson UC Berkeley.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
Multimodal Interaction Dr. Mike Spann
CSC321: Neural Networks Lecture 12: Clustering Geoffrey Hinton.
Swarm Intelligence 虞台文.
Face Recognition Using Evolutionary Approaches
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
Optimization Problems - Optimization: In the real world, there are many problems (e.g. Traveling Salesman Problem, Playing Chess ) that have numerous possible.
(Particle Swarm Optimisation)
Neural and Evolutionary Computing - Lecture 9 1 Evolutionary Neural Networks Design  Motivation  Evolutionary training  Evolutionary design of the architecture.
Particle Swarm Optimization † Spencer Vogel † This presentation contains cheesy graphics and animations and they will be awesome.
An Improved Quantum-behaved Particle Swarm Optimization Algorithm Based on Culture V i   v i 1, v i 2,.. v iD  Gao X. Z 2, Wu Ying 1, Huang Xianlin.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Neural Networks - Berrin Yanıkoğlu1 MLP & Backpropagation Issues.
Paper Review for ENGG6140 Memetic Algorithms
Linear Programming for Solving the DSS Problems
Optimal Over current relay co-ordination using Firefly Algorithm in Electrical Network: A nature inspired approach A presentation on Prepared by: Snehal.
Genetic Algorithms.
Evolutionary Algorithms Jim Whitehead
Particle Swarm Optimization (2)
Deep Feedforward Networks
Department of Computer Science
Scientific Research Group in Egypt (SRGE)
Discrete ABC Based on Similarity for GCP
Goal We present a hybrid optimization approach for solving global optimization problems, in particular automated parameter estimation models. The hybrid.
4. Numerical Integration
PSO -Introduction Proposed by James Kennedy & Russell Eberhart in 1995
Whale Optimization Algorithm
Differential Evolution
Meta-heuristics Introduction - Fabien Tricoire
C.-S. Shieh, EC, KUAS, Taiwan
Probability-based Evolutionary Algorithms
OPTIMIZATION OF PLANAR TRUSS STRUCTURE USING FIREFLY ALGORITHM
Clustering (3) Center-based algorithms Fuzzy k-means
Swarm simulation using anti-Newtonian forces
Going Backwards In The Procedure and Recapitulation of System Identification By Ali Pekcan 65570B.
Xin-She Yang, Nature-Inspired Optimization Algorithms, Elsevier, 2014
metaheuristic methods and their applications
Scientific Research Group in Egypt (SRGE)
SMEM Algorithm for Mixture Models
Metaheuristic methods and their applications. Optimization Problems Strategies for Solving NP-hard Optimization Problems What is a Metaheuristic Method?
Multi-Objective Optimization
Introduction to Simulated Annealing
traveling salesman problem
Optimization and Some Traditional Methods
Boltzmann Machine (BM) (§6.4)
The loss function, the normal equation,
Xin-She Yang, Nature-Inspired Optimization Algorithms, Elsevier, 2014
EE 458 Introduction to Optimization
Mathematical Foundations of BME Reza Shadmehr
Neural networks (1) Traditional multi-layer perceptrons
Statistical Methods for Data Analysis Random number generators
6 Differential Evolution
A Tutorial (Complete) Yiming
SWARM INTELLIGENCE Swarms
Population Methods.
Presentation transcript:

Xin-She Yang, Nature-Inspired Optimization Algorithms, Elsevier, 2014. 8 Firefly Algorithms Xin-She Yang, Nature-Inspired Optimization Algorithms, Elsevier, 2014.

8.1 The Firefly Algorithm FA was first developed by Xin-She Yang in late 2007 and published in 2008. FA was based on the flashing patterns and behavior of fireflies.

8.1.1 Firefly Behavior Two fundamental functions of such flashes Attract mating partners Attract potential prey Two combined factors make most fireflies visible to a limit distance The light intensity I decreases as the distance r increases in terms of I ∝ 1/r2. The air absorbs light, which becomes weaker and weaker as the distance increases.

8.1.2 Standard Firefly Algorithm Three idealized rules All fireflies are unisex One firefly will be attracted to other fireflies regardless of their sex. Attractiveness is proportional to a firefly’s brightness. For any two flashing fireflies, the less brighter one will move toward the brighter one. The attractiveness is proportional to the brightness, both of which decrease as their distance increases. If there is no brighter one than a particular firefly, it will move randomly.

8.1.2 Standard Firefly Algorithm The brightness of a firefly is affected or determined by the landscape of the objective function.

8.1.3 Variations of Light Intensity and Attractiveness There are two important issues Variation of light intensity The brightness I of a firefly at a particular location x can be chosen as I(x) ∝ f(x) Formulation of the attractiveness The attractiveness β is relative; it should be seen in the eyes of the beholder or judged by the other fireflies. Light attenuation Light absorption

The light intensity I (r ) varies according to the inverse square law where Is is the intensity at the source.

For a given medium with a fixed light absorption coefficient γ , the light intensity I varies with the distance r . That is, where I0 is the original light intensity at zero distance r = 0

The combined effect of both the inverse-square law and absorption can be approximated as the following Gaussian form:

Because a firefly’s attractiveness is proportional to the light intensity seen by adjacent fireflies, we can now define the attractiveness β of a firefly by where β0 is the attractiveness at r = 0.

The distance between any two fireflies i and j at xi and xj , respectively, is the Cartesian distance where xi,k is the kth component of the spatial coordinate xi of i th firefly.

The movement of a firefly i attracted to another, more attractive (brighter) firefly j is determined by where the second term is due to the attraction. The third term is randomization, with α being the randomization parameter, and ϵi is a vector of random numbers drawn from a Gaussian distribution or uniform distribution.

8.1.4 Controlling Randomization A further improvement on the convergence of the algorithm is to vary the randomization parameter α so that it decreases gradually as the optima are approaching or where θ ∈ (0, 1]

Simulations indicated that the efficiency may improve if we add an extra term λ ϵi (g∗−xi ) to the updating formula, where g∗ is the current global optimum.

8.2 Algorithm Analysis

8.2.1 Scalings and Limiting Cases In fact, any measure that can effectively characterize the quantities of interest in the optimization problem can be used as the “distance” r . The initial locations of these n fireflies distribute relatively uniformly over the entire search space. As the iterations proceed, the fireflies would converge into all the local optima (including the global ones).

For γ → 0, the attractiveness is constant β = β0,  FA  Particle Swarm Optimization. For γ → ∞, the attractiveness is zero in the sight of other fireflies  FA  Simulated Annealing Because the firefly algorithm is usually a case between these two extremes, it is possible to adjust the parameter γ and α so that it can outperform both simulated annealing and PSO.

A further advantage of FA is that different fireflies will work almost independently. It is thus particularly suitable for parallel implementation.

8.2.3 Special Cases of FA DE, APSO, SA, and HS are special cases of FA.

8.3 Implementation

α0 = 0.5, γ = 1 and β0 = 1 25 fireflies in 20 generations

8.4 Variants of the Firefly Algorithm

8.4.1 FA Variants Discrete firefly algorithm (DFA) Chaotic firefly algorithm (CFA) Lagrangian firefly algorithm (LFA) Memetic firefly algorithm (MFA) Multiobjective discrete firefly algorithm (MDFA) Mulitobjective firefly algorithm (MOFA) Multi-objective enhanced firefly algorithm (MOEFA) Hybrid firefly algorithms (HFA) Parallel firefly algorithm with predation (pFAP)

8.5 Firefly Algorithms in Applications Digital image compression Highly nonlinear, multimodal design problems Antenna design optimization Discrete version of FA that can efficiently solve NP-hard scheduling problems Multi-objective load dispatch problems Classifications and clustering

Non-convex economic dispatch problem with valve-loading effect Economic load dispatch problems with reduced power losses Traveling salesman problem by discrete FA Scheduling jobs on grid computing Mixed integer programming and load dispatch problems

Training neural networks Isospectral spring-mass systems Support vector regression with the chaos-based FA for stock market price forecasting

8.6 Why the Firefly Algorithm is Efficient FA has two major advantages over other algorithms: Automatic subdivision Ability to deal with multimodality

First, FA is based on attraction and attractiveness First, FA is based on attraction and attractiveness. This leads to the fact that the whole population can automatically subdivide into subgroups, and each group can swarm around each mode or local optimum.

Second, this subdivision allows the fireflies to be able to find all optima simultaneously if the population size is sufficiently higher than the number of modes. This automatic subdivision ability makes FA particularly suitable for highly nonlinear, multimodal optimization problems.