1 A classification approach for structure discovery in search spaces of combinatorial optimization problems Daniel Porumbel 1, 2, *, Jin Kao Hao 2, Pascale.

Slides:



Advertisements
Similar presentations
Michele Samorani Manuel Laguna. PROBLEM : In meta-heuristic methods that are based on neighborhood search, whenever a local optimum is encountered, the.
Advertisements

Fast Parallel Similarity Search in Multimedia Databases (Best Paper of ACM SIGMOD '97 international conference)
1 Reinforced Tabu Search (RTS) for Graph Colouring Daniel Porumbel PhD student (joint work with Jin Kao Hao and Pascale Kuntz) Laboratoire d’Informatique.
Tabu Search Strategy Hachemi Bennaceur 5/1/ iroboapp project, 2013.
1 Appendix B: Solving TSP by Dynamic Programming Course: Algorithm Design and Analysis.
Polynomial Time Approximation Schemes Presented By: Leonid Barenboim Roee Weisbert.
Applications of Single and Multiple UAV for Patrol and Target Search. Pinsky Simyon. Supervisor: Dr. Mark Moulin.
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
Using Structure Indices for Efficient Approximation of Network Properties Matthew J. Rattigan, Marc Maier, and David Jensen University of Massachusetts.
Region Segmentation. Find sets of pixels, such that All pixels in region i satisfy some constraint of similarity.
University of CreteCS4831 The use of Minimum Spanning Trees in microarray expression data Gkirtzou Ekaterini.
Optimization via Search CPSC 315 – Programming Studio Spring 2009 Project 2, Lecture 4 Adapted from slides of Yoonsuck Choe.
Nature’s Algorithms David C. Uhrig Tiffany Sharrard CS 477R – Fall 2007 Dr. George Bebis.
Reporter : Mac Date : Multi-Start Method Rafael Marti.
MAE 552 – Heuristic Optimization
A TABU SEARCH APPROACH TO POLYGONAL APPROXIMATION OF DIGITAL CURVES.
CS 1 – Introduction to Computer Science Introduction to the wonderful world of Dr. T Daniel Tauritz, Ph.D. Associate Professor of Computer Science.
1 CSE 417: Algorithms and Computational Complexity Winter 2001 Lecture 25 Instructor: Paul Beame.
Scientific Method Study Guide
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Metaheuristics The idea: search the solution space directly. No math models, only a set of algorithmic steps, iterative method. Find a feasible solution.
Metaheuristics Meta- Greek word for upper level methods
Data Mining Techniques
Unsupervised Learning. CS583, Bing Liu, UIC 2 Supervised learning vs. unsupervised learning Supervised learning: discover patterns in the data that relate.
Algorithms for Network Optimization Problems This handout: Minimum Spanning Tree Problem Approximation Algorithms Traveling Salesman Problem.
Understand local search through visualization and animation A way for debugging and tuning local search The behavior of local search algorithms for solving.
Fixed Parameter Complexity Algorithms and Networks.
Approximation Algorithms for NP-hard Combinatorial Problems Magnús M. Halldórsson Reykjavik University
“Study on Parallel SVM Based on MapReduce” Kuei-Ti Lu 03/12/2015.
Introduction to Data Mining Group Members: Karim C. El-Khazen Pascal Suria Lin Gui Philsou Lee Xiaoting Niu.
Design Techniques for Approximation Algorithms and Approximation Classes.
Modeling Using Colorability Presented by: Kelley Louie CSC 252 Algorithms 252a-aj.
A two-stage approach for multi- objective decision making with applications to system reliability optimization Zhaojun Li, Haitao Liao, David W. Coit Reliability.
On Graph Query Optimization in Large Networks Alice Leung ICS 624 4/14/2011.
GRASP: A Sampling Meta-Heuristic
Heuristic Optimization Methods Tabu Search: Advanced Topics.
Image segmentation Prof. Noah Snavely CS1114
CSCI 3160 Design and Analysis of Algorithms Chengyu Lin.
Clustering What is clustering? Also called “unsupervised learning”Also called “unsupervised learning”
Course: Logic Programming and Constraints
FORS 8450 Advanced Forest Planning Lecture 11 Tabu Search.
Biologically Inspired Computing: Optimisation This is mandatory additional material for `Biologically Inspired Computing’ Contents: Optimisation; hard.
CSE 589 Part VI. Reading Skiena, Sections 5.5 and 6.8 CLR, chapter 37.
Marina Drosou, Evaggelia Pitoura Computer Science Department
An Efficient Linear Time Triple Patterning Solver Haitong Tian Hongbo Zhang Zigang Xiao Martin D.F. Wong ASP-DAC’15.
Introduction to Genetic Algorithms. Genetic Algorithms We’ve covered enough material that we can write programs that use genetic algorithms! –More advanced.
Presenter : Kuang-Jui Hsu Date : 2011/3/24(Thur.).
1 Knowledge Discovery from Transportation Network Data Paper Review Jiang, W., Vaidya, J., Balaporia, Z., Clifton, C., and Banich, B. Knowledge Discovery.
SwinTop: Optimizing Memory Efficiency of Packet Classification in Network Author: Chen, Chang; Cai, Liangwei; Xiang, Yang; Li, Jun Conference: Communication.
Ramakrishna Lecture#2 CAD for VLSI Ramakrishna
Graph Algorithms for Vision Amy Gale November 5, 2002.
IE 312 Review 1. The Process 2 Problem Model Conclusions Problem Formulation Analysis.
FastMap : Algorithm for Indexing, Data- Mining and Visualization of Traditional and Multimedia Datasets.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Clustering Wei Wang. Outline What is clustering Partitioning methods Hierarchical methods Density-based methods Grid-based methods Model-based clustering.
CS 721 Project Implementation of Hypergraph Edge Covering Algorithms By David Leung ( )
Tabu Search Applications Outlines: 1.Application of Tabu Search 2.Our Project with Tabu Search: EACIIT analytics.
Discrete Optimization MA2827 Fondements de l’optimisation discrète Approximate methods: local search Material based.
On the Ability of Graph Coloring Heuristics to Find Substructures in Social Networks David Chalupa By, Tejaswini Nallagatla.
Ning Jin, Wei Wang ICDE 2011 LTS: Discriminative Subgraph Mining by Learning from Search History.
CS 1010– Introduction to Computer Science Daniel Tauritz, Ph.D. Associate Professor of Computer Science Director, Natural Computation Laboratory Academic.
Timetable Problem solving using Graph Coloring
A Genetic Algorithm Approach to K-Means Clustering
Computers versus human brains a cooperative game for scientific discoveries Alain Hertz Polytechnique Montréal Mons, August 23, 2017.
School of Computer Science & Engineering
School of Computer Science & Engineering
Subject Name: Operation Research Subject Code: 10CS661 Prepared By:Mrs
Study Guide for ES205 Yu-Chi Ho Jonathan T. Lee Nov. 7, 2000
Clustering Wei Wang.
Topological Signatures For Fast Mobility Analysis
Presentation transcript:

1 A classification approach for structure discovery in search spaces of combinatorial optimization problems Daniel Porumbel 1, 2, *, Jin Kao Hao 2, Pascale Kuntz 1 1 LINA CS. Laboratory University of Nantes, France 2 LERIA CS. Laboratory University of Angers, France *New a ffiliation from September 2010: LGI2A CS. Lab., Artois University, France

2 Data Mining in Search Spaces Search Space: the set of all (potential) solutions of an optimization problem A heuristic Search Algorithm goes through this Search Space to (try to) find the optimal solution – Information on the structure of the Search Space is essential for heuristic guiding Main drawback: the Search Space is too large for complete enumeration or characterization Use Data Mining to discover hidden structures

3 Towards guided optimization heuristics General objective: Adapt the search process (the heuristic) according to discovered search space properties Our approach: Consider a sample of the search space, i.e. a set of “best” candidate solutions Question: How are these solutions organized/structured ?

4 Graph K-coloring problem A graph G=(V,E) is K-colorable if we can label its vertices with exactly K colors so that any two adjacent vertices have different labels (colors) Graph coloring problem Determine the minimal number K so that G is K-colorable : the chromatic number χ(G) NP-hard problem Our experimental framework : graph coloring problem

5 Encoding of Coloring Solutions Solution C = a vector (c 1 |c 2 |…|c |V| |) of labels/colors Conflict number of C: number of edges with both ends of the same color K-coloring problem ≈ find a optimal solution C* minimizing the conflict number The problem is solved if C* has no conflicts Solution C=(3|1|1|1|2) Conflict number = 2 (V3-V2 and V3-V4)

6 Search Space for K-coloring The set of all solutions has cardinal K |V| ~ a space of dimension |V| Neighborhood relation to pass from solution to solution: Pass from a solution C to neighboring C’ by changing just one color in a conflicting vertex of C Our approach for solving the coloring problem: Tabu Search

7 Tabu Search [Glover and Laguna, Tabu Search, 1987] A local search moving from solution to solution by applying the neighborhood relation – It can move to a worser solution when there is no move that decreases the number of conflicts (no down move) Risk : to repeat a set of moves again and again => Tabu list = a list of temporary forbidden moves

8 Classification of candidate solutions Question: given a set of best solutions, how are they distributed in the search space ? Our approach: 1. compute the distance (dissimilarity) between each two “best” solutions discovered by Tabu Search 2. analyze the solution distribution via the observed distance values one needs computing a distance with a low algorithmic complexity

9 Fast Distance Calculation K-Coloring: the transfer distance between partitions [Régnier, Sur quelques aspects mathematiques des problemes de classification automatique, 1965] Classical complexity : Hungarian algorithm O(|V|+k 3 ) Las Vegas algorithm : O(|V|) time if the partitions are close enough (condition required for our problem) [Porumbel, Hao, Kuntz, An improved algorithm for computing the partition distance, Discrete Applied Mathematics, 2010]

10 High Quality Solutions plotted via Multidimensional Scaling 350 best local minima represented via Multidimensional Scaling DIMACS graph G=dsjc250.5 with k = 27 These points form clusters that can be covered by spheres of small diameter.

11 Tabu Search trajectory: MDS representation We consider a Tabu Search exploring the search space We launch it from a local optimum and we let it explore This figure plots the solutions of high-quality (not worse than the starting point) Intuitively, the visited high-quality colorings are grouped in clusters

12 Consider a Tabu Search process exploring the search space: Record first high-quality solutions and compute all distances (pairwise) between them Histogram: number of pairs of solutions for observed distance value We observe: Small distances: solutions in the same cluster Large distances: solutions in different clusters Trajectory of long Tabu Search processes

13 Integrating learned information in the search process These analyses let us to the clustering hypothesis: The best candidate solutions are grouped into clusters that can be covered by spheres of specific diameter (10%|V|) Question: how to exploit such information to improve a search process? Answer: consider a sphere-based search space organization

14 Integrating learned information in the search process These analyses let us to the clustering hypothesis: The best candidate solutions are grouped into clusters that can be covered by spheres of specific diameter (10%|V|) Question: how to exploit such information to improve a Search process? Answer:consider a sphere-based search space organization

15 Numerical Results Integrating learned information (e.g. clustering information) helped a basic local search to reach competitive results We proposed one of the first local search algorithms that can compete with complex population-based heuristics [Porumbel, Hao & Kuntz, A Search Space Cartography for Guiding Graph Coloring Heuristics, Computers & OR, 2010 ] First coloring with k=223 colors for the well-studied DIMACS graph dsjc1000.9

16 Detailed Numerical Results

17 Conclusions Clustering information can be very useful in guiding any search algorithm We also employed the clustering hypothesis in an evolutionary coloring approach Such techniques can be used for any combinatorial optimization problem – given a distance function that can rapidly computed More advanced techniques can be used to improve our knowledge of search spaces Using learning in optimization seems a promising direction (e.g. the Learning and Intelligent Optimization)