SATzilla: Portfolio-based Algorithm Selection for SAT Lin Xu, Frank Hutter, Holger H. Hoos, and Kevin Leyton-Brown Department of Computer Science University.

Slides:



Advertisements
Similar presentations
Detecting Statistical Interactions with Additive Groves of Trees
Advertisements

Alexander Statnikov1, Douglas Hardin1,2, Constantin Aliferis1,3
Automated Parameter Setting Based on Runtime Prediction: Towards an Instance-Aware Problem Solver Frank Hutter, Univ. of British Columbia, Vancouver, Canada.
Robust Window-based Multi-node Technology- Independent Logic Minimization Jeff L.Cobb Kanupriya Gulati Sunil P. Khatri Texas Instruments, Inc. Dept. of.
Hydra-MIP: Automated Algorithm Configuration and Selection for Mixed Integer Programming Lin Xu, Frank Hutter, Holger H. Hoos, and Kevin Leyton-Brown Department.
On Comparing Classifiers : Pitfalls to Avoid and Recommended Approach
January 23 rd, Document classification task We are interested to solve a task of Text Classification, i.e. to automatically assign a given document.
Machine Learning: Intro and Supervised Classification
Dept of Biomedical Engineering, Medical Informatics Linköpings universitet, Linköping, Sweden A Data Pre-processing Method to Increase.
I/O Placement for FPGAs with Multiple I/O Standards.
Mani Srivastava UCLA - EE Department Room: 6731-H Boelter Hall Tel: WWW: Copyright 2003.
Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
IBM Labs in Haifa © 2005 IBM Corporation Adaptive Application of SAT Solving Techniques Ohad Shacham and Karen Yorav Presented by Sharon Barner.
Dynamic Restarts Optimal Randomized Restart Policies with Observation Henry Kautz, Eric Horvitz, Yongshao Ruan, Carla Gomes and Bart Selman.
Automatic Tuning1/33 Boosting Verification by Automatic Tuning of Decision Procedures Domagoj Babić joint work with Frank Hutter, Holger H. Hoos, Alan.
On the Potential of Automated Algorithm Configuration Frank Hutter, University of British Columbia, Vancouver, Canada. Motivation for automated tuning.
Indian Statistical Institute Kolkata
Paper presentation for CSI5388 PENGCHENG XI Mar. 23, 2005
SATzilla-07: The Design and Analysis of an Algorithm Portfolio for SAT Lin Xu, Frank Hutter, Holger H. Hoos and Kevin Leyton-Brown University of British.
Hierarchical Hardness Models for SAT Lin Xu, Holger H. Hoos and Kevin Leyton-Brown University of British Columbia {xulin730, hoos,
TEMPLATE DESIGN © Genetic Algorithm and Poker Rule Induction Wendy Wenjie Xu Supervised by Professor David Aldous, UC.
Sparse vs. Ensemble Approaches to Supervised Learning
Frank Hutter, Holger Hoos, Kevin Leyton-Brown
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Hierarchical Hardness Models for SAT Hierarchical Hardness Models for SAT Building the hierarchical hardness models 1.The classifier is trained on a set.
1 Combinatorial Problems in Cooperative Control: Complexity and Scalability Carla Gomes and Bart Selman Cornell University Muri Meeting March 2002.
Hardness-Aware Restart Policies Yongshao Ruan, Eric Horvitz, & Henry Kautz IJCAI 2003 Workshop on Stochastic Search.
Prediction model building and feature selection with SVM in breast cancer diagnosis Cheng-Lung Huang, Hung-Chang Liao, Mu- Chen Chen Expert Systems with.
Large-scale Hybrid Parallel SAT Solving Nishant Totla, Aditya Devarakonda, Sanjit Seshia.
Monday, January 31, 2000 Aiming Wu Department of Computing and Information Sciences, KSU Readings: Chown and Dietterich.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A data mining approach to the prediction of corporate failure.
Dept. of Computer and Information Sciences : University of Delaware John Cavazos Department of Computer and Information Sciences University of Delaware.
Benk Erika Kelemen Zsolt
Parallel Algorithm Configuration Frank Hutter, Holger Hoos, Kevin Leyton-Brown University of British Columbia, Vancouver, Canada.
1Computer Sciences Department. Book: Introduction to Algorithms, by: Thomas H. Cormen Charles E. Leiserson Ronald L. Rivest Clifford Stein Electronic:
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Ensemble Methods: Bagging and Boosting
CLASSIFICATION: Ensemble Methods
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
Performance Prediction and Automated Tuning of Randomized and Parametric Algorithms Frank Hutter 1, Youssef Hamadi 2, Holger Hoos 1, and Kevin Leyton-Brown.
Evolutionary Algorithms for Finding Optimal Gene Sets in Micro array Prediction. J. M. Deutsch Presented by: Shruti Sharma.
T. Messelis, S. Haspeslagh, P. De Causmaecker B. Bilgin, G. Vanden Berghe.
Experimental Algorithmics Reading Group, UBC, CS Presented paper: Fine-tuning of Algorithms Using Fractional Experimental Designs and Local Search by Belarmino.
Performance Prediction and Automated Tuning of Randomized and Parametric Algorithms: An Initial Investigation Frank Hutter 1, Youssef Hamadi 2, Holger.
Multiple Instance Learning for Sparse Positive Bags Razvan C. Bunescu Machine Learning Group Department of Computer Sciences University of Texas at Austin.
Reservoir Uncertainty Assessment Using Machine Learning Techniques Authors: Jincong He Department of Energy Resources Engineering AbstractIntroduction.
Patrick De Causmaecker Stefaan Haspeslagh Tommy Messelis.
Classification Ensemble Methods 1
Text Categorization by Boosting Automatically Extracted Concepts Lijuan Cai and Tommas Hofmann Department of Computer Science, Brown University SIGIR 2003.
Machine Learning in Compiler Optimization By Namita Dave.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Tommy Messelis * Stefaan Haspeslagh Burak Bilgin Patrick De Causmaecker Greet Vanden Berghe *
Bump Hunting The objective PRIM algorithm Beam search References: Feelders, A.J. (2002). Rule induction by bump hunting. In J. Meij (Ed.), Dealing with.
Semi-Supervised Clustering
CSE 4705 Artificial Intelligence
SUNNY-CP: A Multicore Tool for Constraint Solving Jacopo Mauro University of Oslo NordConsNet Uppsala, May 2017.
Lin Xu, Holger H. Hoos, Kevin Leyton-Brown
Reading: Pedro Domingos: A Few Useful Things to Know about Machine Learning source: /cacm12.pdf reading.
CS548 Fall 2017 Decision Trees / Random Forest Showcase by Yimin Lin, Youqiao Ma, Ran Lin, Shaoju Wu, Bhon Bunnag Showcasing work by Cano,
Adversarial Evasion-Resilient Hardware Malware Detectors
Lecture 18: Bagging and Boosting
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Lecture 06: Bagging and Boosting
Canonical Computation without Canonical Data Structure
Rohan Yadav and Charles Yuan (rohany) (chenhuiy)
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Presentation transcript:

SATzilla: Portfolio-based Algorithm Selection for SAT Lin Xu, Frank Hutter, Holger H. Hoos, and Kevin Leyton-Brown Department of Computer Science University of British Columbia Mainly based on [Xu, Hutter, Hoos, Leyton-Brown, JAIR, 2008]

SATzilla Excellent performance in SAT competitions [ – 2 second places, one third place in 2003 SAT competition [Leyton-Brown, Nudelman, Andrew, McFadden & Shoham, 2003] – 3 Gold, 1 Silver, 1 Bronze in 2007 SAT competition [Xu, Hutter, Hoos, & Leyton-Brown, 2007] – 3 Gold, 2 Silver in 2009 SAT competition [Xu, Hutter, Hoos & Leyton-Brown, 2009] One gold in every category: Industrial/Handmade/Random Domain independent – MIPzilla, CSP, Planning … Xu, Hutter, Hoos, and Leyton-Brown: SATzilla2 Solvers Instance Feature extractor Algorithm selector

Impact on the SAT community Encouraged research on portfolio-based solvers for SAT Final 2011 SAT competition results [ Xu, Hutter, Hoos, and Leyton-Brown: SATzilla CategoryRandomHandmadeIndustrial Medalists 3S ppfolio (2) ppfolio seq sparrow2011 sattime2011 eagleup march_rw mphasesat_m 3S (3) Ppfolio (2) ppfolio seq (2) clasp glucose glucose (2) glueminisat (2) lingeling contrasat hack cirminisat hack mphasesat64 qutersat Total medals4/97/90/9 3

Outline Introduction Related Work: – Algorithm selection SATzilla overview: – Building SATzilla – Running SATzilla SATzilla07 and beyond Results SATzilla since 2008 Conclusion Xu, Hutter, Hoos, and Leyton-Brown: SATzilla4

Related work: algorithm selection Select the best solver for each instance [Rice, 76] – Regression-based Linear regression (SATzilla) [Leyton-Brown, Nudelman, Andrew, McFadden & Shoham, 2003; Xu, Hutter, Hoos, Leyton-Brown, 2007+] Matchbox [Stern et al., 2010] – Classification-based Logistic regression [Samulowitz & Memisevic, 2007] K-NN [OMahony & Hebrard, 2008] Decision tree [Guerri & Milano, 2004] Xu, Hutter, Hoos, and Leyton-Brown: SATzilla5

What is SATzilla? Portfolio-based algorithm selection based on empirical hardness models [Leyton-Brown et al., 2002;2009] – Use regression models to predict each algorithms performance – Select the predicted best Many new techniques for improving robustness – Pre-solvers – Backup solver – Solver subset selection Xu, Hutter, Hoos, and Leyton-Brown: SATzilla6

SATzilla: prerequisites Identify a target distribution of instances Select a set of candidate solvers, and measure runtime on a training set Identify a set of features based on domain knowledge and collect feature values – Cheap to compute – Informative for prediction Xu, Hutter, Hoos, and Leyton-Brown: SATzilla7

SATzilla core: predictive models Goal: accurately predict solver performance based on cheaply computable features Supervised machine learning: learn function f: features performance – Input: pairs – Output: f Xu, Hutter, Hoos, and Leyton-Brown: SATzilla f w () = w T Performance(y) Features () … Algorithm selector 8

SATzilla: selection based on models Predict each solvers performance Pick solver based on predictions Feature extractor Algorithm selector Instance Compute features Make prediction Run on selected solver Xu, Hutter, Hoos, and Leyton-Brown: SATzilla9

Improving robustness Solver subset selection – More candidate solvers can hurt – select subset of solvers that achieves best overall performance Pre-solvers – Solve some instances quickly without need for feature computation – Increase model robustness Backup solver – In case something goes wrong Feature extractor Algorithm selector Dealing with imperfect predictions: Xu, Hutter, Hoos, and Leyton-Brown: SATzilla Algorithm selector 10

Time to run SATzilla! Xu, Hutter, Hoos, and Leyton-Brown: SATzilla11

Step 1: pre-solving Run pre-solvers for a fixed short amount of time – If solved, done Instance Feature extractor Algorithm selector Xu, Hutter, Hoos, and Leyton-Brown: SATzilla12

Step 2: compute features Run feature extractor to get a set of feature values – If solved, done – If anything fails, simply run backup solver Instance Feature extractor Algorithm selector Xu, Hutter, Hoos, and Leyton-Brown: SATzilla13

Step 3: performance prediction Predict performance of solvers and order them according to runtime predictions Feature extractor Algorithm selector Xu, Hutter, Hoos, and Leyton-Brown: SATzilla14

Step 4: run top-ranked solver Run the top-ranked solver. If a run fails, run the next-ranked solver. Xu, Hutter, Hoos, and Leyton-Brown: SATzilla15 Feature extractor Algorithm selector Instance

SATzilla07: construction Pre-solvers: manually selected Backup solver: overall best Model building: – Predict runtime – Hierarchical hardness models [Xu, Hoos, Leyton-Brown, 2007] – Models handle censored data [Schmee & Hahn, 1979] Solver subset selection: exhaustive search Xu, Hutter, Hoos, and Leyton-Brown: SATzilla16

Improving SATzilla07 (1.) Problem: manual pre-solver selection is not ideal Solution: auto pre-solver selection based on validation Xu, Hutter, Hoos, and Leyton-Brown: SATzilla17 Feature extractor Algorithm selector

Improving SATzilla07 (2.-5.) Problem: overall best solver best backup solver Solution: choose the best solver for instances not solved by pre-solvers and with problematic features Problem: objective is not always runtime Solution: predict and optimize true objective directly also enables cleaner use of incomplete solvers Problem: exhaustive search for solver subset selection is infeasible for large portfolios Solution: local search for solver subset selection Xu, Hutter, Hoos, and Leyton-Brown: SATzilla18

SATzilla vs candidate solvers HANDMADE Xu, Hutter, Hoos, and Leyton-Brown: SATzilla19

SATzilla vs candidate solvers HANDMADE Xu, Hutter, Hoos, and Leyton-Brown: SATzilla Oracle 20

SATzilla vs candidate solvers HANDMADE Pre-solving Feat. Comp. Xu, Hutter, Hoos, and Leyton-Brown: SATzilla Oracle SATzilla07 21

SATzilla vs candidate solvers HANDMADE Xu, Hutter, Hoos, and Leyton-Brown: SATzilla Oracle SATzilla07 (improved) 22

SATzilla09 [Xu et al., 2009] New techniques – New features: clause learning, … – Prediction of feature computation cost Feature time predictor Instance Minimal cost feature extractor Results: 3 Gold, 2 Silver in 2009 SAT competition - Gold for Industrial (SAT) category - Gold for Handmade (UNSAT) category - Gold for Random (SAT+UNSAT) category Xu, Hutter, Hoos, and Leyton-Brown: SATzilla23 Feature extractor Algorithm selector

New Results: Cost sensitive classification Goal: directly optimize the objective [Xu et al., 2011] – Cost sensitive classification for every pair of candidates – Vote to decide the order of candidates Improved performance over regression models, particularly with heterogeneous data sets New SATzilla on SAT competition 2011 Xu, Hutter, Hoos, and Leyton-Brown: SATzilla Random Gap closed: 80% 55% 33% Oracle Handmade Industrial 24

Recent work: Combining algorithm selection and algorithm configuration Problem: what if you only have one strong solver but it is highly parameterized? Hydra: combine algorithm selection and algorithm configuration [Xu et al., 2010; 2011] – Automatically find strong, uncorrelated parameter settings – Construct portfolio from these Results: clear improvement for both SAT and MIP – 2 × to 31 × faster for MIP – 1.2 × to 73 × faster for SAT Xu, Hutter, Hoos, and Leyton-Brown: SATzilla25

Conclusion Portfolio-based algorithm selection can improve performance and robustness SATzilla-like portfolio approaches are the state of the art for many problems New directions since our JAIR paper: – better predictive models – automatic portfolio construction (Hydra) Xu, Hutter, Hoos, and Leyton-Brown: SATzilla26