Hierarchical Hardness Models for SAT Hierarchical Hardness Models for SAT Building the hierarchical hardness models 1.The classifier is trained on a set.

Slides:



Advertisements
Similar presentations
SATzilla: Portfolio-based Algorithm Selection for SAT Lin Xu, Frank Hutter, Holger H. Hoos, and Kevin Leyton-Brown Department of Computer Science University.
Advertisements

Hydra-MIP: Automated Algorithm Configuration and Selection for Mixed Integer Programming Lin Xu, Frank Hutter, Holger H. Hoos, and Kevin Leyton-Brown Department.
Preliminary Results (Synthetic Data) We generate a random 4-ary MRF and we sample training and test data. We forget the structure and start learning with.
Design of Experiments Lecture I
Classification Classification Examples
Integrated Instance- and Class- based Generative Modeling for Text Classification Antti PuurulaUniversity of Waikato Sung-Hyon MyaengKAIST 5/12/2013 Australasian.
IBM Labs in Haifa © 2005 IBM Corporation Adaptive Application of SAT Solving Techniques Ohad Shacham and Karen Yorav Presented by Sharon Barner.
Dynamic Restarts Optimal Randomized Restart Policies with Observation Henry Kautz, Eric Horvitz, Yongshao Ruan, Carla Gomes and Bart Selman.
Automatic Tuning1/33 Boosting Verification by Automatic Tuning of Decision Procedures Domagoj Babić joint work with Frank Hutter, Holger H. Hoos, Alan.
On the Potential of Automated Algorithm Configuration Frank Hutter, University of British Columbia, Vancouver, Canada. Motivation for automated tuning.
CMPUT 466/551 Principal Source: CMU
SATzilla-07: The Design and Analysis of an Algorithm Portfolio for SAT Lin Xu, Frank Hutter, Holger H. Hoos and Kevin Leyton-Brown University of British.
Hierarchical Hardness Models for SAT Lin Xu, Holger H. Hoos and Kevin Leyton-Brown University of British Columbia {xulin730, hoos,
Sparse vs. Ensemble Approaches to Supervised Learning
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
Supervised classification performance (prediction) assessment Dr. Huiru Zheng Dr. Franscisco Azuaje School of Computing and Mathematics Faculty of Engineering.
Ensemble Learning: An Introduction
Frank Hutter, Holger Hoos, Kevin Leyton-Brown
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Instance Hardness and Phase Transitions.
ML ALGORITHMS. Algorithm Types Classification (supervised) Given -> A set of classified examples “instances” Produce -> A way of classifying new examples.
Sparse vs. Ensemble Approaches to Supervised Learning
Sampling Combinatorial Space Using Biased Random Walks Jordan Erenrich, Wei Wei and Bart Selman Dept. of Computer Science Cornell University.
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
What is the Best Multi-Stage Architecture for Object Recognition Kevin Jarrett, Koray Kavukcuoglu, Marc’ Aurelio Ranzato and Yann LeCun Presented by Lingbo.
Attention Deficit Hyperactivity Disorder (ADHD) Student Classification Using Genetic Algorithm and Artificial Neural Network S. Yenaeng 1, S. Saelee 2.
Active Learning for Class Imbalance Problem
Efficient Direct Density Ratio Estimation for Non-stationarity Adaptation and Outlier Detection Takafumi Kanamori Shohei Hido NIPS 2008.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
1 Scaling multi-class Support Vector Machines using inter- class confusion Author:Shantanu Sunita Sarawagi Sunita Sarawagi Soumen Chakrabarti Soumen Chakrabarti.
On Bridging Simulation and Formal Verification Eugene Goldberg Cadence Research Labs (USA) VMCAI-2008, San Francisco, USA.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A data mining approach to the prediction of corporate failure.
Research Ranked Recall: Efficient Classification by Learning Indices That Rank Omid Madani with Michael Connor (UIUC)
Filter + Support Vector Machine for NIPS 2003 Challenge Jiwen Li University of Zurich Department of Informatics The NIPS 2003 challenge was organized to.
Parallel Algorithm Configuration Frank Hutter, Holger Hoos, Kevin Leyton-Brown University of British Columbia, Vancouver, Canada.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Heavy-Tailed Phenomena in Satisfiability and Constraint Satisfaction Problems by Carla P. Gomes, Bart Selman, Nuno Crato and henry Kautz Presented by Yunho.
BING: Binarized Normed Gradients for Objectness Estimation at 300fps
Today Ensemble Methods. Recap of the course. Classifier Fusion
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Ensemble Methods: Bagging and Boosting
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
Performance Prediction and Automated Tuning of Randomized and Parametric Algorithms Frank Hutter 1, Youssef Hamadi 2, Holger Hoos 1, and Kevin Leyton-Brown.
New Measures of Data Utility Mi-Ja Woo National Institute of Statistical Sciences.
Sparse Bayesian Learning for Efficient Visual Tracking O. Williams, A. Blake & R. Cipolloa PAMI, Aug Presented by Yuting Qi Machine Learning Reading.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
T. Messelis, S. Haspeslagh, P. De Causmaecker B. Bilgin, G. Vanden Berghe.
Performance Prediction and Automated Tuning of Randomized and Parametric Algorithms: An Initial Investigation Frank Hutter 1, Youssef Hamadi 2, Holger.
Patch Based Prediction Techniques University of Houston By: Paul AMALAMAN From: UH-DMML Lab Director: Dr. Eick.
USE RECIPE INGREDIENTS TO PREDICT THE CATEGORY OF CUISINE Group 7 – MEI, Yan & HUANG, Chenyu.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
STATISTICS AND OPTIMIZATION Dr. Asawer A. Alwasiti.
Latent Dirichlet Allocation
Implicit Hitting Set Problems Richard M. Karp Erick Moreno Centeno DIMACS 20 th Anniversary.
© Devi Parikh 2008 Devi Parikh and Tsuhan Chen Carnegie Mellon University April 3, ICASSP 2008 Bringing Diverse Classifiers to Common Grounds: dtransform.
Patrick De Causmaecker Stefaan Haspeslagh Tommy Messelis.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Efficient SAT Solving Under Assumptions Alexander Nadel 1 and Vadim Ryvchin 1,2 1 – Intel, Haifa, Israel 2 – Technion, Haifa, Israel SAT’12, Trento, Italy.
Classification Ensemble Methods 1
Balance and Filtering in Structured Satisfiability Problems Henry Kautz University of Washington joint work with Yongshao Ruan (UW), Dimitris Achlioptas.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Inference in Propositional Logic (and Intro to SAT) CSE 473.
Tommy Messelis * Stefaan Haspeslagh Burak Bilgin Patrick De Causmaecker Greet Vanden Berghe *
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
ROC Curves and Operating Points
Canonical Computation without Canonical Data Structure
Machine Learning – a Probabilistic Perspective
Presentation transcript:

Hierarchical Hardness Models for SAT Hierarchical Hardness Models for SAT Building the hierarchical hardness models 1.The classifier is trained on a set of training instances, then tested on validation and test data sets. 2.Three empirical hardness models M sat, M unsat, M uncond are trained using training data with satisfiable, unsatisfiable and mixture instances. Each model is used to predict the runtime of every instance in the test set, and the responses are r sat, r unsat, r uncond. 3.For each test instance, the predicted runtime is selected from one of three response sets based on its classification score and score threshold of sat and unsat. Those thresholds are selected by cross-validation. Classifying SAT Instances Lin Xu, Holger H. Hoos, Kevin Leyton-Brown, University of British Columbia Unconditional model: the model trained on a mixture of satisfiable and unsatisfiable instances. Magic model: if we had an oracle that could determine the satisfiability of an unsolved test instance, we could use models trained on satisfiable instances (M sat ) or unsatisfiable instances (M unsat ) to predict the runtime. Background and MotivationHierarchical Hardness Models One goal of using empirical hardness models is to predict the runtime of an algorithm based on some polytime-computable features. We use ridge linear regression in this research. The response is a linear combination of basis functions: The free parameters  are computed by: Encouraging Results rand3-var, rand3-fix QCP, SW-GCP Four types of SAT-distributions: Random 3-SAT Structured SAT 84 features can be classified into nine categories: problem size, variable-clause graph, variable graph, clause graph, balance features, proximity-to-horn formulae, LP- based, DPLL search space and local search space Solvers for different problem distributions: Random 3-SAT Structured SAT kcnfs, oksolver oksolver, zchaff, sato march_dl, satz satelite, minisat, satzoo Motivation Much simpler and more accurate empirical hardness models can be learned when all instances are either satisfiable or unsatisfiable. Approximating the performance of the magic model offers the possibility of performance gains for empirical hardness models. Using the wrong model could result in a significant reduction in performance. Key point: An accurate, computationally efficient way for distinguishing between satisfiable and unsatisfiable instances. SMLR: Sparse Multinomial Logistic Regression Input vector: the same set of raw features as that in regression Output: the probabilities of an instance belonging to each class 1. High overall classification accuracy 2. Big fraction of instances with high classification scores Distribution: QCP; Solver:satelite Left: rand3-var; Right: QCP Left: rand3-fix; Right: SW-GCP 3. High classification accuracy for a small number of features Experimental Setup Using only five features can achieve over 97% of the accuracy obtained using all features. Local search-based features turned out to be very important for all four data sets. The important features for regression and classification are similar. Using a classifier to distinguish satisfiable and unsatisfiable instances is feasible and reliable. Since all the features are computed for ridge linear regression anyway, the classification is free. NP-hardness of SAT problems indicates that no existing classifier can achieve 100% classification accuracy. Experimental Results Greater accuracy and less bias runtime prediction Distribution: rand3-var; Solver: satz References 1.L. Xu, H. H. Hoos, K. Leyton-Brown. Hierarchical hardness model. Submitted April, E. Nudelman, K. Leyton-Brown, H. H. Hoos, A. Devkar, and Y. Shoham. Understanding random SAT: Beyond the clauses-to-variables ratio. In CP 04, , Acknowledgement: This research was supported by a Precarn scholarship 16 th Annual Canadian Conference on Intelligent Systems, Victoria, BC, May, 2006