Reducing Training Time in a One-shot Machine Learning-based Compiler

Slides:



Advertisements
Similar presentations
Efficient Program Compilation through Machine Learning Techniques Gennady Pekhimenko IBM Canada Angela Demke Brown University of Toronto.
Advertisements

K-means Clustering Ke Chen.
DIMENSIONALITY REDUCTION: FEATURE EXTRACTION & FEATURE SELECTION Principle Component Analysis.
Graduate : Sheng-Hsuan Wang
K Means Clustering , Nearest Cluster and Gaussian Mixture
U NIVERSITY OF D ELAWARE C OMPUTER & I NFORMATION S CIENCES D EPARTMENT Mitigating the Compiler Optimization Phase- Ordering Problem using Machine Learning.
The Disputed Federalist Papers : SVM Feature Selection via Concave Minimization Glenn Fung and Olvi L. Mangasarian CSNA 2002 June 13-16, 2002 Madison,
Spike Sorting Goal: Extract neural spike trains from MEA electrode data Method 1: Convolution of template spikes Method 2: Sort by spikes features.
Paper presentation for CSI5388 PENGCHENG XI Mar. 23, 2005
Dave Lattanzi’s RRT Algorithm. General Concept Use dictionaries for trees Create a randomized stack of nodes Iterate through stack “Extend” each tree.
Feature Selection Presented by: Nafise Hatamikhah
By Fernando Seoane, April 25 th, 2006 Demo for Non-Parametric Classification Euclidean Metric Classifier with Data Clustering.
1 Ensembles of Nearest Neighbor Forecasts Dragomir Yankov, Eamonn Keogh Dept. of Computer Science & Eng. University of California Riverside Dennis DeCoste.
CS Instance Based Learning1 Instance Based Learning.
CISC673 – Optimizing Compilers1/34 Presented by: Sameer Kulkarni Dept of Computer & Information Sciences University of Delaware Phase Ordering.
Review: Intro to recognition Recognition tasks Machine learning approach: training, testing, generalization Example classifiers Nearest neighbor Linear.
Knowledge Acquisition from Game Records Takuya Kojima, Atsushi Yoshikawa Dept. of Computer Science and Information Engineering National Dong Hwa University.
Data Mining Joyeeta Dutta-Moscato July 10, Wherever we have large amounts of data, we have the need for building systems capable of learning information.
CISC Machine Learning for Solving Systems Problems Arch Explorer Lecture 5 John Cavazos Dept of Computer & Information Sciences University of Delaware.
Mobility Limited Flip-Based Sensor Networks Deployment Reporter: Po-Chung Shih Computer Science and Information Engineering Department Fu-Jen Catholic.
Adapting Convergent Scheduling Using Machine Learning Diego Puppin*, Mark Stephenson †, Una-May O’Reilly †, Martin Martin †, and Saman Amarasinghe † *
Self-organizing Maps Kevin Pang. Goal Research SOMs Research SOMs Create an introductory tutorial on the algorithm Create an introductory tutorial on.
1 A Bayesian Method for Guessing the Extreme Values in a Data Set Mingxi Wu, Chris Jermaine University of Florida September 2007.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Dept. of Computer and Information Sciences : University of Delaware John Cavazos Department of Computer and Information Sciences University of Delaware.
Apache Mahout. Mahout Introduction Machine Learning Clustering K-means Canopy Clustering Fuzzy K-Means Conclusion.
Generalized Fuzzy Clustering Model with Fuzzy C-Means Hong Jiang Computer Science and Engineering, University of South Carolina, Columbia, SC 29208, US.
Analysis of algorithms Analysis of algorithms is the branch of computer science that studies the performance of algorithms, especially their run time.
CISC Machine Learning for Solving Systems Problems Presented by: Alparslan SARI Dept of Computer & Information Sciences University of Delaware
Pattern Recognition April 19, 2007 Suggested Reading: Horn Chapter 14.
Click to edit Master subtitle style 2/23/10 Time and Space Optimization of Document Content Classifiers Dawei Yin, Henry S. Baird, and Chang An Computer.
CISC Machine Learning for Solving Systems Problems John Cavazos Dept of Computer & Information Sciences University of Delaware
CISC Machine Learning for Solving Systems Problems Presented by: Satyajeet Dept of Computer & Information Sciences University of Delaware Automatic.
A split-and-merge framework for 2D shape summarization D. Gerogiannis, C. Nikou and A. Likas Department of Computer Science, University of Ioannina, Greece.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
DDM Kirk. LSST-VAO discussion: Distributed Data Mining (DDM) Kirk Borne George Mason University March 24, 2011.
On Utillizing LVQ3-Type Algorithms to Enhance Prototype Reduction Schemes Sang-Woon Kim and B. John Oommen* Myongji University, Carleton University*
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
CISC Machine Learning for Solving Systems Problems Presented by: Eunjung Park Dept of Computer & Information Sciences University of Delaware Solutions.
U NIVERSITY OF D ELAWARE C OMPUTER & I NFORMATION S CIENCES D EPARTMENT Intelligent Compilation John Cavazos Computer & Information Sciences Department.
Support-Vector Networks C Cortes and V Vapnik (Tue) Computational Models of Intelligence Joon Shik Kim.
A Parallel Mixture of SVMs for Very Large Scale Problems Ronan Collobert Samy Bengio Yoshua Bengio Prepared : S.Y.C. Neural Information Processing Systems,
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Optimal Relay Placement for Indoor Sensor Networks Cuiyao Xue †, Yanmin Zhu †, Lei Ni †, Minglu Li †, Bo Li ‡ † Shanghai Jiao Tong University ‡ HK University.
Color Image Segmentation Mentor : Dr. Rajeev Srivastava Students: Achit Kumar Ojha Aseem Kumar Akshay Tyagi.
A Binary Linear Programming Formulation of the Graph Edit Distance Presented by Shihao Ji Duke University Machine Learning Group July 17, 2006 Authors:
Accelerating K-Means Clustering with Parallel Implementations and GPU Computing Janki Bhimani Miriam Leeser Ningfang Mi
BAHIR DAR UNIVERSITY Institute of technology Faculty of Computing Department of information technology Msc program Distributed Database Article Review.
Machine Learning with Spark MLlib
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
School of Computer Science & Engineering
Prepared by: Mahmoud Rafeek Al-Farra
Generalization Bounds for Indefinite Kernel Machines
Saisai Gong, Wei Hu, Yuzhong Qu
Machine Learning I & II.
Palm Oil Plantation Area Clusterization for Monitoring
Chapter 15 QUERY EXECUTION.
An Introduction to Support Vector Machines
Machine Learning Week 1.
MILEPOST GCC Lecture 4 John Cavazos
Nearest-Neighbor Classifiers
Prepared by: Mahmoud Rafeek Al-Farra
Christophe Dubach, Timothy M. Jones and Michael F.P. O’Boyle
Presented by: Divya Muppaneni
Department of Computer Science University of York
University of Wisconsin-Madison
Implementation of neural gas on Cell Broadband Engine
Introduction to Algorithm and its Complexity Lecture 1: 18 slides
Concave Minimization for Support Vector Machine Classifiers
Adaptive Occlusion Boundary Extraction for Depth Inference
Presentation transcript:

Reducing Training Time in a One-shot Machine Learning-based Compiler John Thomson, Michael O'Boyle, Grigori Bursin, Björn Franke Presented by: Muhsin Zahid UGUR Dept of Computer & Information Sciences University of Delaware

A brief introduction of the paper The cluster-based approach Results

A brief introduction Iterative compilation Performance Training cost

The cluster-based approach

The cluster-based approach (the steps in detail) Clustering Clustered using GustafsonKessel algorithm Distances are minimized

The cluster-based approach (the steps in detail) - cont. Training Find the best optimization settings Build a model One-shot compilation Use a nearest neighbor model Deployment Extracted features input to nearest neighbor classifier Benchmark compiled and executed

Cluster approach 6 typical programs represent the clusters Select 4000 random flag settings Best performing one recorded

Standard Random Training Selection Use random selection to select programs to train on 6 benchmarks Robust mean performance

Generating the upper bound Apply 4000 different optimizations A reasonable upper bound limit

Results

Results (cont.)

Results (cont.)

Conclusion Reduce the amount of training Better characterize the program-space 1.14 speedup on EEMBCv2

Questions?

Thank you.