Classification Breakdown

Slides:



Advertisements
Similar presentations
Lazy Paired Hyper-Parameter Tuning
Advertisements

Integrated Instance- and Class- based Generative Modeling for Text Classification Antti PuurulaUniversity of Waikato Sung-Hyon MyaengKAIST 5/12/2013 Australasian.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
Particle swarm optimization for parameter determination and feature selection of support vector machines Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen,
Sparse vs. Ensemble Approaches to Supervised Learning
General Mining Issues a.j.m.m. (ton) weijters Overfitting Noise and Overfitting Quality of mined models (some figures are based on the ML-introduction.
Genetic Algorithm What is a genetic algorithm? “Genetic Algorithms are defined as global optimization procedures that use an analogy of genetic evolution.
Intelligible Models for Classification and Regression
© 2013 IBM Corporation Efficient Multi-stage Image Classification for Mobile Sensing in Urban Environments Presented by Shashank Mujumdar IBM Research,
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Selective Sampling on Probabilistic Labels Peng Peng, Raymond Chi-Wing Wong CSE, HKUST 1.
M. Sulaiman Khan Dept. of Computer Science University of Liverpool 2009 COMP527: Data Mining Classification: Evaluation February 23,
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 01: Training, Testing, and Tuning Datasets.
CLassification TESTING Testing classifier accuracy
Efficient Model Selection for Support Vector Machines
Feature Selection in Nonlinear Kernel Classification Olvi Mangasarian & Edward Wild University of Wisconsin Madison Workshop on Optimization-Based Data.
1 1 Slide Evaluation. 2 2 n Interactive decision tree construction Load segmentchallenge.arff; look at dataset Load segmentchallenge.arff; look at dataset.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Special topics on text mining [ Part I: text classification ] Hugo Jair Escalante, Aurelio Lopez, Manuel Montes and Luis Villaseñor.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 5.
Machine Learning in Stock Price Trend Forecasting BY PRADEEP KUMAR REDDY MUSKU.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Gary M. Weiss Alexander Battistin Fordham University.
Online Multiple Kernel Classification Steven C.H. Hoi, Rong Jin, Peilin Zhao, Tianbao Yang Machine Learning (2013) Presented by Audrey Cheong Electrical.
USE RECIPE INGREDIENTS TO PREDICT THE CATEGORY OF CUISINE Group 7 – MEI, Yan & HUANG, Chenyu.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Classification Ensemble Methods 1
Validation methods.
Competition II: Springleaf Sha Li (Team leader) Xiaoyan Chong, Minglu Ma, Yue Wang CAMCOS Fall 2015 San Jose State University.
Final Report (30% final score) Bin Liu, PhD, Associate Professor.
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Developing outcome prediction models for acute intracerebral hemorrhage patients: evaluation of a Support Vector Machine based method A. Jakab 1, L. Lánczi.
A distributed PSO – SVM hybrid system with feature selection and parameter optimization Cheng-Lung Huang & Jian-Fan Dun Soft Computing 2008.
Does one size really fit all? Evaluating classifiers in a Bag-of-Visual-Words classification Christian Hentschel, Harald Sack Hasso Plattner Institute.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Data Science Credibility: Evaluating What’s Been Learned
Experience Report: System Log Analysis for Anomaly Detection
7. Performance Measurement
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Introduction to Machine Learning
Evaluating Classifiers
Classification with Gene Expression Data
Debesh Jha and Kwon Goo-Rak
ANOMALY DETECTION FRAMEWORK FOR BIG DATA
SLAQ: Quality-Driven Scheduling for Distributed Machine Learning
Evaluating Techniques for Image Classification
CSE 4705 Artificial Intelligence
Can Computer Algorithms Guess Your Age and Gender?
TeamMember1 TeamMember2 Machine Learning Project 2016/2017-2
Source: Procedia Computer Science(2015)70:
CS6604 Project Ensemble Classification
Schizophrenia Classification Using
Damiano Bolzoni, Sandro Etalle, Pieter H. Hartel
Classifying enterprises by economic activity
Discriminative Frequent Pattern Analysis for Effective Classification
Evaluating Models Part 2
Applying SVM to Data Bypass Prediction
CSCI N317 Computation for Scientific Applications Unit Weka
Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen, Zne-Jung Lee
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Intro to Machine Learning
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Machine Learning in Practice Lecture 27
Introduction to Machine learning
THE ASSISTIVE SYSTEM SHIFALI KUMAR BISHWO GURUNG JAMES CHOU
Outlines Introduction & Objectives Methodology & Workflow
Presenter: Donovan Orn
Presentation transcript:

Classification Breakdown Performance Evaluation of Machine Learning Algorithms by K-Fold Cross-Validation and Hold-Out Validation for Classification of Survey Write-in Responses Andrea Roberson (andrea.roberson@census.gov) Any views expressed are those of the author and not necessarily those of the U.S. Census Bureau. 50% HV vs 2 KFCV Example Background Methodology Results Conclusions Analysis of KFCV over HV on large write-in datasets We can use KFCV over HV with varying values of k for quality and more efficient classification. The computational time of fitting 100 folds was under 8.5 minutes. Accuracy increased with KFCV by up to 3%. This finding has been absent from most literature on Cross Validation. The Annual Capital Expenditures Survey (ACES) provides national estimates of capital investments in U.S. businesses. U.S. Census Bureau studies identified areas of improvement in our editing processes in order to improve the timeliness and quality of our estimates while reducing cost. A U.S. Census Bureau Economic Edit Reduction team identified edits and processes that can be automated. Suggestions included automating the manual examination of ACES survey write-ins. Technique LR 20% HV 10% HV 6.7% HV 5% HV 4% HV 2% HV 1% HV 0.913 0.916 0.907 0.917 0.899 5 KFCV 10 KFCV 15 KFCV 20 KFCV 25 KFCV 50 KFCV 100 KFCV 0.927 0.938 0.930 0.931 0.932 We compared the predictive performance of Support Vector Machines (SVMs) and Logistic Regression (LR) to predict write-in responses using various fine tuning parameters. After selecting LR for our production model, we calculated the percentage of cases where KFCV prevails over HV. Current Work The experiments were performed using Python with and an Intel core i7 processor with 16GB RAM. We consider large to be over 1,000 instances, our dataset has 18,789 instances. LR was applied on 5 KFCV and 20% HV and then again, for several values of folds between 5 and 100. A summary of these results is shown below A is the number of cases when KFCV gives better accuracy than HV. B is the number of cases when HV gives better accuracy than KFCV. C is the percentage of A over the 7 cases. Develop U.S. Census Bureau-wide best practices to apply Cross-Validation for automated text analysis. Analyze other U.S. Census Bureau training sets of various sizes that have been developed for text classification. Consider other ML methods such as Random Forests and Ensembles. Predict the upper limit of k for computational efficiency of KFCV over HV. Classification Breakdown 13% 13% Goal 63% Research and deploy a Machine Learning (ML) classifier to accurately predict the correct class of capital expenditures 10% Challenges Data was acquired from ACES survey staff for the years 2015 and 2016 50% HV 2 KFCV References There are many strategies available for Cross-validation, a method to determine the generalizability of a model to unseen data. It is unclear whether Hold-Out validation (HV) is a better validation scheme than K-Fold Cross-validation (KFCV) for large write-in survey datasets. 50% HV vs 2 KFCV Example Wong, T. (2015). Performance Evaluation of Classification Algorithms by k-fold and leave-one-out- cross validation. Pattern Recognition, vol.48, no. 9, 2839-2846. Zhang, Y. and Yang, Y. (2015). Cross-validation for selecting a model selection procedure. Journal of Econometrics, vol.187, no. 1, 95-112. U.S. Census Bureau. (2018). Annual Capital Expenditures Survey. < https://www.census.gov/programs-surveys/aces.html > 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 Dataset A B C ACES 7 100% training evaluation