Preventing Overfitting Problem: We don’t want to these algorithms to fit to ``noise’’ Reduced-error pruning : –breaks the samples into a training set and.

Slides:



Advertisements
Similar presentations
Data Pre-processing Data Cleaning : Sampling:
Advertisements

DECISION TREES. Decision trees  One possible representation for hypotheses.
Chapter 4 Pattern Recognition Concepts: Introduction & ROC Analysis.
Imbalanced data David Kauchak CS 451 – Fall 2013.
Demo: Classification Programs C4.5 CBA Minqing Hu CS594 Fall 2003 UIC.
Classification Algorithms
Learning Algorithm Evaluation
Preventing Overfitting Problem: We don’t want to these algorithms to fit to ``noise’’ The generated tree may overfit the training data –Too many branches,
Evaluating Diagnostic Accuracy of Prostate Cancer Using Bayesian Analysis Part of an Undergraduate Research course Chantal D. Larose.
Lecture Notes for Chapter 4 Part III Introduction to Data Mining
Assessing and Comparing Classification Algorithms Introduction Resampling and Cross Validation Measuring Error Interval Estimation and Hypothesis Testing.
Cost-Sensitive Classifier Evaluation Robert Holte Computing Science Dept. University of Alberta Co-author Chris Drummond IIT, National Research Council,
CS 8751 ML & KDDEvaluating Hypotheses1 Sample error, true error Confidence intervals for observed hypothesis error Estimators Binomial distribution, Normal.
© Vipin Kumar CSci 8980 Fall CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance Computing Research Center Department of Computer.
ROC Curves.
1 The Expected Performance Curve Samy Bengio, Johnny Mariéthoz, Mikaela Keller MI – 25. oktober 2007 Kresten Toftgaard Andersen.
Inspection Policies for Diagnostic Classification Brandon Blakeley Advisers: Dr. MK Jeong and Dr. Endre Boros.
Jeremy Wyatt Thanks to Gavin Brown
Evaluation of Learning Models
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 07: Cost-Sensitive Measures.
Comparison of Classification Methods for Customer Attrition Analysis Xiaohua Hu, Ph.D. Drexel University Philadelphia, PA, 19104
Evaluating Classifiers
Evaluation – next steps
Link Reconstruction from Partial Information Gong Xiaofeng, Li Kun & C. H. Lai
Mohammad Ali Keyvanrad
Performance measurement. Must be careful what performance metric we use For example, say we have a NN classifier with 1 output unit, and we code ‘1 =
Evaluating Hypotheses Reading: Coursepack: Learning From Examples, Section 4 (pp )
Copyright © 2003, SAS Institute Inc. All rights reserved. Cost-Sensitive Classifier Selection Ross Bettinger Analytical Consultant SAS Services.
Data Mining: Classification & Predication Hosam Al-Samarraie, PhD. Centre for Instructional Technology & Multimedia Universiti Sains Malaysia.
Evaluating What’s Been Learned. Cross-Validation Foundation is a simple idea – “ holdout ” – holds out a certain amount for testing and uses rest for.
Experiments in Machine Learning COMP24111 lecture 5 Accuracy (%) A BC D Learning algorithm.
Classification Performance Evaluation. How do you know that you have a good classifier? Is a feature contributing to overall performance? Is classifier.
Evaluating Results of Learning Blaž Zupan
Computational Intelligence: Methods and Applications Lecture 16 Model evaluation and ROC Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
Practical Issues of Classification Underfitting and Overfitting –Training errors –Generalization (test) errors Missing Values Costs of Classification.
Classification Algorithms Decision trees Rule-based induction Neural networks Memory(Case) based reasoning Genetic algorithms Bayesian networks Basic Principle.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
Evaluating Classification Performance
Quiz 1 review. Evaluating Classifiers Reading: T. Fawcett paper, link on class website, Sections 1-4 Optional reading: Davis and Goadrich paper, link.
Data Mining and Decision Support
Chapter 5: Credibility. Introduction Performance on the training set is not a good indicator of performance on an independent set. We need to predict.
Evaluating Classifiers Reading: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)An introduction to ROC analysis.
Testing Predictive Performance of Ecological Niche Models A. Townsend Peterson, STOLEN FROM Richard Pearson.
Chapter 5 – Evaluating Predictive Performance Data Mining for Business Analytics Shmueli, Patel & Bruce.
ROC curve estimation. Index Introduction to ROC ROC curve Area under ROC curve Visualization using ROC curve.
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
Chapter 5: Credibility. Introduction Performance on the training set is not a good indicator of performance on an independent set. We need to predict.
Evaluation of Learning Models Evgueni Smirnov. Overview Motivation Metrics for Classifier’s Evaluation Methods for Classifier’s Evaluation Comparing Data.
Evaluating Classifiers. Reading for this topic: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)
2011 Data Mining Industrial & Information Systems Engineering Pilsung Kang Industrial & Information Systems Engineering Seoul National University of Science.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Evaluating Classifiers
Evaluation – next steps
Performance Evaluation 02/15/17
Evaluating Results of Learning
Machine Learning Week 10.
Data Mining Classification: Alternative Techniques
Features & Decision regions
Diagnosis II Dr. Brent E. Faught, Ph.D. Assistant Professor
Evaluating Classifiers (& other algorithms)
Learning Algorithm Evaluation
CSCI N317 Computation for Scientific Applications Unit Weka
Model Evaluation and Selection
Intro to Machine Learning
Roc curves By Vittoria Cozza, matr
Evaluating Classifiers
ECE – Pattern Recognition Lecture 8 – Performance Evaluation
Presentation transcript:

Preventing Overfitting Problem: We don’t want to these algorithms to fit to ``noise’’ Reduced-error pruning : –breaks the samples into a training set and a test set. The tree is induced completely on the training set. –Working backwards from the bottom of the tree, the subtree starting at each nonterminal node is examined. If the error rate on the test cases improves by pruning it, the subtree is removed. The process continues until no improvement can be made by pruning a subtree, The error rate of the final tree on the test cases is used as an estimate of the true error rate.

Decision Tree Pruning : physician fee freeze = n: | adoption of the budget resolution = y: democrat (151.0) | adoption of the budget resolution = u: democrat (1.0) | adoption of the budget resolution = n: | | education spending = n: democrat (6.0) | | education spending = y: democrat (9.0) | | education spending = u: republican (1.0) physician fee freeze = y: | synfuels corporation cutback = n: republican (97.0/3.0) | synfuels corporation cutback = u: republican (4.0) | synfuels corporation cutback = y: | | duty free exports = y: democrat (2.0) | | duty free exports = u: republican (1.0) | | duty free exports = n: | | | education spending = n: democrat (5.0/2.0) | | | education spending = y: republican (13.0/2.0) | | | education spending = u: democrat (1.0) physician fee freeze = u: | water project cost sharing = n: democrat (0.0) | water project cost sharing = y: democrat (4.0) | water project cost sharing = u: | | mx missile = n: republican (0.0) | | mx missile = y: democrat (3.0/1.0) | | mx missile = u: republican (2.0) Simplified Decision Tree: physician fee freeze = n: democrat (168.0/2.6) physician fee freeze = y: republican (123.0/13.9) physician fee freeze = u: | mx missile = n: democrat (3.0/1.1) | mx missile = y: democrat (4.0/2.2) | mx missile = u: republican (2.0/1.0) Evaluation on training data (300 items): Before Pruning After Pruning Size Errors Size Errors Estimate 25 8( 2.7%) 7 13( 4.3%) ( 6.9%) <

False Positives True Positives False Negatives Actual Predicted Evaluation of Classification Systems Training Set: examples with class values for learning. Test Set: examples with class values for evaluating. Evaluation: Hypotheses are used to infer classification of examples in the test set; inferred classification is compared to known classification. Accuracy: percentage of examples in the test set that are classified correctly.

Model Evaluation Analytic goal: achieve understanding –Exploratory evaluation : understand a novel area of study –Experimental evaluation : support or refute some models Engineering goal : solve a practical problem Estimator of classifiers : accuracy –Accuracy : how well does a model classify –Higher accuracy does not necessarily imply better performance on target task

Confusion Metrics A: True + C : False - B : False + D : True - Actual Class Predicted class + - Y N Entries are counts of correct classifications and counts of errors Other evaluation metrics True positive rate (TP) = A/(A+C)= 1- false negative rate False positive rate (FP)= B/(B+D) = 1- true negative rate Sensitivity = true positive rate Specificity = true negative rate Positive predictive value = A/(A+B) Recall = A/(A+C) = true positive rate = sensitivity Precision = A/(A+B) = PPV

Probabilistic Interpretation of CM Class Distribution Confusion matrix P (+) : P (-) prior probabilities approximated by class frequencies Defined for a particular training set Defined for a particular classifier P(+ | Y) P(- | N) Posterior probabilities likelihoods approximated using error frequencies P(Y |+) P(Y |- )

More Than Accuracy Cost and Benefits –Medical diagnosis : cost of falsely indicating “cancer” is different from cost of missing a true cancer case –Fraud detection : cost of falsely challenging customer is different from cost of leaving fraud undetected –Customer segmentation : Benefit of not contacting a non-buyer is different from benefit of contacting a buyer

Model Evaluation within Context Must take costs and distributions into account Calculate expected profit: profit = P(+)*(TP*B(Y, +) + (1-TP)*C(N, +)) + P(-)*((1-FP)*B(N, -) + FP*C(Y, -)) Choose the classifier that maximises profit Benefits of correct classificationcosts of incorrect classification

Lift & Cumulative Response Curves Lift = P(+ | Y)/P(+) : How much better with model than without

ROC Curves ROC (Receiver Operating Characteristic) Space : Each classifier is represented by plotting its (FP, TP) pair Good measure of overall model performance and target conditions are completely unknown Good model : maximising AUC (Area under Curve) Better estimation : a good model extends the ROC Convex Hull

Parametric Models : Parametrically Summarise Data

Contributory Models : retain training data points; each potentially affects the estimation at new point