ECE – Pattern Recognition Lecture 8 – Performance Evaluation

Slides:



Advertisements
Similar presentations
Chapter 4 Pattern Recognition Concepts: Introduction & ROC Analysis.
Advertisements

Statistical Significance and Performance Measures
Learning Algorithm Evaluation
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Other Classification Techniques 1.Nearest Neighbor Classifiers 2.Support Vector Machines.
Assessing and Comparing Classification Algorithms Introduction Resampling and Cross Validation Measuring Error Interval Estimation and Hypothesis Testing.
Model Evaluation Metrics for Performance Evaluation
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Supervised classification performance (prediction) assessment Dr. Huiru Zheng Dr. Franscisco Azuaje School of Computing and Mathematics Faculty of Engineering.
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 07: Cost-Sensitive Measures.
Classification II (continued) Model Evaluation
Evaluating Classifiers
SPH 247 Statistical Analysis of Laboratory Data May 19, 2015SPH 247 Statistical Analysis of Laboratory Data1.
Chapter 4 Pattern Recognition Concepts continued.
Evaluation – next steps
Data Analysis 1 Mark Stamp. Topics  Experimental design o Training set, test set, n-fold cross validation, thresholding, imbalance, etc.  Accuracy o.
CSSE463: Image Recognition Day 11 Lab 4 (shape) tomorrow: feel free to start in advance Lab 4 (shape) tomorrow: feel free to start in advance Test Monday.
Model Evaluation l Metrics for Performance Evaluation –How to evaluate the performance of a model? l Methods for Performance Evaluation –How to obtain.
ECE 471/571 – Lecture 2 Bayesian Decision Theory 08/25/15.
ECE 471/571 – Lecture 6 Dimensionality Reduction – Fisher’s Linear Discriminant 09/08/15.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
Classification Evaluation. Estimating Future Accuracy Given available data, how can we reliably predict accuracy on future, unseen data? Three basic approaches.
CSSE463: Image Recognition Day 11 Due: Due: Written assignment 1 tomorrow, 4:00 pm Written assignment 1 tomorrow, 4:00 pm Start thinking about term project.
ECE 471/571 - Lecture 19 Review 11/12/15. A Roadmap 2 Pattern Classification Statistical ApproachNon-Statistical Approach SupervisedUnsupervised Basic.
Evaluating Classifiers Reading: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)An introduction to ROC analysis.
ECE 471/571 – Lecture 21 Syntactic Pattern Recognition 11/19/15.
ECE 471/571 – Lecture 3 Discriminant Function and Normal Density 08/27/15.
Evaluating Classifiers. Reading for this topic: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
7. Performance Measurement
Syntactic Pattern Recognition 04/28/17
Machine Learning – Classification David Fenyő
ECE 471/571 - Lecture 19 Review 02/24/17.
Evaluating Classifiers
Nonparametric Density Estimation – k-nearest neighbor (kNN) 02/20/17
Performance Evaluation 02/15/17
CSSE463: Image Recognition Day 11
Evaluating Results of Learning
ECE 471/571 – Lecture 18 Classifier Fusion 04/12/17.
Unsupervised Learning - Clustering 04/03/17
Chapter 7 – K-Nearest-Neighbor
Measuring Success in Prediction
Unsupervised Learning - Clustering
Machine Learning Week 10.
Data Mining Classification: Alternative Techniques
Features & Decision regions
CSSE463: Image Recognition Day 11
Experiments in Machine Learning
ECE 471/571 – Lecture 12 Perceptron.
ECE 471/571 – Review 1.
Learning Algorithm Evaluation
Parametric Estimation
Hairong Qi, Gonzalez Family Professor
Model Evaluation and Selection
Evaluating Models Part 1
ECE – Pattern Recognition Lecture 16: NN – Back Propagation
ECE – Pattern Recognition Lecture 15: NN - Perceptron
ECE – Pattern Recognition Lecture 13 – Decision Tree
CSSE463: Image Recognition Day 11
Hairong Qi, Gonzalez Family Professor
CSSE463: Image Recognition Day 11
Hairong Qi, Gonzalez Family Professor
ECE – Lecture 1 Introduction.
ECE – Pattern Recognition Lecture 10 – Nonparametric Density Estimation – k-nearest-neighbor (kNN) Hairong Qi, Gonzalez Family Professor Electrical.
Bayesian Decision Theory
Hairong Qi, Gonzalez Family Professor
ECE – Pattern Recognition Lecture 4 – Parametric Estimation
Hairong Qi, Gonzalez Family Professor
ECE – Pattern Recognition Lecture 14 – Gradient Descent
ECE – Pattern Recognition Midterm Review
Presentation transcript:

ECE471-571 – Pattern Recognition Lecture 8 – Performance Evaluation Hairong Qi, Gonzalez Family Professor Electrical Engineering and Computer Science University of Tennessee, Knoxville http://www.eecs.utk.edu/faculty/qi Email: hqi@utk.edu

Pattern Classification Statistical Approach Non-Statistical Approach Supervised Unsupervised Decision-tree Basic concepts: Baysian decision rule (MPP, LR, Discri.) Basic concepts: Distance Agglomerative method Syntactic approach Parameter estimate (ML, BL) k-means Non-Parametric learning (kNN) Winner-takes-all LDF (Perceptron) Kohonen maps NN (BP) Mean-shift Support Vector Machine Deep Learning (DL) Dimensionality Reduction FLD, PCA Performance Evaluation ROC curve (TP, TN, FN, FP) cross validation Stochastic Methods local opt (GD) global opt (SA, GA) Classifier Fusion majority voting NB, BKS

Some Terminology Often used in automatic target recognition and medical diagnosis True positive The object is there and our classifier says it is there True negative The object is not there and our classifier says it is not there False negative (false misses) The object is there and our classifier says it is not there False positive (false hits) The object is not there and our classifier says it is there

Some Terminology (cont’) Sensitivity Probability of a true-positive = TP/(TP+FN) Specificity Probability of a true-negative = TN/(TN+FP) The probability of a correct decision = (TP+TN)/N, where N is the total number of samples

Some Terminology (cont’) Precision = TP/(TP+FP) Recall = TP/(TP+FN) Accuracy = (TP+TN)/(TP+TN+FP+FN) TP+FP: Total number of elements classified positives TP+FN: Total number of elements that actually belong to positives

Parameters vs. Performance Once we have designed our classifier, we invariably have some parameters we’d like to adjust. E.g., Prior probability The optimal classifier is one with sensitivity as close to 100% as possible, and at the same time with specificity as close to 100% as possible

ROC – Receiver Operating Characteristic Each curve represents the performance of a particular classifier as some parameter is varied over its range Of the three curves, the one with the sharpest bend, which passes closest to the upper left corner, is the best Calculate the area above the curve, the one with the smallest area is the best TPR: TP out of the total actual positives (Sensitivity or Recall) FPR: FP out of the total actual negatives (1-Specificity) TNR 1 0.5 FNR TPR 0.5 0.5 0.5 1 FPR

ROC (cont’d) http://en.wikipedia.org/wiki/Receiver_operating_characteristic

Determine ROC from Test Data Apparent error rate vs. true error rate Apparent error rate: counting the number of elements in the training set that are misclassified. However, this error rate leads to an optimistic result since the classifier has been designed to minimize the number of misclassifications of the training set Solutions???

Solution 1 – Separating the training set and the test set Divide the training set into two parts (randomly) and build the classifier using half of the data, then test the classifier on the other half. This other half is called the validation set. Clarification: training set, test set, validation set

Solution 2 – The Leave-One-Out Approach Assume there are n points in the training set. Remove point 1 from the set and design the classifier (determine the pdf) using the other n-1 points. Then test the classifier on point 1. Repeat for all points. The resulting error rate can be shown to be an almost unbiased estimate of the expected true error rate This requires we design n classifiers. However, we only need to do it once.

m-Fold Cross Validation A generalization to both solution 1 and 2. The training set is randomly divided into m disjoint sets of equal size. The classifier is trained m times, each time with a different set held out as a validation set For example, when m = 3, m1+m2 as training, test on m3 m1+m3 as training, test on m2 m2+m3 as training, test on m1 When m=2, it is solution 1 When m=n, it is solution 2 (n is the number of samples in the original training set)