System Evaluation To evaluate the error probability of the designed Pattern Recognition System Resubstitution Method – Apparent Error Overoptimistic Holdout.

Slides:



Advertisements
Similar presentations
Lecture Notes for Chapter 4 (2) Introduction to Data Mining
Advertisements

Model generalization Test error Bias, variance and complexity
Evaluation (practice). 2 Predicting performance  Assume the estimated error rate is 25%. How close is this to the true error rate?  Depends on the amount.
Evaluation.
Assuming normally distributed data! Naïve Bayes Classifier.
Importance Sampling. What is Importance Sampling ? A simulation technique Used when we are interested in rare events Examples: Bit Error Rate on a channel,
The Mean Square Error (MSE):. Now, Examples: 1) 2)
Credibility: Evaluating what’s been learned. Evaluation: the key to success How predictive is the model we learned? Error on the training data is not.
Part 2b Parameter Estimation CSE717, FALL 2008 CUBS, Univ at Buffalo.
© sebastian thrun, CMU, The KDD Lab Intro: Outcome Analysis Sebastian Thrun Carnegie Mellon University
Evaluation.
SAMPLING DISTRIBUTIONS. SAMPLING VARIABILITY
OMS 201 Review. Range The range of a data set is the difference between the largest and smallest data values. It is the simplest measure of dispersion.
Performance Evaluation: Estimation of Recognition rates J.-S. Roger Jang ( 張智星 ) CSIE Dept., National Taiwan Univ.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation X = {
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation Given.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Conduction Heat Transfer John Virtue Homework #2 01/28/00.
Chapter 7 Probability and Samples: The Distribution of Sample Means
1  The goal is to estimate the error probability of the designed classification system  Error Counting Technique  Let classes  Let data points in class.
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 01: Training, Testing, and Tuning Datasets.
2015 AprilUNIVERSITY OF HAIFA, DEPARTMENT OF STATISTICS, SEMINAR FOR M.A 1 Hastie, Tibshirani and Friedman.The Elements of Statistical Learning (2nd edition,
CLassification TESTING Testing classifier accuracy
沈致远. Test error(generalization error): the expected prediction error over an independent test sample Training error: the average loss over the training.
Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 5 of Data Mining by I. H. Witten, E. Frank and M. A. Hall 報告人:黃子齊
When estimating the mean, how large a sample must be used in order to assure a given level of confidence? Use the formula:
Error estimation Data Mining II Year Lluís Belanche Alfredo Vellido.
Introduction With such a wide variety of algorithms to choose from, which one is best? Are there any reasons to prefer one algorithm over another? Occam’s.
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation Given.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 5.
Performance of Resampling Variance Estimation Techniques with Imputed Survey data.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Occam’s Razor No Free Lunch Theorem Minimum.
CpSc 810: Machine Learning Evaluation of Classifier.
Credibility: Evaluating what’s been learned This Lecture based on Ch 5 of Witten & Frank Plan for this week 3 classes before Midterm Paper and Survey discussion.
The Bias-Variance Trade-Off Oliver Schulte Machine Learning 726.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
Model Evaluation l Metrics for Performance Evaluation –How to evaluate the performance of a model? l Methods for Performance Evaluation –How to obtain.
Bab /57 Bab 4 Classification: Basic Concepts, Decision Trees & Model Evaluation Part 2 Model Overfitting & Classifier Evaluation.
S-005 Sampling. Some basic terminology Population –Larger group (the research population) –Sometimes can be quite small Sample –A subset of the population.
Confidence Interval & Unbiased Estimator Review and Foreword.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Occam’s Razor No Free Lunch Theorem Minimum.
CSCI 347, Data Mining Evaluation: Cross Validation, Holdout, Leave-One-Out Cross Validation and Bootstrapping, Sections 5.3 & 5.4, pages
Chapter 5 Sampling Distributions. The Concept of Sampling Distributions Parameter – numerical descriptive measure of a population. It is usually unknown.
Sampling Theory and Some Important Sampling Distributions.
Validation methods.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 04: GAUSSIAN CLASSIFIERS Objectives: Whitening.
BUS304 – Chapter 7 Estimating Population Mean 1 Review – Last Week  Sampling error The radio station claims that on average a household in San Diego spends.
Q1) A precedes C. B precedes D & E. C precedes F & G. D precedes G.
Ex St 801 Statistical Methods Inference about a Single Population Mean (CI)
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Computational Intelligence: Methods and Applications Lecture 15 Model selection and tradeoffs. Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
Rodney Nielsen Many of these slides were adapted from: I. H. Witten, E. Frank and M. A. Hall Data Science Credibility: Evaluating What’s Been Learned Predicting.
THE CENTRAL LIMIT THEOREM. Sampling Distribution of Sample Means Definition: A distribution obtained by using the means computed from random samples of.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
Data Science Credibility: Evaluating What’s Been Learned
CS479/679 Pattern Recognition Dr. George Bebis
STATISTICAL INFERENCE
Bootstrap Confidence Intervals using Percentiles
POINT ESTIMATOR OF PARAMETERS
Chapter 8 Estimation.
Pattern Recognition and Training
Pattern Recognition and Training
ECE – Pattern Recognition Lecture 4 – Parametric Estimation
Presentation transcript:

System Evaluation To evaluate the error probability of the designed Pattern Recognition System Resubstitution Method – Apparent Error Overoptimistic Holdout Method - Unreliable Leave-one-out Method – Time Consuming Cross-Validation (Jack-Knife) Bootstrap Method – Biased, Smaller Variance but Time Consuming

Problem 10.2 on Page 480 In a two-class problem, the classifier to be used in the minimum Euclidean distance. Assume N i samples from class ω i for i=1,2. Show that the leave-one-out method estimate can be obtained from the resubstitution method if the distance of x from the class means d i (x), i=1,2, are modified as e i (x)=[N i /(N i -1)] 2 d i (x), if x belongs to ω i. Show that in this case, the leave-one-out method always results in larger error estimates than the resubstitution method.