Evaluating Hypotheses

Slides:



Advertisements
Similar presentations
Estimation of Means and Proportions
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Previous Lecture: Distributions. Introduction to Biostatistics and Bioinformatics Estimation I This Lecture By Judy Zhong Assistant Professor Division.
Evaluating Classifiers
Sampling: Final and Initial Sample Size Determination
POINT ESTIMATION AND INTERVAL ESTIMATION
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Class notes for ISE 201 San Jose State University
CS 8751 ML & KDDEvaluating Hypotheses1 Sample error, true error Confidence intervals for observed hypothesis error Estimators Binomial distribution, Normal.
Topic 2: Statistical Concepts and Market Returns
Evaluating Hypotheses
Experimental Evaluation
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Standard error of estimate & Confidence interval.
One Sample  M ean μ, Variance σ 2, Proportion π Two Samples  M eans, Variances, Proportions μ1 vs. μ2 σ12 vs. σ22 π1 vs. π Multiple.
1 Machine Learning: Lecture 5 Experimental Evaluation of Learning Algorithms (Based on Chapter 5 of Mitchell T.., Machine Learning, 1997)
Estimation Basic Concepts & Estimation of Proportions
1 Machine Learning: Experimental Evaluation. 2 Motivation Evaluating the performance of learning systems is important because: –Learning systems are usually.
Random Sampling, Point Estimation and Maximum Likelihood.
Mid-Term Review Final Review Statistical for Business (1)(2)
Estimating a Population Proportion
King Saud University Women Students
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
CpSc 881: Machine Learning Evaluating Hypotheses.
机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心. 课程基本信息  主讲教师:陈昱 Tel :  助教:程再兴, Tel :  课程网页:
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Machine Learning Chapter 5. Evaluating Hypotheses
1 CSI5388 Current Approaches to Evaluation (Based on Chapter 5 of Mitchell T.., Machine Learning, 1997)
Chapter5: Evaluating Hypothesis. 개요 개요 Evaluating the accuracy of hypotheses is fundamental to ML. - to decide whether to use this hypothesis - integral.
1 Evaluation of Learning Models Literature: Literature: T. Mitchel, Machine Learning, chapter 5 T. Mitchel, Machine Learning, chapter 5 I.H. Witten and.
Statistics 300: Elementary Statistics Sections 7-2, 7-3, 7-4, 7-5.
© Copyright McGraw-Hill 2004
Ch4: 4.3The Normal distribution 4.4The Exponential Distribution.
8.1 Estimating µ with large samples Large sample: n > 30 Error of estimate – the magnitude of the difference between the point estimate and the true parameter.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate accuracy.
Chapter Nine Hypothesis Testing.
Sampling and Sampling Distributions
Chapter Eight Estimation.
Chapter 6 Inferences Based on a Single Sample: Estimation with Confidence Intervals Slides for Optional Sections Section 7.5 Finite Population Correction.
ESTIMATION.
Two-Sample Hypothesis Testing
STATISTICAL INFERENCE
Chapter 4. Inference about Process Quality
Computational Learning Theory
STAT 312 Chapter 7 - Statistical Intervals Based on a Single Sample
Computational Learning Theory
Random Variable.
Maximum Likelihood Estimation
Evaluating Classifiers
Statistics in Applied Science and Technology
CONCEPTS OF HYPOTHESIS TESTING
CONCEPTS OF ESTIMATION
Econometric Models The most basic econometric model consists of a relationship between two variables which is disturbed by a random error. We need to use.
Computational Learning Theory
Random Variable.
Evaluating Hypotheses
Section 3: Estimating p in a binomial distribution
Computational Learning Theory
CHAPTER 6 Statistical Inference & Hypothesis Testing
LESSON 18: CONFIDENCE INTERVAL ESTIMATION
Parametric Methods Berlin Chen, 2005 References:
Evaluating Hypothesis
Chapter 8 Estimation.
Propagation of Error Berlin Chen
Machine Learning: Lecture 5
Presentation transcript:

Evaluating Hypotheses CS 9633 Machine Learning Evaluating Hypotheses Adapted from notes by Tom Mitchell http://www-2.cs.cmu.edu/~tom/mlbook-chapter-slides.html Computer Science Department CS 9633 Machine Learning

Computer Science Department Three big questions Given the observed accuracy of a hypothesis over a limited sample of data, how well does this estimate its accuracy over additional data? Given that one hypothesis outperforms another over some sample of data, how probable is it that this hypothesis is more accurate in general? When data is limited, what is the best way to use this data to both learn a hypothesis and estimate its accuracy? Computer Science Department CS 9633 Machine Learning

Difficulties in estimating hypotheses with small data sets Bias in the estimate Performance over training data provides an over-optimistic estimate of performance More likely with rich hypothesis space Typically use test set to estimate accuracy Variance in the estimate Measured accuracy over test set can still differ from true accuracy The smaller the test sample, the greater the expected variance Computer Science Department CS 9633 Machine Learning

Estimating Hypothesis Accuracy We want to know the accuracy of the hypothesis in classifying future instances We also want to know the probably error of the accuracy estimate Computer Science Department CS 9633 Machine Learning

Two Questions More Precisely Given a hypothesis h and a data sample containing n examples drawn at random according to the distribution D, what is the best estimate of the accuracy of h over future instances drawn from the same distribution? What is the probably error of the accuracy estimate? Computer Science Department CS 9633 Machine Learning

Sample Error and True Error Sample error: error rate of the hypothesis over the sample of data that is available True Error: the error rate of the hypothesis over the entire unknown distribution D of examples Computer Science Department CS 9633 Machine Learning

Computer Science Department Sample Error Definition: The sample error (denoted errors(h)) of hypothesis h with respect to target function f and data sample S is Where n is the number of examples in S, and the quantity d(f(x),h(x)) is 1 if f(x)h(x) and 0 otherwise. Computer Science Department CS 9633 Machine Learning

Computer Science Department True Error Definition: The true error (denoted errorD(h)) of hypothesis h with respect to target function f and distribution D , is the probability that h will misclassify an instance drawn at random according to D. Computer Science Department CS 9633 Machine Learning

Computer Science Department The problem We want to know errorD(h) but we can only obtain errors(h) So we must determine how well the sample error estimates the true error. Computer Science Department CS 9633 Machine Learning

Confidence Intervals for Discrete-Valued Hypotheses We wish to estimate the true error for some discrete-valued hypothesis h, based on its observed sample error over a sample S where The sample S contains n examples drawn independent of one another, and independent of h, according to the probability distribution D n 30 Hypothesis h commits r errors over these n examples Errors(h)  r/n Computer Science Department CS 9633 Machine Learning

Computer Science Department Important Result Given no other information, the most probable value of errorD(h) is errors(h) With approximately 95% probability, the true errorD(h) lies in the interval Computer Science Department CS 9633 Machine Learning

Computer Science Department More General Result Confidence Level N% 50% 68% 80% 90% 95% 98% 99% Constant zN 0.67 1.00 1.28 1.64 1.96 2.33 2.58 Computer Science Department CS 9633 Machine Learning

Where the Binomial Distribution Applies Notation Let n be the number of trials Let r be the number of successes Let p be the probability of success Conditions for Application Two possible outcomes (success and failure) The probability of success is the same on each trial There are n trials, where n is constant The n trials are independent Computer Science Department CS 9633 Machine Learning

The Binomial Distribution Computer Science Department CS 9633 Machine Learning

Mean (Expected Value) and Variance Computer Science Department CS 9633 Machine Learning

Difference in sampling error and true error errors(h) is an estimator of errorD(h) Is it an unbiased estimator? The estimation bias of an estimator Y for an arbitrary parameter p is E[Y]-p If the estimation bias is zero, Y is an unbiased estimator of P Computer Science Department CS 9633 Machine Learning

Answer to previous question What is expected value of r? E(r) = What is expected value of r/n? E(r/n) = Is r/n unbiased estimate of p? Computer Science Department CS 9633 Machine Learning

Important distinction Estimation bias is a number Inductive bias is a set of assertions Computer Science Department CS 9633 Machine Learning

Variance of an estimator Problem: r = 15 n = 50 What is variance of errors(h)? Computer Science Department CS 9633 Machine Learning

General Expression for Variance Computer Science Department CS 9633 Machine Learning

Computer Science Department Confidence Intervals Definition: An N% confidence interval for some parameter p is an interval that is expected with probability N% to contain p. Computer Science Department CS 9633 Machine Learning

Finding Confidence Intervals Question: For a given value of N how can we find the size of the interval that contains N% of the probability mass? Answer: Approximate binomial distribution with normal distribution Rule of thumb: ok for n>30 Computer Science Department CS 9633 Machine Learning

Computer Science Department Results If a random variable Y obeys a normal distribution with mean  and standard deviation , then the measured value y of Y will fall in the following interval N% of the time   zN Equivalently, the mean  will fall in the following interval N% of the time y  zN Computer Science Department CS 9633 Machine Learning

N% confidence interval 2 approximations When estimating standard deviation used errors(h) as approximation of errorD(h) Used normal distribution to approximate binomial Computer Science Department CS 9633 Machine Learning