Hypothesis testing Some general concepts: Null hypothesisH 0 A statement we “wish” to refute Alternative hypotesisH 1 The whole or part of the complement.

Slides:



Advertisements
Similar presentations
Hypothesis testing Another judgment method of sampling data.
Advertisements

Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Lecture XXIII.  In general there are two kinds of hypotheses: one concerns the form of the probability distribution (i.e. is the random variable normally.
Topics Today: Case I: t-test single mean: Does a particular sample belong to a hypothesized population? Thursday: Case II: t-test independent means: Are.
Inferential Statistics
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Chap 9: Testing Hypotheses & Assessing Goodness of Fit Section 9.1: INTRODUCTION In section 8.2, we fitted a Poisson dist’n to counts. This chapter will.
Objectives (BPS chapter 24)
Likelihood ratio tests
The Mean Square Error (MSE):. Now, Examples: 1) 2)
Elementary hypothesis testing
4.1 All rights reserved by Dr.Bill Wan Sing Hung - HKBU Lecture #4 Studenmund (2006): Chapter 5 Review of hypothesis testing Confidence Interval and estimation.
Part 2b Parameter Estimation CSE717, FALL 2008 CUBS, Univ at Buffalo.
Elementary hypothesis testing
9-1 Hypothesis Testing Statistical Hypotheses Statistical hypothesis testing and confidence interval estimation of parameters are the fundamental.
Presenting: Lihu Berman
Hypothesis : Statement about a parameter Hypothesis testing : decision making procedure about the hypothesis Null hypothesis : the main hypothesis H 0.
Elementary hypothesis testing Purpose of hypothesis testing Type of hypotheses Type of errors Critical regions Significant levels Hypothesis vs intervals.
Topic 2: Statistical Concepts and Market Returns
T-test.
Probability & Statistics for Engineers & Scientists, by Walpole, Myers, Myers & Ye ~ Chapter 10 Notes Class notes for ISE 201 San Jose State University.
Hypothesis Testing for Population Means and Proportions
Chapter 3 Hypothesis Testing. Curriculum Object Specified the problem based the form of hypothesis Student can arrange for hypothesis step Analyze a problem.
STATISTICAL INFERENCE PART VI
July 3, Department of Computer and Information Science (IDA) Linköpings universitet, Sweden Minimal sufficient statistic.
July 3, A36 Theory of Statistics Course within the Master’s program in Statistics and Data mining Fall semester 2011.
Inferences About Process Quality
1 Inference About a Population Variance Sometimes we are interested in making inference about the variability of processes. Examples: –Investors use variance.
Chapter 9 Hypothesis Testing.
5-3 Inference on the Means of Two Populations, Variances Unknown
The Neymann-Pearson Lemma Suppose that the data x 1, …, x n has joint density function f(x 1, …, x n ;  ) where  is either  1 or  2. Let g(x 1, …,
Hypothesis Testing and T-Tests. Hypothesis Tests Related to Differences Copyright © 2009 Pearson Education, Inc. Chapter Tests of Differences One.
Hypothesis Testing – Introduction
1/2555 สมศักดิ์ ศิวดำรงพงศ์
The paired sample experiment The paired t test. Frequently one is interested in comparing the effects of two treatments (drugs, etc…) on a response variable.
Statistical Decision Theory
9-1 Hypothesis Testing Statistical Hypotheses Definition Statistical hypothesis testing and confidence interval estimation of parameters are.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Confidence intervals and hypothesis testing Petter Mostad
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005 Dr. John Lipp Copyright © Dr. John Lipp.
STATISTICAL INFERENCE PART VI HYPOTHESIS TESTING 1.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
1 9 Tests of Hypotheses for a Single Sample. © John Wiley & Sons, Inc. Applied Statistics and Probability for Engineers, by Montgomery and Runger. 9-1.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Consistency An estimator is a consistent estimator of θ, if , i.e., if
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
STATISTICAL INFERENCE PART IV CONFIDENCE INTERVALS AND HYPOTHESIS TESTING 1.
"Classical" Inference. Two simple inference scenarios Question 1: Are we in world A or world B?
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Statistical Significance Hypothesis Testing.
ENGR 610 Applied Statistics Fall Week 7 Marshall University CITE Jack Smith.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Example The strength of concrete depends, to some extent on the method used for drying it. Two different drying methods were tested independently on specimens.
Hypothesis Testing and Statistical Significance
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Chapter 5 STATISTICAL INFERENCE: ESTIMATION AND HYPOTHESES TESTING
LECTURE 33: STATISTICAL SIGNIFICANCE AND CONFIDENCE (CONT.)
Chapter 4. Inference about Process Quality
Hypothesis Testing – Introduction
Chapter 9 Hypothesis Testing.
STATISTICAL INFERENCE PART IV
Discrete Event Simulation - 4
Simple Linear Regression
Presentation transcript:

Hypothesis testing Some general concepts: Null hypothesisH 0 A statement we “wish” to refute Alternative hypotesisH 1 The whole or part of the complement of H 0 Common case: The statement is about an unknown parameter,   H 0 :    H 1 :    –  (  \  ) where  is a well-defined subset of the parameter space 

Simple hypothesis:  (or  –  ) contains only one point (one single value) Composite hypothesis: The opposite of simple hypothesis Critical region (Rejection region) A subset C of the sample space for the random sample X = (X 1, …, X n ) such that we reject H 0 if X  C (and accept (better phrase: do not reject ) H 0 otherwise ). The complement of C, i.e. C will be referred to as the acceptance region C is usually defined in terms of a statistic, T(X), called the test statistic

Simple null and alternative hypotheses Errors in hypothesis testing: Type I errorRejecting a true H 0 Type II errorAccepting a false H 0 Significance level  The probability of Type I error Also referred to as the size of the test or the risk level Risk of Type II error  The probability of Type II error Power  The probability of rejecting a false H 0, i.e. the probability of the complement of Type II error = = 1 – 

Writing it more “mathematically”: Classical approach: Fix  and then find a test that makes  desirably small A low value of  does not imply a low value of , rather the contrary Most powerful test A test which minimizes  for a fixed value of  is called a most powerful test (or best test) of size 

Neyman-Pearson lemma x = (x 1, …, x n ) a random sample from a distribution with p.d.f. f (x;  ) We wish to testH 0 :  =  0 (simple hypothesis) versus H 1 :  =  1 (simple hypothesis) The the most powerful test of size  has a critical region of the form where A is some non-negative constant. Proof: Se the course book Note! Both hypothesis are simple

Example:

How to find B ? If  1 >  0 then B must satisfy

If the sample x comes from a distribution belonging to the one-parameter exponential family:

“Pure significance tests” Assume we wish to test H 0 :  =  0 with a test of size  Test statistic T(x) is observed to the value t Case 1: H 1 :  >  0 The P-value is defined as Pr(T(x)  t | H 0 ) Case 2: H 1 :  <  0 The P-value is defined as Pr(T(x)  t | H 0 ) If the P-value is less than  H 0 is rejected

Case 3: H 1 :    0 The P-value is defined as the probability that T(x) is as extreme as the observed value, including that it can be extreme in two directions from H 0 In general: Consider we just have a null hypothesis, H 0, that could specify the value of a parameter (like above) a particular distribution independence between two or more variables … Important is that H 0 specifies something under which calculations are feasible Given a test statistic T = t the P-value is defined as Pr (T is as extreme as t | H 0 )

Uniformly most powerful tests (UMP) Generalizations of some concepts to composite (null and) alternative hypotheses: H 0 :    H 1 :    –  (  \  ) Power function: Size:

A test of size  is said to be uniformly most powerful (UMP) if If H 0 is simple but H 1 is composite and we have found a best test (Neyman- Pearson) for H 0 vs. H 1 ’:  =  1 where  1   – , then if this best test takes the same form for all  1   – , the test is UMP. Univariate cases: H 0 :  =  0 vs. H 1 :  >  0 (or H 1 :  <  0 ) usually UMP test is found H 0 :  =  0 vs. H 1 :    0 usually UMP test is not found

Unbiased test: A test is said to be unbiased if  (  )   for all    –  Similar test: A test is said to be similar if  (  ) =  for all    Invariant test: Assume that the hypotheses of a test are unchanged if a transformation of sample data is applied. If the critical region is not changed by this transformation, thes test is said to be invariant. Consistent test: If a test depends on the sample size n such that  (  ) =  n (  ). If lim n   n (  ) = 1 the test is said to be consistent. Efficiency: Two test of the pair of simple hypotheses H 0 and H 1. If n 1 and n 2 are the minimum sample sizes for test 1 and 2 resp. to achieve size  and power  , then the relative efficiency of test1 vs. test 2 is defined as n 2 / n 1

(Maximum) Likelihood Ratio Tests Consider again that we wish to test H 0 :    H 1 :    –  (  \  ) The Maximum Likelihood Ratio Test (MLRT) is defined as rejecting H 0 if 0    1 For simple H 0  gives a UMP test MLRT is asymptotically most powerful unbiased MLRT is asymptotically similar MLRT is asymptotically efficient

If H 0 is simple, i.e. H 0 :  =  0 the MLRT is simplified to Example

Distribution of  Sometimes  has a well-defined distribution: e.g.   A can be shown to be an ordinary t-test when the sample is from the normal distribution with unknown variance and H 0 :  =  0 Often, this is not the case. Asymptotic result: Under H 0 it can be shown that –2ln  is asymptotically  2 -distributed with d degrees of freedom, where d is the difference in estimated parameters (including “nuisance” parameters) between

Example Exp (  ) cont.

Score tests

Wald tests Score and Wald tests are particularly used in Generalized Linear Models

Confidence sets and confidence intervals Definition: Let x be a random sample from a distribution with p.d.f. f (x ;  ) where  is an unknown parameter with parameter space , i.e.   . If S X is a subset of , depending on X such that then S X is said to be a confidence set for  with confidence coeffcient (level) 1 –  For a one-dimensional parameter  we rather refer to this set as a confidence interval

Pivotal quantities A pivotal quantity is a function g of the unknown parameter  and the observations in the sample, i.e. g = g (x ;  ) whose distribution is known and independent of . Examples:

To obtain a confidence set from a pivotal quantity we write a probability statement as (1) For a one-dimensional  and g monotonic, the probability statement can be re- written as where now the limits are random variables, and the resulting observed confidence interval becomes For a k-dimensional  the transformation of (1) to a confidence set is more complicated but feasible.

In particular, a point estimator of  is often used to construct the pivotal quantity. Example:

Using the asymptotic normality of MLE:s One-dimensional parameter  : k-dimensional parameter  :

Construction of confidence intervals from hypothesis tests: Assume a test of H 0 :  =  0 vs. H 1 :    0 with critical region C(  0 ). Then a confidence set for  with confidence coefficient 1 –  is