Presenting: Lihu Berman

Slides:



Advertisements
Similar presentations
Dept of Bioenvironmental Systems Engineering National Taiwan University Lab for Remote Sensing Hydrology and Spatial Modeling STATISTICS Hypotheses Test.
Advertisements

Detection Chia-Hsin Cheng. Wireless Access Tech. Lab. CCU Wireless Access Tech. Lab. 2 Outlines Detection Theory Simple Binary Hypothesis Tests Bayes.
Shortest Vector In A Lattice is NP-Hard to approximate
Point Estimation Notes of STAT 6205 by Dr. Fan.
Hypothesis testing Another judgment method of sampling data.
Lecture XXIII.  In general there are two kinds of hypotheses: one concerns the form of the probability distribution (i.e. is the random variable normally.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Chap 9: Testing Hypotheses & Assessing Goodness of Fit Section 9.1: INTRODUCTION In section 8.2, we fitted a Poisson dist’n to counts. This chapter will.
The General Linear Model. The Simple Linear Model Linear Regression.
Likelihood ratio tests
Parameter Estimation using likelihood functions Tutorial #1
1 Introduction to Computability Theory Lecture12: Reductions Prof. Amos Israeli.
Derandomized DP  Thus far, the DP-test was over sets of size k  For instance, the Z-Test required three random sets: a set of size k, a set of size k-k’
8. Statistical tests 8.1 Hypotheses K. Desch – Statistical methods of data analysis SS10 Frequent problem: Decision making based on statistical information.
Maximum likelihood (ML) and likelihood ratio (LR) test
The Mean Square Error (MSE):. Now, Examples: 1) 2)
Hypothesis testing Some general concepts: Null hypothesisH 0 A statement we “wish” to refute Alternative hypotesisH 1 The whole or part of the complement.
Section 7.1 Hypothesis Testing: Hypothesis: Null Hypothesis (H 0 ): Alternative Hypothesis (H 1 ): a statistical analysis used to decide which of two competing.
Elementary hypothesis testing
Maximum likelihood (ML) and likelihood ratio (LR) test
1 STATISTICAL INFERENCE PART I EXPONENTIAL FAMILY & POINT ESTIMATION.
STATISTICAL INFERENCE PART VI
July 3, Department of Computer and Information Science (IDA) Linköpings universitet, Sweden Minimal sufficient statistic.
July 3, A36 Theory of Statistics Course within the Master’s program in Statistics and Data mining Fall semester 2011.
Lecture II-2: Probability Review
The Neymann-Pearson Lemma Suppose that the data x 1, …, x n has joint density function f(x 1, …, x n ;  ) where  is either  1 or  2. Let g(x 1, …,
Maximum Likelihood Estimation
Data Selection In Ad-Hoc Wireless Sensor Networks Olawoye Oyeyele 11/24/2003.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
Hypothesis Testing – Introduction
STATISTICS HYPOTHESES TEST (I) Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering National Taiwan University.
Chapter 5 Sampling and Statistics Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Inference for the mean vector. Univariate Inference Let x 1, x 2, …, x n denote a sample of n from the normal distribution with mean  and variance 
The paired sample experiment The paired t test. Frequently one is interested in comparing the effects of two treatments (drugs, etc…) on a response variable.
1 Power and Sample Size in Testing One Mean. 2 Type I & Type II Error Type I Error: reject the null hypothesis when it is true. The probability of a Type.
Statistical Decision Theory
1 7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to.
An Empirical Likelihood Ratio Based Goodness-of-Fit Test for Two-parameter Weibull Distributions Presented by: Ms. Ratchadaporn Meksena Student ID:
1 2. Independence and Bernoulli Trials Independence: Events A and B are independent if It is easy to show that A, B independent implies are all independent.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
IE241: Introduction to Hypothesis Testing. We said before that estimation of parameters was one of the two major areas of statistics. Now let’s turn to.
One Random Variable Random Process.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
1 Chapter 9 Detection of Spread-Spectrum Signals.
CHAPTER 5 SIGNAL SPACE ANALYSIS
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Consistency An estimator is a consistent estimator of θ, if , i.e., if
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Basic Concepts of Encoding Codes and Error Correction 1.
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
Chapter 5 Statistical Inference Estimation and Testing Hypotheses.
WS 2007/08Prof. Dr. J. Schütze, FB GW KI 1 Hypothesis testing Statistical Tests Sometimes you have to make a decision about a characteristic of a population.
Discrete Random Variables. Introduction In previous lectures we established a foundation of the probability theory; we applied the probability theory.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
Week 31 The Likelihood Function - Introduction Recall: a statistical model for some data is a set of distributions, one of which corresponds to the true.
ELEC 303 – Random Signals Lecture 17 – Hypothesis testing 2 Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 2, 2009.
G. Cowan Lectures on Statistical Data Analysis Lecture 6 page 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem 2Random variables and.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
Example The strength of concrete depends, to some extent on the method used for drying it. Two different drying methods were tested independently on specimens.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate accuracy.
STATISTICS HYPOTHESES TEST (I)
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Hypothesis Testing – Introduction
Chapter 9 Hypothesis Testing.
Presentation transcript:

Presenting: Lihu Berman Hypothesis Testing Presenting: Lihu Berman

Agenda Basic concepts Neyman-Pearson lemma UMP Invariance CFAR

Basic concepts X is a random vector with distribution is a parameter, belonging to the parameter space disjoint covering of the parameter space denotes the hypothesis that Binary test of hypotheses: vs. M-ary test: vs.

Basic concepts (cont.) If then is said to be a simple hypothesis Otherwise, it is said to be a composite hypothesis Example: vs. simple vs. composite hypotheses RADAR – Is there a target or not ? Physical model – is the coin we are tossing fair or not ? A two-sided test: the alternative lies on both sides of A one-sided test (for scalar ): vs.

Basic concepts (cont.) Introduce the test function for binary test of hypotheses: is a disjoint covering of the measurement space If the measurement is in the Acceptance region – is accepted If it is in the Rejection region – is rejected, and is accepted.

Basic concepts (cont.) Probability of False-Alarm (a.k.a. Size): simple: composite: i.e. the worst case Detection probability (a.k.a. Power): simple: composite:

Basic concepts (cont.) Receiving Operating Characteristics (ROC): Chance line The best test of size has the largest among all tests of that size

The Neyman-Pearson lemma and let denote the density function of X, then: Is the most powerful test of size for testing which of the two simple hypotheses is in force, for some Let:

The Neyman-Pearson (cont.) Proof: Let denote any test satisfying: Obviously:

The Neyman-Pearson (cont.) Note 1: If then the most powerful test is: Note 2: Introduce the likelihood function: Then the most powerful test can be written as:

The Neyman-Pearson (cont.) Note 3: Choosing the threshold k. Denote by the probability density of the likelihood function under , then: Note 4: If is not continuous (i.e. ) Then the previous equation might not work! In that case, use the test: Toss a coin, and choose if heads up

Binary comm. in AWGN Source Mapper ‘1’ = Enemy spotted. ‘0’ = All is clear. Prior probabilities unknown !

Binary comm. in AWGN Natural logarithm is monotone, enabling the use of Log-Likelihood

Binary comm. (cont.)

Binary comm. (cont.) Assume equal energies: and define

Binary comm. (cont.)

Binary comm. (cont.)

Binary comm. (cont.)

UMP Tests The Neyman-Pearson lemma holds for simple hypotheses. Uniformly Most Powerful tests generalize to composite hypotheses A test is UMP of size , if for any other test , we have:

UMP Tests (cont.) Consider scalar R.Vs whose PDFs are parameterized by scalar If the likelihood-ratio is monotone non-decreasing in x for any pair Karlin-Rubin Theorem (for UMP one-sided tests): Is the UMP test of size for testing , then the test:

UMP Tests (cont.) Proof: begin with fixed values By the Neyman-Pearson lemma, the most powerful test of size for testing is: As likelihood is monotone, we may replace it with the threshold test

UMP Tests (cont.) The test is independent of , so the argument holds for every making the most powerful test of size for testing the composite alternative vs. the simple hypothesis Consider now the power function At . For any because is more powerful than the test A similar argument holds for any

UMP Tests (cont.) Thus, we conclude that is non-decreasing Consequently, is also a test whose size satisfies Finally, no test with size can have power , as it would contradict Neyman-Pearson, in

A note on sufficiency The statistic T(x) is sufficient for if and only if No other statistic which can be calculated from the same sample provides any additional information as to the value of the parameter Fisher-Neyman factorization theorem: The statistic T(x) is sufficient for if and only if One can write the likelihood-ratio in terms of the sufficient statistic

UMP Tests (cont.) UMP one-sided tests exist for a host of problems ! Theorem: the one-parameter exponential family of distributions with density: has a monotone likelihood ratio in the sufficient statistic provided that is non-decreasing Proof:

UMP Tests (cont.) Example:

UMP Tests (cont.) Therefore, the test is the Uniformly Most Powerful test of size for testing

Invariance Revisit the binary communication example, but with a slight change. Source Mapper So what?! Let us continue with the log-likelihood as before… Oops

Invariance (cont.) Intuitively: search for a statistic that is invariant to the nuisance parameter Project the measurement on the subspace orthogonal to the disturbance! Optimal signals ?

Invariance (formal discussion) Let G denote a group of transformations. X has probability distribution:

Invariance (cont.) Revisit the previous example (AWGN channel with unknown bias) The measurement is distributed as

Invariance (cont.) organizes the measurements x into equivalent classes where:

Invariance (cont.)

Invariance (cont.) Let us show that is indeed a maximal invariant statistic

Invariance (another example) Consider the group of transformations: The hypothesis test problem is invariant to G

Invariance (another example) What statistic is invariant to the scale of S ? The angle between the measurement and the signal-subspace (or the subspace orthogonal to it: ) In fact, Z is a maximal invariant statistic to a broader group G’, that includes also rotation in the subspace. G’ is specifically appropriate for channels that introduce rotation in as well as gain

Invariance (UMPI & summary) Invariance may be used to compress measurements into statistics of low dimensionality, that satisfy invariance conditions. It is often possible to find a UMP test within the class of invariant tests. Steps when applying invariance: 1. Find a meaningful group of transformations, for which the hypothesis testing problem is invariant. 2. Find a maximal invariant statistic M, and construct a likelihood ratio test. 3. If M has a monotone likelihood ratio, then the test is UMPI for testing one sided hypotheses of the form Note: Sufficiency principals may facilitate this process.

CFAR (introductory example) Project the measurement on the signal space. A UMP test ! The False-Alarm Rate is Constant. Thus: CFAR

CFAR (cont.) m depends now on the unknown . Test is useless. Certainly not CFAR Redraw the problem as: Utilize Invariance !!

CFAR (cont.) As before: Change slightly: independent

CFAR (cont.) The distribution of is completely characterized under even though is unknown !!! Thus, we can set a threshold in the test: in order to obtain CFAR ! Furthermore, as the likelihood ratio for non-central t is monotone, this test is UMPI for testing in the distribution when is unknown !

CFAR (cont.) The actual probability of detection depends on the actual value of the SNR

Summary Basic concepts Neyman-Pearson lemma UMP Invariance CFAR