Parameter Estimation and Hypothesis Testing

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Probability and Statistics Basic concepts II (from a physicist point of view) Benoit CLEMENT – Université J. Fourier / LPSC
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
Statistics review of basic probability and statistics.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Ch11 Curve Fitting Dr. Deshi Ye
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Topic 2: Statistical Concepts and Market Returns
Chapter 11 Multiple Regression.
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 41 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Inference for regression - Simple linear regression
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Today’s lesson Confidence intervals for the expected value of a random variable. Determining the sample size needed to have a specified probability of.
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Confidence intervals and hypothesis testing Petter Mostad
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
© Copyright McGraw-Hill 2004
NON-LINEAR REGRESSION Introduction Section 0 Lecture 1 Slide 1 Lecture 6 Slide 1 INTRODUCTION TO Modern Physics PHYX 2710 Fall 2004 Intermediate 3870 Fall.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
Machine Learning 5. Parametric Methods.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
Chapter 13 Understanding research results: statistical inference.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
CHAPTER 4 ESTIMATES OF MEAN AND ERRORS. 4.1 METHOD OF LEAST SQUARES I n Chapter 2 we defined the mean  of the parent distribution and noted that the.
Richard Kass/F02P416 Lecture 6 1 Lecture 6 Chi Square Distribution (  2 ) and Least Squares Fitting Chi Square Distribution (  2 ) (See Taylor Ch 8,
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Data Modeling Patrice Koehl Department of Biological Sciences
Virtual University of Pakistan
Chapter 4: Basic Estimation Techniques
The Maximum Likelihood Method
Physics 114: Lecture 13 Probability Tests & Linear Fitting
Chapter 4 Basic Estimation Techniques
Parameter Estimation and Fitting to Data
Ex1: Event Generation (Binomial Distribution)
Chapter 4. Inference about Process Quality
Probability Distributions and Monte Carlo Techniques
Virtual COMSATS Inferential Statistics Lecture-26
Ch3: Model Building through Regression
Chapter 11: Simple Linear Regression
Goodness of Fit x² -Test
Lecture 8 Preview: Interval Estimates and Hypothesis Testing
The Maximum Likelihood Method
Correlation and Simple Linear Regression
The Maximum Likelihood Method
Physics 114: Exam 2 Review Material from Weeks 7-11
I271B Quantitative Methods
CHAPTER 29: Multiple Regression*
Modelling data and curve fitting
Chi Square Distribution (c2) and Least Squares Fitting
Discrete Event Simulation - 4
Lecture 3 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Simple Linear Regression
Simple Linear Regression and Correlation
Sampling Distributions (§ )
Introduction to Statistics − Day 4
Chapter 6 Confidence Intervals.
Chapter 9 Estimation: Additional Topics
Correlation and Simple Linear Regression
Presentation transcript:

Parameter Estimation and Hypothesis Testing Maximum likelihood Least squares Hypothesis tests Goodness-of-fit Elton S. Smith Jefferson Lab Con ayuda de Eduardo Medinaceli y Cristian Peña

Dictionary / Diccionario systematic sistematico spin espin background trasfondo scintillator contador de centelleo random aleatorio (al azar) scaler escalar histogram histograma degrees of freedom grados de libertad signal sen~al power potencia maximum likelihood maxima verosimilitud test statistic prueba estadistica least squares minimos cuadrados goodness bondad/fidelidad chi-square ji-cuadrado estimation estima confidence intervals and limits intervalos de confianza y limites significance significancia frequentist frecuentista? conditional p p condicional nuissance parameters parametros molestosos? (A given B) (A dado B) outline indice fit ajuste (encaje) asymmetry asimetria parent distribution distr del patron variance varianza signal-to-background fraction sen~al/trasfondo root-mean-square raiz quadrada media (rms) sample muestra biased sesgo flip-flopping indeciso counting experiments experimentos de conteo   multiple scattering dispersion multiple correlations correlaciones weight pesadas

Parameter estimation

Properties of estimators

An estimator for the mean

An estimator for the variance

The Likelihood function via example We have a data set given by N data pairs (xi, yi±si) graphically represented below. The goal is to determine the fixed, but unknown, m = f(x). s is known or estimated from the data.

Gaussian probabilities (least-squares) We assume that at a fixed value of xi, we have made a measurement yi and that the measurement was drawn from a Gaussian probability distribution with mean y(xi) = a + bxi and variance si2.

c2 minimization

Solution for linear fit For simplicity, assume constant s = si. Then solve two simultaneous equations for 2 unknowns: Parameter uncertainties can be estimated from the curvature of the c2 function.

Parameter uncertainties In a graphical method the uncertainty in the parameter estimator q0 is obtained by changing c2 by one unit. c2(q0±sq) = c2(q0) + 1 In general, using maximum likelihood lnL(q0±sq) = lnL(q0) – 1/2

The Likelihood function via example What does the fit look like? ROOT fit to ‘pol1’ Additional information about the fit: c2 and probability Were the assumptions for the fit valid? We will return to this question after a discussion of hypothesis testing

More about the likelihood method Recall likelihood for least squares: But the probability density depends on application Proceed as before maximizing lnL (c2 has minus sign). The values of the estimated parameters might not be very different, but the uncertainties can be greatly affected.

Applications si2 = constant si2 = yi si2 = Y(x) Poisson distribution (see PDG Eq. 32.12) Stirling’s approx ln(n!) ~ n ln(n) - n

Statistical tests In addition to estimating parameters, one often wants to assess the validity of statements concerning the underlying distribution Hypothesis tests provide a rule for accepting or rejecting hypotheses depending on the outcome of an experiment. [Comparison of H0 vs H1] In goodness-of-fit tests one gives the probability to obtain a level of incompatibility with a certain hypothesis that is greater than or equal to the level observed in the actual data. [How good is my assumed H0?] Formulation of the relevant question is critical to deciding how to answer it.

Selecting events background signal

Other ways to select events

Test statistics

Significance level and power of a test

Efficiency of event selection

Linear test statistic

Fisher discriminant

Testing significance/goodness of fit Quantify the level of agreement between the data and a hypothesis without explicit reference to alternative hypotheses. This is done by defining a goodness-of-fit statistic, and the goodness-of-fit is quantified using the p-value. For the case when the c2 is the goodness-of-fit statistic, then the p-value is given by The p-value is a function of the observed value c2obs and is therefore itself a random variable. If the hypothesis used to compute the p-value is true, then p will be uniformly distributed between zero and one.

c2 distribution Gaussian-like

p-value for c2 distribution

Using the goodness-of-fit Data generated using Y(x) = 6.0 + 0.2 x, s = 0.5 Compare three different polynomial fits to the same data. y(x) = p0 + p1x y(x) = p0 y(x) = p0 + p1x + p2x2

c2/DOF vs degrees of freedom

Summary of second lecture Parameter estimation Illustrated the method of maximum likelihood using the least squares assumption Reviewed how hypotheses are tested Use of goodness-of-fit statistics to determine the validity of underlying assumptions used to determine parent parameters

Constructing a test statistic