G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 21 Statistical Methods for Particle Physics Lecture 2: statistical tests, multivariate.

Slides:



Advertisements
Similar presentations
Computing and Statistical Data Analysis / Stat 4
Advertisements

Lecture XXIII.  In general there are two kinds of hypotheses: one concerns the form of the probability distribution (i.e. is the random variable normally.
Statistical Data Analysis Stat 3: p-values, parameter estimation
G. Cowan RHUL Physics Profile likelihood for systematic uncertainties page 1 Use of profile likelihood to determine systematic uncertainties ATLAS Top.
8. Statistical tests 8.1 Hypotheses K. Desch – Statistical methods of data analysis SS10 Frequent problem: Decision making based on statistical information.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
G. Cowan 2007 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 3 Lecture 1 Probability Random variables, probability densities,
G. Cowan RHUL Physics Comment on use of LR for limits page 1 Comment on definition of likelihood ratio for limits ATLAS Statistics Forum CERN, 2 September,
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Statistics for HEP / LAL Orsay, 3-5 January 2012 / Lecture 1
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 41 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability.
G. Cowan RHUL Physics Statistical Methods for Particle Physics / 2007 CERN-FNAL HCP School page 1 Statistical Methods for Particle Physics (2) CERN-FNAL.
G. Cowan Discovery and limits / DESY, 4-7 October 2011 / Lecture 1 1 Statistical Methods for Discovery and Limits Lecture 1: Introduction, Statistical.
G. Cowan RHUL Physics Statistical Methods for Particle Physics / 2007 CERN-FNAL HCP School page 1 Statistical Methods for Particle Physics CERN-FNAL Hadron.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Statistics for HEP / NIKHEF, December 2011 / Lecture 1 1 Statistical Methods for Particle Physics Lecture 1: Introduction, Statistical Tests,
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
G. Cowan SUSSP65, St Andrews, August 2009 / Statistical Methods 3 page 1 Statistical Methods in Particle Physics Lecture 3: Multivariate Methods.
Hypothesis Testing.
G. Cowan Lectures on Statistical Data Analysis Lecture 7 page 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
G. Cowan Statistical Methods in Particle Physics1 Statistical Methods in Particle Physics Day 3: Multivariate Methods (II) 清华大学高能物理研究中心 2010 年 4 月 12—16.
IE241: Introduction to Hypothesis Testing. We said before that estimation of parameters was one of the two major areas of statistics. Now let’s turn to.
G. Cowan Lectures on Statistical Data Analysis Lecture 1 page 1 Lectures on Statistical Data Analysis London Postgraduate Lectures on Particle Physics;
G. Cowan CERN Academic Training 2010 / Statistics for the LHC / Lecture 11 Statistics for the LHC Lecture 1: Introduction Academic Training Lectures CERN,
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
G. Cowan CLASHEP 2011 / Topics in Statistical Data Analysis / Lecture 21 Topics in Statistical Data Analysis for HEP Lecture 2: Statistical Tests CERN.
G. Cowan CLASHEP 2011 / Topics in Statistical Data Analysis / Lecture 1 1 Topics in Statistical Data Analysis for HEP Lecture 1: Parameter Estimation CERN.
G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1 Lecture 2 1 Probability Definition, Bayes’ theorem, probability densities and their properties,
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 31 Introduction to Statistics − Day 3 Lecture 1 Probability Random variables, probability.
G. Cowan, RHUL Physics Discussion on significance page 1 Discussion on significance ATLAS Statistics Forum CERN/Phone, 2 December, 2009 Glen Cowan Physics.
G. Cowan RHUL Physics LR test to determine number of parameters page 1 Likelihood ratio test to determine best number of parameters ATLAS Statistics Forum.
G. Cowan S0S 2010 / Statistical Tests and Limits 1 Statistical Tests and Limits Lecture 1: general formalism IN2P3 School of Statistics Autrans, France.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan TAE 2010 / Statistics for HEP / Lecture 11 Statistical Methods for HEP Lecture 1: Introduction Taller de Altas Energías 2010 Universitat de Barcelona.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Lecture 4 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
G. Cowan Weizmann Statistics Workshop, 2015 / GDC Lecture 21 Statistical Methods for Particle Physics Lecture 2: hypothesis tests I; multivariate methods.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
G. Cowan IDPASC School of Flavour Physics, Valencia, 2-7 May 2013 / Statistical Analysis Tools 1 Statistical Analysis Tools for Particle Physics IDPASC.
1 Introduction to Statistics − Day 2 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 6 page 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1 Statistical Data Analysis: Lecture 5 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Statistical methods for HEP / Freiburg June 2011 / Lecture 1 1 Statistical Methods for Discovery and Limits in HEP Experiments Day 1: Introduction,
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
G. Cowan CERN Academic Training 2010 / Statistics for the LHC / Lecture 21 Statistics for the LHC Lecture 2: Discovery Academic Training Lectures CERN,
Statistics for HEP Lecture 1: Introduction and basic formalism
iSTEP 2016 Tsinghua University, Beijing July 10-20, 2016
Tutorial on Statistics TRISEP School 27, 28 June 2016 Glen Cowan
Comment on Event Quality Variables for Multivariate Analyses
Lecture 4 1 Probability (90 min.)
Computing and Statistical Data Analysis Stat 5: Multivariate Methods
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
TAE 2018 Benasque, Spain 3-15 Sept 2018 Glen Cowan Physics Department
Computing and Statistical Data Analysis / Stat 6
Computing and Statistical Data Analysis / Stat 7
TRISEP 2016 / Statistics Lecture 2
Statistical Analysis Tools for Particle Physics
SUSSP65, St Andrews, August 2009 / Statistical Methods 3
Statistics for HEP Lecture 1: Introduction and basic formalism
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Presentation transcript:

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 21 Statistical Methods for Particle Physics Lecture 2: statistical tests, multivariate methods Graduierten-Kolleg RWTH Aachen February 2014 Glen Cowan Physics Department Royal Holloway, University of London TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAA

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 22 Outline 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general concepts, test statistics, multivariate methods, goodness-of-fit tests 3 Parameter estimation general concepts, maximum likelihood, variance of estimators, least squares 4 Hypothesis tests for discovery and exclusion discovery significance, sensitivity, setting limits 5 Further topics systematic errors, Bayesian methods, MCMC

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 23 Hypotheses A hypothesis H specifies the probability for the data, i.e., the outcome of the observation, here symbolically: x. x could be uni-/multivariate, continuous or discrete. E.g. write x ~ f (x|H). x could represent e.g. observation of a single particle, a single event, or an entire “experiment”. Possible values of x form the sample space S (or “data space”). Simple (or “point”) hypothesis: f (x|H) completely specified. Composite hypothesis: H contains unspecified parameter(s). The probability for x given H is also called the likelihood of the hypothesis, written L(x|H).

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 24 Frequentist statistical tests Consider a hypothesis H 0 and alternative H 1. A test of H 0 is defined by specifying a critical region w of the data space such that there is no more than some (small) probability , assuming H 0 is correct, to observe the data there, i.e., P(x  w | H 0 ) ≤  Need inequality if data are discrete. α is called the size or significance level of the test. If x is observed in the critical region, reject H 0. data space Ω critical region w

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 25 Definition of a test (2) But in general there are an infinite number of possible critical regions that give the same significance level . So the choice of the critical region for a test of H 0 needs to take into account the alternative hypothesis H 1. Roughly speaking, place the critical region where there is a low probability to be found if H 0 is true, but high if H 1 is true:

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 26 Type-I, Type-II errors Rejecting the hypothesis H 0 when it is true is a Type-I error. The maximum probability for this is the size of the test: P(x  W | H 0 ) ≤  But we might also accept H 0 when it is false, and an alternative H 1 is true. This is called a Type-II error, and occurs with probability P(x  S  W | H 1 ) =  One minus this is called the power of the test with respect to the alternative H 1 : Power =  

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 27 Rejecting a hypothesis Note that rejecting H 0 is not necessarily equivalent to the statement that we believe it is false and H 1 true. In frequentist statistics only associate probability with outcomes of repeatable observations (the data). In Bayesian statistics, probability of the hypothesis (degree of belief) would be found using Bayes’ theorem: which depends on the prior probability  (H). What makes a frequentist test useful is that we can compute the probability to accept/reject a hypothesis assuming that it is true, or assuming some alternative is true.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 28 Physics context of a statistical test Event Selection: the event types in question are both known to exist. Example: separation of different particle types (electron vs muon) or known event types (ttbar vs QCD multijet). Use the selected sample for further study. Search for New Physics: the null hypothesis is H 0 : all events correspond to Standard Model (background only), and the alternative is H 1 : events include a type whose existence is not yet established (signal plus background) Many subtle issues here, mainly related to the high standard of proof required to establish presence of a new phenomenon. The optimal statistical test for a search is closely related to that used for event selection.

Aachen 2014 / Statistics for Particle Physics, Lecture 29 For each reaction we consider we will have a hypothesis for the pdf of, e.g., Example of a multivariate statistical test Suppose the result of a measurement for an individual event is a collection of numbers x 1 = number of muons, x 2 = mean p T of jets, x 3 = missing energy,... follows some n-dimensional joint pdf, which depends on the type of event produced, i.e., was it etc. E.g. call H 0 the background hypothesis (the event type we want to reject); H 1 is signal hypothesis (the type we want). G. Cowan

Aachen 2014 / Statistics for Particle Physics, Lecture 210 Defining a multivariate critical region Each event is a point in x-space; critical region is now defined by a ‘decision boundary’ in this space. What kind of boundary is best? Cuts? Linear? Nonlinear?

Aachen 2014 / Statistics for Particle Physics, Lecture 211 Test statistics The critical region’s boundary can be defined by We can work out the pdfs Critical region’s boundary is now a single ‘cut’ on t. So the original n-dimensional problem becomes in effect 1-dimensional. where t(x 1,…, x n ) is a scalar test statistic. G. Cowan

Aachen 2014 / Statistics for Particle Physics, Lecture 212 Signal/background efficiency Probability to reject background hypothesis for background event (background efficiency): Probability to accept a signal event as signal (signal efficiency): G. Cowan g(t|s) g(t|b) accept breject b

Aachen 2014 / Statistics for Particle Physics, Lecture 213 Purity of event selection Suppose only one background type b; overall fractions of signal and background events are  s and  b (prior probabilities). Suppose we select signal events with t > t cut. What is the ‘purity’ of our selected sample? Here purity means the probability to be signal given that the event was accepted. Using Bayes’ theorem we find: So the purity depends on the prior probabilities as well as on the signal and background efficiencies. G. Cowan

Aachen 2014 / Statistics for Particle Physics, Lecture 214 Constructing a test statistic How can we choose a test’s critical region in an ‘optimal way’? Neyman-Pearson lemma states: To get the highest power for a given significance level in a test of H 0, (background) versus H 1, (signal) the critical region should have inside the region, and ≤ c outside, where c is a constant which determines the power. Equivalently, optimal scalar test statistic is N.B. any monotonic function of this is leads to the same test. G. Cowan

Aachen 2014 / Statistics for Particle Physics, Lecture 215 Why Neyman-Pearson doesn’t always help The problem is that we usually don’t have explicit formulae for the pdfs P(x|H 0 ), P(x|H 1 ). Instead we may have Monte Carlo models for signal and background processes, so we can produce simulated data, and enter each event into an n-dimensional histogram. Use e.g. M bins for each of the n dimensions, total of M n cells. But n is potentially large, → prohibitively large number of cells to populate with Monte Carlo data. Compromise: make Ansatz for form of test statistic with fewer parameters; determine them (e.g. using MC) to give best discrimination between signal and background.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 216 Multivariate methods Many new (and some old) methods: Fisher discriminant Neural networks Kernel density methods Support Vector Machines Decision trees Boosting Bagging New software for HEP, e.g., TMVA, Höcker, Stelzer, Tegenfeldt, Voss, Voss, physics/ StatPatternRecognition, I. Narsky, physics/

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 217

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 218

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 219 2

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 220 Determining the coefficients w The numerator of J(w) is where The denominator is where. So the quantity to maximize is:

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 221

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 222

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 223

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 224

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 225

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 226

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 227

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 228

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 229

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 230

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 31 Neural network example from LEP II Signal: e  e  → W  W  (often 4 well separated hadron jets) Background: e  e  → qqgg (4 less well separated hadron jets) ← input variables based on jet structure, event shape,... none by itself gives much separation. Neural network output: (Garrido, Juste and Martinez, ALEPH )

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 32 Overtraining training sample independent validation sample If decision boundary is too flexible it will conform too closely to the training points → overtraining. Monitor by applying classifier to independent validation sample.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 233 Choose classifier that minimizes error function for validation sample. # of adjustable parameters in classifier ~ “flexibility” of boundary Monitor misclassification (error) rate versus “flexibility” of the boundary. For training sample this will continue to decrease.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 34 Particle i.d. in MiniBooNE Detector is a 12-m diameter tank of mineral oil exposed to a beam of neutrinos and viewed by 1520 photomultiplier tubes: H.J. Yang, MiniBooNE PID, DNP06 Search for  to e oscillations required particle i.d. using information from the PMTs.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 35 Decision trees Out of all the input variables, find the one for which with a single cut gives best improvement in signal purity: Example by MiniBooNE experiment, B. Roe et al., NIM 543 (2005) 577 where w i. is the weight of the ith event. Resulting nodes classified as either signal/background. Iterate until stop criterion reached based on e.g. purity or minimum number of events in a node. The set of cuts defines the decision boundary.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 36 Finding the best single cut The level of separation within a node can, e.g., be quantified by the Gini coefficient, calculated from the (s or b) purity as: For a cut that splits a set of events a into subsets b and c, one can quantify the improvement in separation by the change in weighted Gini coefficients: where, e.g., Choose e.g. the cut to the maximize  ; a variant of this scheme can use instead of Gini e.g. the misclassification rate:

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 37 Decision trees (2) The terminal nodes (leaves) are classified a signal or background depending on majority vote (or e.g. signal fraction greater than a specified threshold). This classifies every point in input-variable space as either signal or background, a decision tree classifier, with discriminant function f(x) = 1 if x in signal region,  1 otherwise Decision trees tend to be very sensitive to statistical fluctuations in the training sample. Methods such as boosting can be used to stabilize the tree.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 38

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 40

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 41 At each step the classifiers f k (x) are characterized by a given error rate ε k,

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 42

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 43 <

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 44

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 45 Monitoring overtraining From MiniBooNE example: Performance stable after a few hundred trees.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2page 46 Comparison of boosting algorithms A number of boosting algorithms on the market; differ in the update rule for the weights. B. Roe et al., NIM 543 (2005) 577

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 247 Testing significance / goodness-of-fit Suppose hypothesis H predicts pdf observations for a set of We observe a single point in this space: What can we say about the validity of H in light of the data? Decide what part of the data space represents less compatibility with H than does the point less compatible with H more compatible with H (Not unique!)

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 248 p-values where  (H) is the prior probability for H. Express ‘goodness-of-fit’ by giving the p-value for H: p = probability, under assumption of H, to observe data with equal or lesser compatibility with H relative to the data we got. This is not the probability that H is true! In frequentist statistics we don’t talk about P(H) (unless H represents a repeatable observation). In Bayesian statistics we do; use Bayes’ theorem to obtain For now stick with the frequentist approach; result is p-value, regrettably easy to misinterpret as P(H).

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 249 p-value example: testing whether a coin is ‘fair’ i.e. p = is the probability of obtaining such a bizarre result (or more so) ‘by chance’, under the assumption of H. Probability to observe n heads in N coin tosses is binomial: Hypothesis H: the coin is fair (p = 0.5). Suppose we toss the coin N = 20 times and get n = 17 heads. Region of data space with equal or lesser compatibility with H relative to n = 17 is: n = 17, 18, 19, 20, 0, 1, 2, 3. Adding up the probabilities for these values gives:

G. Cowan 50 Distribution of the p-value The p-value is a function of the data, and is thus itself a random variable with a given distribution. Suppose the p-value of H is found from a test statistic t(x) as Aachen 2014 / Statistics for Particle Physics, Lecture 2 The pdf of p H under assumption of H is In general for continuous data, under assumption of H, p H ~ Uniform[0,1] and is concentrated toward zero for Some (broad) class of alternatives. pHpH g(p H |H) 01 g(p H |H′)

G. Cowan 51 Using a p-value to define test of H 0 So the probability to find the p-value of H 0, p 0, less than  is Aachen 2014 / Statistics for Particle Physics, Lecture 2 We started by defining critical region in the original data space (x), then reformulated this in terms of a scalar test statistic t(x). We can take this one step further and define the critical region of a test of H 0 with size  as the set of data space where p 0 ≤ . Formally the p-value relates only to H 0, but the resulting test will have a given power with respect to a given alternative H 1.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 252 The significance of an observed signal Suppose we observe n events; these can consist of: n b events from known processes (background) n s events from a new process (signal) If n s, n b are Poisson r.v.s with means s, b, then n = n s + n b is also Poisson, mean = s + b: Suppose b = 0.5, and we observe n obs = 5. Should we claim evidence for a new discovery? Give p-value for hypothesis s = 0:

G. Cowan 53 Significance from p-value Often define significance Z as the number of standard deviations that a Gaussian variable would fluctuate in one direction to give the same p-value. 1 - TMath::Freq TMath::NormQuantile Aachen 2014 / Statistics for Particle Physics, Lecture 2

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 254 Pearson’s  2 statistic Test statistic for comparing observed data (n i independent) to predicted mean values For n i ~ Poisson( i ) we have V[n i ] = i, so this becomes (Pearson’s  2 statistic)  2 = sum of squares of the deviations of the ith measurement from the ith prediction, using  i as the ‘yardstick’ for the comparison.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 255 Pearson’s  2 test If n i are Gaussian with mean i and std. dev.  i, i.e., n i ~ N( i,  i 2 ), then Pearson’s  2 will follow the  2 pdf (here for  2 = z): If the n i are Poisson with i >> 1 (in practice OK for i > half dozen) then the Poisson dist. becomes Gaussian and therefore Pearson’s  2 statistic here as well follows the  2 pdf. The  2 value obtained from the data then gives the p-value:

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 256 The ‘  2 per degree of freedom’ Recall that for the chi-square pdf for N degrees of freedom, This makes sense: if the hypothesized i are right, the rms deviation of n i from i is  i, so each term in the sum contributes ~ 1. One often sees  2 /N reported as a measure of goodness-of-fit. But... better to give  2 and N separately. Consider, e.g., i.e. for N large, even a  2 per dof only a bit greater than one can imply a small p-value, i.e., poor goodness-of-fit.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 2 57 Pearson’s  2 with multinomial data Ifis fixed, then we might model n i ~ binomial I.e.with p i = n i / n tot.~ multinomial. In this case we can take Pearson’s  2 statistic to be If all p i n tot >> 1 then this will follow the chi-square pdf for N  1 degrees of freedom.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 258 Example of a  2 test ← This gives for N = 20 dof. Now need to find p-value, but... many bins have few (or no) entries, so here we do not expect  2 to follow the chi-square pdf.

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 259 Using MC to find distribution of  2 statistic The Pearson  2 statistic still reflects the level of agreement between data and prediction, i.e., it is still a ‘valid’ test statistic. To find its sampling distribution, simulate the data with a Monte Carlo program: Here data sample simulated 10 6 times. The fraction of times we find  2 > 29.8 gives the p-value: p = 0.11 If we had used the chi-square pdf we would find p =

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 260 Extra slides

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 261

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 262

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 263

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 264

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 265

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 266

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 267

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 268

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 269

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 270

G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 271 Kernel-based PDE (KDE, Parzen window) Consider d dimensions, N training events, x 1,..., x N, estimate f (x) with Use e.g. Gaussian kernel: kernel bandwidth (smoothing parameter) Need to sum N terms to evaluate function (slow); faster algorithms only count events in vicinity of x (k-nearest neighbor, range search).