1 Practical Statistics for Physicists Sardinia Lectures Oct 2008 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
DO’S AND DONT’S WITH LIKELIHOODS Louis Lyons Oxford (CDF)
Psyc 235: Introduction to Statistics
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
1 χ 2 and Goodness of Fit Louis Lyons IC and Oxford SLAC Lecture 2’, Sept 2008.
1 Do’s and Dont’s with L ikelihoods Louis Lyons Oxford and IC CDF IC, February 2008.
1 χ 2 and Goodness of Fit Louis Lyons Oxford Mexico, November 2006 Lecture 3.
Chapter 1 Probability Theory (i) : One Random Variable
1 Practical Statistics for Physicists LBL January 2008 Louis Lyons Oxford
1 Do’s and Dont’s with L ikelihoods Louis Lyons Oxford CDF Mexico, November 2006.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
1 Practical Statistics for Physicists Dresden March 2010 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC
1 Do’s and Dont’s with L ikelihoods Louis Lyons Oxford CDF CERN, October 2006.
1 Practical Statistics for Physicists SLUO Lectures February 2007 Louis Lyons Oxford
1 Practical Statistics for Particle Physicists CERN Academic Lectures October 2006 Louis Lyons Oxford
Lec 6, Ch.5, pp90-105: Statistics (Objectives) Understand basic principles of statistics through reading these pages, especially… Know well about the normal.
1 Do’s and Dont’s with L ikelihoods Louis Lyons Oxford CDF Manchester, 16 th November 2005.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Overview Parameters and Statistics Probabilities The Binomial Probability Test.
K. Desch – Statistical methods of data analysis SS10 3. Distributions 3.1 Binomial distribution Error of an efficiency Estimator for the efficiency (when.
Discrete Probability Distributions
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
1 Practical Statistics for Physicists Stockholm Lectures Sept 2008 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC
Modern Navigation Thomas Herring
The Binomial, Poisson, and Normal Distributions Modified after PowerPoint by Carlos J. Rosas-Anderson.
Problem A newly married couple plans to have four children and would like to have three girls and a boy. What are the chances (probability) their desire.
Random Variables and Probability Distributions Modified from a presentation by Carlos J. Rosas-Anderson.
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
PBG 650 Advanced Plant Breeding
1 Probability and Statistics  What is probability?  What is statistics?
Probability theory 2 Tron Anders Moger September 13th 2006.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Review and Preview This chapter combines the methods of descriptive statistics presented in.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Random Sampling, Point Estimation and Maximum Likelihood.
Theory of Probability Statistics for Business and Economics.
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
 A probability function is a function which assigns probabilities to the values of a random variable.  Individual probability values may be denoted by.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Basic Business Statistics.
 A probability function is a function which assigns probabilities to the values of a random variable.  Individual probability values may be denoted by.
BINOMIALDISTRIBUTION AND ITS APPLICATION. Binomial Distribution  The binomial probability density function –f(x) = n C x p x q n-x for x=0,1,2,3…,n for.
Introduction to Behavioral Statistics Probability, The Binomial Distribution and the Normal Curve.
Lab 3b: Distribution of the mean
Statistics Lectures: Questions for discussion Louis Lyons Imperial College and Oxford CERN Summer Students July
June 11, 2008Stat Lecture 10 - Review1 Midterm review Chapters 1-5 Statistics Lecture 10.
1 2 nd Pre-Lab Quiz 3 rd Pre-Lab Quiz 4 th Pre-Lab Quiz.
1 Practical Statistics for Physicists CERN Latin American School March 2015 Louis Lyons Imperial College and Oxford CMS expt at LHC
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Statistical Data Analysis: Lecture 4 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
Chapter 5 Joint Probability Distributions and Random Samples  Jointly Distributed Random Variables.2 - Expected Values, Covariance, and Correlation.3.
Chapter 5 Probability Distributions 5-1 Overview 5-2 Random Variables 5-3 Binomial Probability Distributions 5-4 Mean, Variance and Standard Deviation.
G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1 Statistical Data Analysis: Lecture 5 1Probability, Bayes’ theorem 2Random variables and.
1 χ 2 and Goodness of Fit & L ikelihood for Parameters Louis Lyons Imperial College and Oxford CERN Summer Students July 2014.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Business Statistics,
1 Practical Statistics for Physicists CERN Summer Students July 2014 Louis Lyons Imperial College and Oxford CMS expt at LHC
R. Kass/Sp07P416/Lecture 71 More on Least Squares Fit (LSQF) In Lec 5, we discussed how we can fit our data points to a linear function (straight line)
THE NORMAL DISTRIBUTION
Biostatistics Class 3 Probability Distributions 2/15/2000.
Stats 242.3(02) Statistical Theory and Methodology.
Practical Statistics for Physicists
Physics 114: Lecture 13 Probability Tests & Linear Fitting
χ2 and Goodness of Fit & Likelihood for Parameters
Probability and Statistics for Particle Physics
Practical Statistics for Physicists
Practical Statistics for Physicists
BAYES and FREQUENTISM: The Return of an Old Controversy
Chapter 5 Joint Probability Distributions and Random Samples
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Lecture 2 Binomial and Poisson Probability Distributions
Presentation transcript:

1 Practical Statistics for Physicists Sardinia Lectures Oct 2008 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC

2 Topics Introduction Learning to love the Error Matrix ******************************************* 1) Do’s and Dont’s with L ikelihoods 2) χ 2 and Goodness of Fit 3) Discovery and p-values 4) Bayes and Frequentism Plenty of time for discussion

3 Some of the questions to be addressed What is coverage, and do we really need it? Should we insist on at least a 5σ effect to claim discovery? How should p-values be combined? If two different models both have respectable χ 2 probabilities, can we reject one in favour of other? Are there different possibilities for quoting the sensitivity of a search? How do upper limits change as the number of observed events becomes smaller than the predicted background? Combine 1 ± 10 and 3 ± 10 to obtain a result of 6 ± 1? What is the Punzi effect and how can it be understood?

4 Books Statistics for Nuclear and Particle Physicists Cambridge University Press, 1986 Available from CUP Errata in these lectures

5 Other Books CDF Statistics Committee BaBar Statistics Working Group

6

7 Introductory remarks Probability and Statistics Random and systematic errors Conditional probability Variance Combining errors Combining experiments Binomial, Poisson and Gaussian distributions

8

9 Parameter Determination Goodness of fit Hypothesis testing THEORY  DATA DATA  THEORY N.B. Parameter values not sensible if goodness of fit is poor/bad

10 Why do we need errors? Affects conclusion about our result e.g. Result / theory = If ± 0.050, data compatible with theory If ± 0.005, data incompatible with theory If ± 0.7, need better experiment Historical experiment at Harwell testing General Relativity

11 Random + Systematic Errors Random/Statistical: Limited accuracy, Poisson counts Spread of answers on repetition (Method of estimating) Systematics: May cause shift, but not spread e.g. Pendulum g = 4π 2 L/τ, τ = T/n Statistical errors: T, L Systematics: T, L Calibrate: Systematic  Statistical More systematics: Formula for undamped, small amplitude, rigid, simple pendulum Might want to correct to g at sea level: Different correction formulae Ratio of g at different locations: Possible systematics might cancel. Correlations relevant

12 Presenting result Quote result as g ± σ stat ± σ syst Or combine errors in quadrature  g ± σ Other extreme: Show all systematic contributions separately Useful for assessing correlations with other measurements Needed for using: improved outside information, combining results using measurements to calculate something else.

13 Bayes’ Theorem

14

15 Combining errors z = x - y δz = δx – δy [1] Why σ z 2 = σ x 2 + σ y 2 ? [2]

16 Combining errors z = x - y δz = δx – δy [1] Why σ z 2 = σ x 2 + σ y 2 ? [2] 1)[1] is for specific δx, δy Could be so on average ? N.B. Mneumonic, not proof 2) σ z 2 = δz 2 = δx 2 + δy 2 – 2 δx δy = σ x 2 + σ y 2 provided…………..

17 3) Averaging is good for you: N measurements x i ± σ [1] x i ± σ or [2] x i ± σ/ √N ? 4) Tossing a coin: Score 0 for tails, 2 for heads (1 ± 1 ) After 100 tosses, [1] 100 ± 100 or [2] 100 ± 10 ? Prob(0 or 200) = (1/2) 99 ~ Compare age of Universe ~ seconds

18 Rules for different functions 1)Linear: z = k 1 x 1 + k 2 x 2 + ……. σ z = k 1 σ 1 & k 2 σ 2 & means “combine in quadrature” N.B. Fractional errors NOT relevant e.g. z = x – y z = your height x = position of head wrt moon y = position of feet wrt moon x and y measured to 0.1% z could be -30 miles

19 Rules for different functions 2) Products and quotients z = x α y β ……. σ z /z = α σ x /x & β σ y /y Useful for x 2, xy, x/y,…….

20 3) Anything else: z = z(x 1, x 2, …..) σ z = ∂z/∂x 1 σ 1 & ∂z/∂x 2 σ 2 & ……. OR numerically: z 0 = z(x 1, x 2, x 3 ….) z 1 = f(x 1 +σ 1, x 2, x 3 ….) z 2 = f(x 1, x 2 + σ 2, x 3 ….) σ z = (z 1 -z 0 ) & (z 2 -z 0 ) & …. N.B. All formulae approximate (except 1)) – assumes small errors

21 Introductory remarks Probability and Statistics Random and systematic errors Conditional probability Variance Combining errors Combining experiments Binomial, Poisson and Gaussian distributions

22

23 To consider…… Is it possible to combine 1 ± 10 and 2 ± 9 to get a best combined value of 6 ± 1 ? Answer later.

24 Difference between averaging and adding Isolated island with conservative inhabitants How many married people ? Number of married men = 100 ± 5 K Number of married women = 80 ± 30 K Total = 180 ± 30 K Wtd average = 99 ± 5 K CONTRAST Total = 198 ± 10 K GENERAL POINT: Adding (uncontroversial) theoretical input can improve precision of answer Compare “kinematic fitting”

25 Binomial Distribution Fixed N independent trials, each with same prob of success p What is prob of s successes? e.g. Throw dice 100 times. Success = ‘6’. What is prob of 0, 1,…. 49, 50, 51,… 99, 100 successes? Effic of track reconstrn = 98%. For 500 tracks, prob that 490, 491, , 500 reconstructed. Ang dist is cos θ? Prob of 52/70 events with cosθ > 0 ? (More interesting is statistics question)

26 P s = N! p s (1-p) N-s, as is obvious (N-s)! s! Expected number of successes = ΣnP n = Np, as is obvious Variance of no. of successes = Np(1-p) Variance ~ Np, for p~0 ~ N(1-p) for p~1 NOT Np in general. NOT n ±√n e.g. 100 trials, 99 successes, NOT 99 ± 10

27 Statistics: Estimate p and σ p from s (and N) p = s/N σ p 2 = 1/N s/N (1 – s/N) If s = 0, p = 0 ± 0 ? If s = 1, p = 1.0 ± 0 ? Limiting cases: ● p = const, N  ∞: Binomial  Gaussian μ = Np, σ 2 = Np(1-p) ● N  ∞, p  0, Np = const: Bin  Poisson μ = Np, σ 2 = Np {N.B. Gaussian continuous and extends to -∞}

28

29 Poisson Distribution Prob of n independent events occurring in time t when rate is r (constant) e.g. events in bin of histogram NOT Radioactive decay for t ~ τ Limit of Binomial (N  ∞, p  0, Np  μ) P n = e -r t (r t) n /n! = e - μ μ n /n! (μ = r t) = r t = μ (No surprise!) σ 2 n = μ “n ±√n” BEWARE 0 ± 0 ? μ  ∞: Poisson  Gaussian, with mean = μ, variance =μ Important for χ 2

30 For your thought Poisson P n = e - μ μ n /n! P 0 = e – μ P 1 = μ e –μ P 2 = μ 2 /2 e -μ For small μ, P 1 ~ μ, P 2 ~ μ 2 /2 If probability of 1 rare event ~ μ, why isn’t probability of 2 events ~ μ 2 ?

31 P 2 = μ 2 e -μ or P 2 = μ 2 /2 e -μ ? 1) P n = e -μ μ n /n! sums to unity 2) n! comes from corresponding Binomial factor N!/{s!(N-s)!} 3) If first event occurs at t 1 with prob μ, average prob of second event in t-t 1 is μ/2. (Identical events) 4) Cow kicks and horse kicks, each producing scream. Find prob of 2 screams in time interval t, by P 2 and P 2 2c, 0h (c 2 e -c ) (e -h ) (½ c 2 e -c ) (e -h ) 1c,1h (ce -c ) (he -h ) (ce -c ) (he -h ) 0c,2h (e -c ) (h 2 e -h ) (e -c ) (½ h 2 e -h ) Sum (c 2 + hc + h 2 ) e -(c+h) ½(c 2 + 2hc + h 2 ) e -(c+h) 2 screams (c+h) 2 e -(c+h) ½(c+h) 2 e -(c+h) Wrong OK

32

33 Relation between Poisson and Binomial PeopleMaleFemale PatientsCuredRemain ill Decaying nuclei ForwardsBackwards Cosmic raysProtonsOther particles N people in lecture, m males and f females (N = m + f ) Assume these are representative of basic rates : ν people νp males ν(1-p) females Probability of observing N people = P Poisson = e –ν ν N /N! Prob of given male/female division = P Binom = p m (1-p) f Prob of N people, m male and f female = P Poisson P Binom = e –νp ν m p m * e -ν(1-p) ν f (1-p) f m! f ! = Poisson prob for males * Poisson prob for females

34 Gaussian or Normal

36 Relevant for Goodness of Fit

37

38 Gaussian = N (r, 0, 1) Breit Wigner = 1/{π * (r 2 + 1)}

39

40 Learning to love the Error Matrix Introduction via 2-D Gaussian Understanding covariance Using the error matrix Combining correlated measurements Estimating the error matrix

41 Correlations Basic issue: For 1 parameter, quote value and error For 2 (or more) parameters, (e.g. gradient and intercept of straight line fit) quote values + errors + correlations Just as the concept of variance for single variable is more general than Gaussian distribution, so correlation in more variables does not require multi-dim Gaussian But more simple to introduce concept this way

42 Element E ij - Diagonal E ij = variances Off-diagonal E ij = covariances

43

44

45

46

47

48

49

50

51

52

53

54 Small error Example: Chi-sq Lecture x best outside x 1  x 2 y best outside y 1  y 2

55 a b x y

56

57

58

59 Conclusion Error matrix formalism makes life easy when correlations are relevant

60 Next time: L ikelihoods What it is How it works: Resonance Error estimates Detailed example: Lifetime Several Parameters Extended maximum L Do’s and Dont’s with L 