1 Practical Statistics for Physicists Stockholm Lectures Sept 2008 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
DO’S AND DONT’S WITH LIKELIHOODS Louis Lyons Oxford (CDF)
1 BAYES and FREQUENTISM: The Return of an Old Controversy Louis Lyons Oxford University LBL, January 2008.
Lecture 6 CS5661 Pairwise Sequence Analysis-V Relatedness –“Not just important, but everything” Modeling Alignment Scores –Coin Tosses –Unit Distributions.
Psyc 235: Introduction to Statistics
ฟังก์ชั่นการแจกแจงความน่าจะเป็น แบบไม่ต่อเนื่อง Discrete Probability Distributions.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
U eatworms.swmed.edu/~leon u
1 χ 2 and Goodness of Fit Louis Lyons IC and Oxford SLAC Lecture 2’, Sept 2008.
1 Do’s and Dont’s with L ikelihoods Louis Lyons Oxford and IC CDF IC, February 2008.
1 Practical Statistics for Physicists Sardinia Lectures Oct 2008 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC
1 χ 2 and Goodness of Fit Louis Lyons Oxford Mexico, November 2006 Lecture 3.
Chapter 1 Probability Theory (i) : One Random Variable
1 Practical Statistics for Physicists LBL January 2008 Louis Lyons Oxford
1 Do’s and Dont’s with L ikelihoods Louis Lyons Oxford CDF Mexico, November 2006.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
1 Practical Statistics for Physicists Dresden March 2010 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC
1 Do’s and Dont’s with L ikelihoods Louis Lyons Oxford CDF CERN, October 2006.
Introduction to experimental errors
1 Practical Statistics for Physicists SLUO Lectures February 2007 Louis Lyons Oxford
1 Practical Statistics for Particle Physicists CERN Academic Lectures October 2006 Louis Lyons Oxford
1 Do’s and Dont’s with L ikelihoods Louis Lyons Oxford CDF Manchester, 16 th November 2005.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Overview Parameters and Statistics Probabilities The Binomial Probability Test.
K. Desch – Statistical methods of data analysis SS10 3. Distributions 3.1 Binomial distribution Error of an efficiency Estimator for the efficiency (when.
Discrete Probability Distributions
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Lecture 6: Descriptive Statistics: Probability, Distribution, Univariate Data.
TOPLHCWG. Introduction The ATLAS+CMS combination of single-top production cross-section measurements in the t channel was performed using the BLUE (Best.
Problem A newly married couple plans to have four children and would like to have three girls and a boy. What are the chances (probability) their desire.
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Chapter 5 Sampling Distributions
PBG 650 Advanced Plant Breeding
1 Probability and Statistics  What is probability?  What is statistics?
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Review and Preview This chapter combines the methods of descriptive statistics presented in.
Random Sampling, Point Estimation and Maximum Likelihood.
Theory of Probability Statistics for Business and Economics.
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
 A probability function is a function which assigns probabilities to the values of a random variable.  Individual probability values may be denoted by.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Basic Business Statistics.
 A probability function is a function which assigns probabilities to the values of a random variable.  Individual probability values may be denoted by.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #23.
BINOMIALDISTRIBUTION AND ITS APPLICATION. Binomial Distribution  The binomial probability density function –f(x) = n C x p x q n-x for x=0,1,2,3…,n for.
Be humble in our attribute, be loving and varying in our attitude, that is the way to live in heaven.
Statistics Lectures: Questions for discussion Louis Lyons Imperial College and Oxford CERN Summer Students July
June 11, 2008Stat Lecture 10 - Review1 Midterm review Chapters 1-5 Statistics Lecture 10.
1 Practical Statistics for Physicists CERN Latin American School March 2015 Louis Lyons Imperial College and Oxford CMS expt at LHC
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
Chapter 5 Joint Probability Distributions and Random Samples  Jointly Distributed Random Variables.2 - Expected Values, Covariance, and Correlation.3.
R.Kass/F02 P416 Lecture 1 1 Lecture 1 Probability and Statistics Introduction: l The understanding of many physical phenomena depend on statistical and.
Hypothesis Testing. Suppose we believe the average systolic blood pressure of healthy adults is normally distributed with mean μ = 120 and variance σ.
1 χ 2 and Goodness of Fit & L ikelihood for Parameters Louis Lyons Imperial College and Oxford CERN Summer Students July 2014.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Business Statistics,
1 Practical Statistics for Physicists CERN Summer Students July 2014 Louis Lyons Imperial College and Oxford CMS expt at LHC
R. Kass/Sp07P416/Lecture 71 More on Least Squares Fit (LSQF) In Lec 5, we discussed how we can fit our data points to a linear function (straight line)
Stats 242.3(02) Statistical Theory and Methodology.
Practical Statistics for Physicists
Statistical Modelling
Lecture 1 Probability and Statistics
χ2 and Goodness of Fit & Likelihood for Parameters
Probability and Statistics for Particle Physics
Practical Statistics for Physicists
Practical Statistics for Physicists
BAYES and FREQUENTISM: The Return of an Old Controversy
Chapter 5 Joint Probability Distributions and Random Samples
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Lecture 2 Binomial and Poisson Probability Distributions
Presentation transcript:

1 Practical Statistics for Physicists Stockholm Lectures Sept 2008 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC

2 Topics 1) Introduction Learning to love the Error Matrix 2) Do’s and Dont’s with L ikelihoods χ 2 and Goodness of Fit 3) Discovery and p-values 4) Bayes and Frequentism Plenty of time for discussion

3 Some of the questions to be addressed What is coverage, and do we really need it? Should we insist on at least a 5σ effect to claim discovery? How should p-values be combined? If two different models both have respectable χ 2 probabilities, can we reject one in favour of other? Are there different possibilities for quoting the sensitivity of a search? How do upper limits change as the number of observed events becomes smaller than the predicted background? Combine 1 ± 10 and 3 ± 10 to obtain a result of 6 ± 1? What is the Punzi effect and how can it be understood?

4 Books Statistics for Nuclear and Particle Physicists Cambridge University Press, 1986 Available from CUP Errata in these lectures

5 Other Books CDF Statistics Committee BaBar Statistics Working Group

6

7 Introductory remarks Probability and Statistics Random and systematic errors Conditional probability Variance Combining errors Combining experiments Binomial, Poisson and Gaussian distributions

8

9 Parameter Determination Goodness of fit Hypothesis testing THEORY  DATA DATA  THEORY N.B. Parameter values not sensible if goodness of fit is poor/bad

10 Why do we need errors? Affects conclusion about our result e.g. Result / theory = If ± 0.050, data compatible with theory If ± 0.005, data incompatible with theory If ± 0.7, need better experiment Historical experiment at Harwell testing General Relativity

11 Random + Systematic Errors Random/Statistical: Limited accuracy, Poisson counts Spread of answers on repetition (Method of estimating) Systematics: May cause shift, but not spread e.g. Pendulum g = 4π 2 L/τ, τ = T/n Statistical errors: T, L Systematics: T, L Calibrate: Systematic  Statistical More systematics: Formula for undamped, small amplitude, rigid, simple pendulum Might want to correct to g at sea level: Different correction formulae Ratio of g at different locations: Possible systematics might cancel. Correlations relevant

12 Presenting result Quote result as g ± σ stat ± σ syst Or combine errors in quadrature  g ± σ Other extreme: Show all systematic contributions separately Useful for assessing correlations with other measurements Needed for using: improved outside information, combining results using measurements to calculate something else.

13

14

15 Combining errors z = x - y δz = δx – δy [1] Why σ z 2 = σ x 2 + σ y 2 ? [2]

16 Combining errors z = x - y δz = δx – δy [1] Why σ z 2 = σ x 2 + σ y 2 ? [2] 1)[1] is for specific δx, δy Could be so on average ? N.B. Mneumonic, not proof 2) σ z 2 = δz 2 = δx 2 + δy 2 – 2 δx δy = σ x 2 + σ y 2 provided…………..

17 3) Averaging is good for you: N measurements x i ± σ [1] x i ± σ or [2] x i ± σ/ √N ? 4) Tossing a coin: Score 0 for tails, 2 for heads (1 ± 1 ) After 100 tosses, [1] 100 ± 100 or [2] 100 ± 10 ? Prob(0 or 200) = (1/2) 99 ~ Compare age of Universe ~ seconds

18 Rules for different functions 1)Linear: z = k 1 x 1 + k 2 x 2 + ……. σ z = k 1 σ 1 & k 2 σ 2 & means “combine in quadrature” 2) Products and quotients z = x α y β ……. σ z /z = α σ x /x & β σ y /y

19 3) Anything else: z = z(x 1, x 2, …..) σ z = ∂z/∂x 1 σ 1 & ∂z/∂x 2 σ 2 & ……. OR numerically: z 0 = z(x 1, x 2, x 3 ….) Z 1 = f(x 1 +σ 1, x 2, x 3 ….) Z 2 = f(x 1, x 2 + σ 2, x 3 ….) σ z = (z 1 -z 0 ) & (z 2 -z 0 ) & ….

20

21 To consider…… Is it possible to combine 1 ± 10 and 2 ± 9 to get a best combined value of 6 ± 1 ? Answer later.

22 Difference between averaging and adding Isolated island with conservative inhabitants How many married people ? Number of married men = 100 ± 5 K Number of married women = 80 ± 30 K Total = 180 ± 30 K Wtd average = 99 ± 5 K CONTRAST Total = 198 ± 10 K GENERAL POINT: Adding (uncontroversial) theoretical input can improve precision of answer Compare “kinematic fitting”

23 Binomial Distribution Fixed N independent trials, each with same prob of success p What is prob of s successes? e.g. Throw dice 100 times. Success = ‘6’. What is prob of 0, 1,…. 49, 50, 51,… 99, 100 successes? Effic of track reconstrn = 98%. For 500 tracks, prob that 490, 491, , 500 reconstructed. Ang dist is cos θ? Prob of 52/70 events with cosθ > 0 ? (More interesting is statistics question)

24 P s = N! p s (1-p) N-s, as is obvious (N-s)! s! Expected number of successes = ΣnP n = Np, as is obvious Variance of no. of successes = Np(1-p) Variance ~ Np, for p~0 ~ N(1-p) for p~1 NOT Np in general. NOT n ±√n

25 Statistics: Estimate p and σ p from s (and N) p = s/N σ p 2 = 1/N s/N (1 – s/N) If s = 0, p = 0 ± 0 ? If s = 1, p = 1.0 ± 0 ? Limiting cases: ● p = const, N  ∞: Binomial  Gaussian μ = Np, σ 2 = Np(1-p) ● N  ∞, p  0, Np = const: Bin  Poisson μ = Np, σ 2 = Np {N.B. Gaussian continuous and extends to -∞}

26

27 Poisson Distribution Prob of n independent events occurring in time t when rate is r (constant) e.g. events in bin of histogram NOT Radioactive decay for t ~ τ Limit of Binomial (N  ∞, p  0, Np  μ) P n = e -r t (r t) n /n! = e - μ μ n /n! (μ = r t) = r t = μ (No surprise!) σ 2 n = μ “n ±√n” BEWARE 0 ± 0 ? μ  ∞: Poisson  Gaussian, with mean = μ, variance =μ Important for χ 2

28 For your thought Poisson P n = e - μ μ n /n! P 0 = e – μ P 1 = μ e –μ P 2 = μ 2 /2 e -μ For small μ, P 1 ~ μ, P 2 ~ μ 2 /2 If probability of 1 rare event ~ μ, Why isn’t probability of 2 events ~ μ 2 ?

29

30 Relation between Poisson and Binomial PeopleMaleFemale PatientsCuredRemain ill Decaying nuclei ForwardsBackwards Cosmic raysProtonsOther particles N people in lecture, m males and f females (N = m + f ) Assume these are representative of basic rates : ν people νp males ν(1-p) females Probability of observing N people = P Poisson = e –ν ν N /N! Prob of given male/female division = P Binom = p m (1-p) f Prob of N people, m male and f female = P Poisson P Binom = e –νp ν m p m * e -ν(1-p) ν f (1-p) f m! f ! = Poisson prob for males * Poisson prob for females

31 Gaussian or Normal

32

33 Relevant for Goodness of Fit

34

35 Gaussian = N (r, 0, 1) Breit Wigner = 1/{π * (r 2 + 1)}

36

37 Learning to love the Error Matrix Introduction via 2-D Gaussian Understanding covariance Using the error matrix Combining correlated measurements Estimating the error matrix

38 Element E ij - Diagonal E ij = variances Off-diagonal E ij = covariances

39

40

41

42

43

44

45

46

47

48

49

50 Small error Example: Lecture 2 x best outside x 1  x 2 y best outside y 1  y 2

51 a b x y

52

53

54

55 Conclusion Error matrix formalism makes life easy when correlations are relevant

56 Next time: L ikelihoods What it is How it works: Resonance Error estimates Detailed example: Lifetime Several Parameters Extended maximum L Do’s and Dont’s with L 