Practical Statistics for Physicists

Slides:



Advertisements
Similar presentations
Topic 12: Multiple Linear Regression
Advertisements

Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Prediction, Correlation, and Lack of Fit in Regression (§11. 4, 11
1 Practical Statistics for Physicists Sardinia Lectures Oct 2008 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC
1 χ 2 and Goodness of Fit Louis Lyons Oxford Mexico, November 2006 Lecture 3.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
1 Practical Statistics for Physicists LBL January 2008 Louis Lyons Oxford
1 Def: Let and be random variables of the discrete type with the joint p.m.f. on the space S. (1) is called the mean of (2) is called the variance of (3)
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Spatial autoregressive methods Nr245 Austin Troy Based on Spatial Analysis by Fortin and Dale, Chapter 5.
1 Practical Statistics for Physicists Dresden March 2010 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC
Level 1 Laboratories University of Surrey, Physics Dept, Level 1 Labs, Oct 2007 Handling & Propagation of Errors : A simple approach 1.
1 Practical Statistics for Physicists SLUO Lectures February 2007 Louis Lyons Oxford
1 Practical Statistics for Particle Physicists CERN Academic Lectures October 2006 Louis Lyons Oxford
K. Desch – Statistical methods of data analysis SS10 2. Probability 2.3 Joint p.d.f.´s of several random variables Examples: Experiment yields several.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
1 Practical Statistics for Physicists Stockholm Lectures Sept 2008 Louis Lyons Imperial College and Oxford CDF experiment at FNAL CMS expt at LHC
Lecture II-2: Probability Review
Modern Navigation Thomas Herring
TOPLHCWG. Introduction The ATLAS+CMS combination of single-top production cross-section measurements in the t channel was performed using the BLUE (Best.
Introduction l Example: Suppose we measure the current (I) and resistance (R) of a resistor. u Ohm's law relates V and I: V = IR u If we know the uncertainties.
880.P20 Winter 2006 Richard Kass Propagation of Errors Suppose we measure the branching fraction BR(Higgs  +  - ) using the number of produced Higgs.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
Short Resume of Statistical Terms Fall 2013 By Yaohang Li, Ph.D.
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
MTH 161: Introduction To Statistics
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
GG 313 Geological Data Analysis Lecture 13 Solution of Simultaneous Equations October 4, 2005.
LECTURE 3: ANALYSIS OF EXPERIMENTAL DATA
1 Practical Statistics for Physicists CERN Latin American School March 2015 Louis Lyons Imperial College and Oxford CMS expt at LHC
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
The Simple Linear Regression Model: Specification and Estimation ECON 4550 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s.
LESSON 6: REGRESSION 2/21/12 EDUC 502: Introduction to Statistics.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Chapter 3: Uncertainty "variation arises in data generated by a model" "how to transform knowledge of this variation into statements about the uncertainty.
1 Practical Statistics for Physicists CERN Summer Students July 2014 Louis Lyons Imperial College and Oxford CMS expt at LHC
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Louis Lyons Imperial College
Practical Statistics for Physicists
Notes on Weighted Least Squares Straight line Fit Passing Through The Origin Amarjeet Bhullar November 14, 2008.
Chapter 11: Linear Regression and Correlation
χ2 and Goodness of Fit & Likelihood for Parameters
(5) Notes on the Least Squares Estimate
Practical Statistics for Physicists
B&A ; and REGRESSION - ANCOVA B&A ; and
Regression Diagnostics
CH 5: Multivariate Methods
Probablity Density Functions
Statistics in Applied Science and Technology
Correlation and Simple Linear Regression
Chapter 6: MULTIPLE REGRESSION ANALYSIS
Probability & Statistics Probability Theory Mathematical Probability Models Event Relationships Distributions of Random Variables Continuous Random.
Modelling data and curve fitting
The regression model in matrix form
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Correlation and Simple Linear Regression
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005
Trip Generation II Meeghat Habibian Transportation Demand Analysis
Chapter 3 Statistical Concepts.
5.4 General Linear Least-Squares
Simple Linear Regression and Correlation
Computing and Statistical Data Analysis / Stat 7
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
#21 Marginalize vs. Condition Uninteresting Fitted Parameters
Principal Component Analysis
Moments of Random Variables
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Practical Statistics for Physicists Louis Lyons Imperial College CMS expt at LHC l.lyons@physics.ox.ac.uk Islamabad Aug 2017

Combining Experiments N.B. Better to combine data! * BEWARE 100±10 2±1? 1±1 or 50.5±5?

For your thought Poisson Pn = e-μ μn/n! P0 = e–μ P1 = μ e–μ P2 = μ2 /2 e-μ For small μ, P1 ~ μ, P2 ~ μ2/2 If probability of 1 rare event ~ μ, why isn’t probability of 2 events ~ μ2 ?

Correlations Basic issue: For 1 parameter, quote value and uncertainty For 2 (or more) parameters, (e.g. gradient and intercept of straight line fit) quote values + uncertainties + correlations Just as the concept of variance for single variable is more general than Gaussian distribution, so correlation in more variables does not require multi-dim Gaussian But more simple to introduce concept this way

Learning to love the Covariance Matrix Introduction via 2-D Gaussian Understanding covariance Using the covariance matrix Combining correlated measurements Estimating the covariance matrix

Reminder of 1-D Gaussian or Normal y = 1 exp{-(x-µ)2/(22)} 2  Reminder of 1-D Gaussian or Normal Significance of σ i) RMS of Gaussian = σ (hence factor of 2 in definition of Gaussian) ii) At x = μ±σ, y = ymax/√e ~0.606 ymax (i.e. σ = half-width at ‘half’-height) iii) Fractional area within μ±σ = 68% iv) Height at max = 1/(σ√2π)

Element Eij - <(xi – xi) (xj – xj)> Diagonal Eij = variances Off-diagonal Eij = covariances

Towards the Covariance Matrix x and y uncorrelated P(x,y} = G1(x) G2(y) G1(x) = 1/(√2πσx) exp{-x2/2σx2} G1(x) = 1/(√2πσy) exp{-y2/2σy2} P(x,y) = 1/(2πσxσy) exp{-0.5(x2/σx2+y2/σy2)}

Inverse Covariance Matrix Covariance Matrix

( ( ) ) 8x2 + 2y2 =1 0.5(13x’2 + 6√3x’y’ + 7y’2) =1 Inverse Covariance 13/2 3√3/2 (1/32)* 7 -3√3/2 3√3/2 7/2 -3√3/2 13 ( ) ( ) Inverse Covariance covariance matrix matrix Correlation coefficient ρ = covariance/σ(x’)σ(y’) = (-3√3/2)/sqrt(7*13) = -0.27 7/32 = (0.468)2 = σ(x`)2 1/6.5 = (0.392)2 1/8 = eigenvalue of covariance matrix = σ(x)2

Covariance matrix

͂ Using the Covariance Matrix σy2 = DED Function of variables y = y(xa, xb) Given covariance matrix for xa, xb, what is σy ? Differentiate, square, average σy2 = DED ͂

(ii) Change of variables xa = xa(pi,pj) xb = xb(pi,pj) e.g Cartesian to polars; or Points in x.y  intercept and gradient of line Given cov matrix for pi,pj, what is cov matrix for xa,xb ? Differentiate, calculate δxaδxb, and average

Ex = TEpT BEWARE! ͂

Example from Particle Physics

Examples of correlated variables

Using the Covariance Matrix

Small error Example: Straight line fitting xbest outside x1  x2 ybest outside y1  y2

b y a x

Uncertainty on Ωdark energy When combining pairs of variables, the uncertainties on the combined parameters can be much smaller than any of the individual uncertainties e.g. Ωdark energy

Estimating the Covariance Matrix

Conclusion Covariance matrix formalism makes life easy when correlations are relevant