Probability Distributions and Monte Carlo Techniques

Slides:



Advertisements
Similar presentations
Random Variables ECE460 Spring, 2012.
Advertisements

Outline input analysis input analyzer of ARENA parameter estimation
Statistics review of basic probability and statistics.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
Review of Basic Probability and Statistics
Chapter 4 Discrete Random Variables and Probability Distributions
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1 Statistical Data Analysis: Lecture 2 1Probability, Bayes’ theorem 2Random variables and.
Statistics.
Probability and Probability Distributions
3-1 Introduction Experiment Random Random experiment.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Introduction to Statistics − Day 2
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Lecture II-2: Probability Review
Modern Navigation Thomas Herring
Chapter 21 Random Variables Discrete: Bernoulli, Binomial, Geometric, Poisson Continuous: Uniform, Exponential, Gamma, Normal Expectation & Variance, Joint.
Standard error of estimate & Confidence interval.
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
Short Resume of Statistical Terms Fall 2013 By Yaohang Li, Ph.D.
1 Probability and Statistics  What is probability?  What is statistics?
Probability theory 2 Tron Anders Moger September 13th 2006.
PROBABILITY & STATISTICAL INFERENCE LECTURE 3 MSc in Computing (Data Analytics)
Statistics for Engineer Week II and Week III: Random Variables and Probability Distribution.
Noise & Uncertainty ASTR 3010 Lecture 7 Chapter 2.
Statistics for Data Miners: Part I (continued) S.T. Balke.
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Theory of Probability Statistics for Business and Economics.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 1 – Slide 1 of 34 Chapter 11 Section 1 Random Variables.
LECTURER PROF.Dr. DEMIR BAYKA AUTOMOTIVE ENGINEERING LABORATORY I.
Lab 3b: Distribution of the mean
1 Introduction to Statistical Methods for High Energy Physics Glen Cowan 2006 CERN Summer Student Lectures CERN Summer Student Lectures on Statistics Glen.
G. Cowan Lectures on Statistical Data Analysis Lecture 1 page 1 Lectures on Statistical Data Analysis London Postgraduate Lectures on Particle Physics;
Statistical Data Analysis and Simulation Jorge Andre Swieca School Campos do Jordão, January,2003 João R. T. de Mello Neto.
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
Uniovi1 Some statistics books, papers, etc. G. Cowan, Statistical Data Analysis, Clarendon, Oxford, 1998 see also R.J. Barlow,
IE 300, Fall 2012 Richard Sowers IESE. 8/30/2012 Goals: Rules of Probability Counting Equally likely Some examples.
Computer simulation Sep. 9, QUIZ 2 Determine whether the following experiments have discrete or continuous out comes A fair die is tossed and the.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
NON-LINEAR REGRESSION Introduction Section 0 Lecture 1 Slide 1 Lecture 6 Slide 1 INTRODUCTION TO Modern Physics PHYX 2710 Fall 2004 Intermediate 3870 Fall.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Statistical Data Analysis: Lecture 4 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1 Statistical Data Analysis: Lecture 5 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Introduction to Probability - III John Rundle Econophysics PHYS 250
MECH 373 Instrumentation and Measurements
Lecture 1 Probability and Statistics
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005
Parameter Estimation and Hypothesis Testing
Probability and Statistics for Particle Physics
Introduction to Statistics − Day 2
Parameter Estimation and Fitting to Data
Probability Distributions and Monte Carlo Techniques
Ex1: Event Generation (Binomial Distribution)
Appendix A: Probability Theory
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Introduction, class rules, error analysis Julia Velkovska
Introduction to Instrumentation Engineering
Discrete Event Simulation - 4
Lecture 2 – Monte Carlo method in finance
Lecture 3 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Counting Statistics and Error Prediction
2.0 Probability Concepts definitions: randomness, parent population, random variable, probability, statistical independence, probability of multiple events,
Computing and Statistical Data Analysis / Stat 7
Introduction to Statistics − Day 4
Computing and Statistical Data Analysis / Stat 10
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Presentation transcript:

Probability Distributions and Monte Carlo Techniques Common probability distributions Binomial, Poisson, Gaussian, Exponential Sums and differences Characteristic functions Central Limits Theorem Generation of random distributions Elton S. Smith Jefferson Lab Con ayuda de Eduardo Medinaceli y Cristian Peña

Selected references CDF Statistics Committee Particle Data Group http://pdg.lbl.gov/2009/reviews/contents_sports.html CERN Summer Student Lecture Programme Course, August 2009 Introduction to Statistics (Four Lectures) by G. Cowan (University of London) http://indico.cern.ch/conferenceDisplay.py?confId=57559 CDF Statistics Committee http://www-cdf.fnal.gov/physics/statistics/ F. James, Statistical Methods in Experimental Physics, 2nd ed. 2006 (W.T. Eadie et al.1971) R. Bevington, Data Reduction and Error Analysis for the Physical Sciences (2002) H. Cramer, Mathematical Methods of Statistics (1946)

Dictionary / Diccionario systematic sistematico spin espin background trasfondo scintillator contador de centelleo random aleatorio (al azar) scaler escalar histogram histograma degrees of freedom grados de libertad signal sen~al power potencia maximum likelihood maxima verosimilitud test statistic prueba estadistica least squares minimos cuadrados goodness bondad/fidelidad chi-square ji-cuadrado estimation estima confidence intervals and limits intervalos de confianza y limites significance significancia frequentist frecuentista? conditional p p condicional nuissance parameters parametros molestosos? (A given B) (A dado B) outline indice fit ajuste (encaje) asymmetry asimetria parent distribution distr del patron variance varianza signal-to-background fraction sen~al/trasfondo root-mean-square raiz quadrada media (rms) sample muestra biased sesgo flip-flopping indeciso counting experiments experimentos de conteo   multiple scattering dispersion multiple correlations correlaciones weight pesadas

We quantify all these uncertainties using the concept of PROBABILITY Theory of quantum mechanics is not deterministic Present even for “perfect” measurements Example: Lifetime of a radioactive nucleus Random measurement uncertainties or “errors” Present even without quantum effects Example: limited accuracy of measurement Things we know in principle, but don’t in practice Example: uncontrolled parameters during measurement We quantify all these uncertainties using the concept of PROBABILITY

Interpretation of probability Relative frequency (classical) If ‘A’ is the outcome of a repeatable experiment Subjective probability (Bayesian) If ‘A’ is a hypothesis (statement that is true or false) In particle physics the classical or ‘frequentist’ interpretation is most common, but the Bayesian approach can be useful for non-repeatable phenomena, e.g. probability the Higgs boson exists

Uniform random variable probability probability density distribution (pdf) P(x) = f(x)dx a f(x) x a Particle counter detects a particle if it hits anywhere over its sensitive length ‘a’. Absent any knowledge about the source of particles, we would assign a uniform probability distribution to the position x, of the particle whenever the counter registers a hit.

Experimental uncertainties Assume a measurement is made of a quantity m. Let x be the value of a single measurement and this measurement has an uncertainty s. Now, assume this measurement is repeated many times, and assume that each measurement is made independently of other measurements. In that case, the measurement x can be considered a random variable, with a probability distribution that will be centered on m, but with a value that differs from m by an amount that is approximately s. But what is this probability distribution?

Distribution of measurements Raw asymmetries for 1999 HAPPEX running period, in ppm, broken down by data set. Circles are for left spectrometer, triangles for right. Dashed line is the average for the entire run. Aniol Phys Rev C69 (2004) 065501

Distribution of experimental measurements Run asymmetries for 1999 HAPPEX running period, with mean subtracted off and normalized by statistical error Aniol Phys Rev C69 (2004) 065501

Gaussian (Normal) distribution Typical of experimental random uncertainties Cumulative distribution function (units ~ 1/s) s (dimensionless) Named “standard Gaussian” when m=0 and s=1 m

Moments: Defined for all distributions Expectation value of x nth moment of a random variable x nth central moment of a random variable x mean variance Same units! root-mean-square

Example: uniform distribution f(x) x a s = 0.29a m = a/2 s = a/√12 For a Gaussian distribution m = “m” s = “s”

Histograms: representation of pdfs using data normalization

Discrete distributions: binomial Biased coin toss N trials probability of success p probability of failure (1-p) 0 ≤ p ≤ 1 parameters random variable mean variance

Binomial distribution examples

Discrete distributions: Poisson Parent distribution for counting experiments Limiting case of binomial distribution Limit for p 0 Mean successes Np m n ≥ 0 parameter random variable mean s = √m variance s/m = 1/√m

Poisson distribution examples The mean m can be any positive real number For small values of m, there is a significant probability n=0 The distribution approaches Gaussian for values of m ≥ 10

Counting experiments The Poisson approximation is appropriate for counting experiments where the data represent the number of items observed per unit time. Example: 10 nA beam (1011 particles/s) produce 104 interactions/s, i.e p ~ 10-7 Here m = 104 (1 s of data), and s = √m = 102 The uncertainty s is called the statistical error or uncertainty Note also: 10% chance to get zero events is

Another continuous pdf: exponential Proper decay time for unstable particle Population growth parameter random variable mean Lack of memory: f(t-t0 | t ≥ t0) = f(t) variance s = m

Characteristic functions f1(u) <--> f1(x) f2(u) <--> f2(y) Form a new random variable z = ax + by For independent variables x and y, f(x,y) = f1(x) f2(y) Then f(u) = f1(au)f2(bu) <--> g(z) Allows computation of pdfs for sums and differences of random variables with known distributions

Example: rules for sums Gaussian sum of G(m1,s1) and G(m2,s2) gives G(m=m1+m2,s=√s12+s22) Poisson sum of P(m1) and P(m2) gives P(m=m1+m2) Note: Difference also works for G, not P!

Central Limit Theorem Any random variable that is the sum of many small contributions (with arbitrary pdfs) has a Gaussian pdf. xi are (independent) random variables with means mi and variances si2 Variable of interest is For large n: Example: Multiple scattering distributions are approximately Gaussian because they result from the sum of many individual scatters

Correlations Let x,y be two random variables with a joint probability distribution f(x,y) Marginal probability distributions (integrate over y) Conditional probability distributions (fix y=y0) Averages

Covariance

Examples

Propagation of errors uncertainties Physical quantities of interest are often combinations of more than one measurements sums m = mx + my s2 = sx2 + sy2 + 2sxsy rxy (-1 ≤ rxy ≤ 1) products m = mx my s2/(x2y2) = sx2/x2 + sy2/y2 + 2sxsy rxy/(xy) If x and y are independent, then rxy = 0

Error in the average If xi are independent measurements of a quantity with mean m and distributed with the same sigma s Then the average of xi average m s of average s/√N

Monte Carlo Method (p,q,f) Numerical technique for computing the distribution of particle interactions Each interaction is assumed to be governed by the conservation of energy and momentum and the probabilistic laws of quantum mechanics Perfect modeling of the interaction will require the use of correct probability distribution for each variable (e.g. momentum and angles of each particle) including all correlations, although much can be learned with reasonable approximations Sequences of random numbers are used to generate Monte Carlo or “simulated” data to be compared to actual measurements. Differences between the true and simulated data can be used to improve understanding the process under study (p,q,f)

Random number generators Computer generates “pseudo-random” numbers, which are deterministic by depend on an input “seed” (often Unix time used) Many interactions are simulated Each interaction requires the generation a series of random numbers (p, q and f in the present example) Poor random number generators will repeat themselves and/or have periodic correlations (e.g. between the first and third number generated) Very good algorithms are available, e.g. TRandom3 from ROOT have a periodicity of 219937-1.

The acceptance-rejection method

The transform method Uses every single random number produced to generate the distribution f(x) of interest Integrate distribution Normalize distribution Generate random number r on [0,1]. Then compute x = FN-1(r). The variable x will be distributed according to the function f(x).

Example of the transform method Many other examples in the Review of Particle Physics

Summary of first lecture Defined probability Described various common probability distributions Demonstrated how new distributions could be generated from combinations of known distributions Described the Monte Carlo method for numerically simulating physical processes. Next lecture will focus on interpreting data to extract information about the parent distributions, namely statistics.

Backup slides

Window pair asymmetries Window pair asymmetries for 1999 HAPPEX running period, normalized by square root of beam intensity, with mean value subtracted off, in ppm. Aniol Phys Rev C69 (2004) 065501