Basic Counting StatisticsBasic Counting StatisticsBasic Counting StatisticsBasic Counting Statistics.

Slides:



Advertisements
Similar presentations
Chapter 3 Properties of Random Variables
Advertisements

AP Statistics Course Review.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Sampling Distributions (§ )
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Chapter 7 Sampling and Sampling Distributions
Statistics. Large Systems Macroscopic systems involve large numbers of particles.  Microscopic determinism  Macroscopic phenomena The basis is in mechanics.
Chapter 6 Continuous Random Variables and Probability Distributions
Statistics.
Lec 6, Ch.5, pp90-105: Statistics (Objectives) Understand basic principles of statistics through reading these pages, especially… Know well about the normal.
Some standard univariate probability distributions
K. Desch – Statistical methods of data analysis SS10 3. Distributions 3.1 Binomial distribution Error of an efficiency Estimator for the efficiency (when.
OMS 201 Review. Range The range of a data set is the difference between the largest and smallest data values. It is the simplest measure of dispersion.
Some standard univariate probability distributions
Sampling Distributions & Point Estimation. Questions What is a sampling distribution? What is the standard error? What is the principle of maximum likelihood?
Modern Navigation Thomas Herring
Normal and Sampling Distributions A normal distribution is uniquely determined by its mean, , and variance,  2 The random variable Z = (X-  /  is.
Standard error of estimate & Confidence interval.
Statistical Methods For Engineers ChE 477 (UO Lab) Larry Baxter & Stan Harding Brigham Young University.
Body size distribution of European Collembola Lecture 9 Moments of distributions.
W. Udo Schröder, 2009 Principles Meas 1 Principles of Measurement.
Introduction to Error Analysis
Standard Statistical Distributions Most elementary statistical books provide a survey of commonly used statistical distributions. The reason we study these.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
1 Probability and Statistics  What is probability?  What is statistics?
Some standard univariate probability distributions Characteristic function, moment generating function, cumulant generating functions Discrete distribution.
Noise & Uncertainty ASTR 3010 Lecture 7 Chapter 2.
Topic 5 Statistical inference: point and interval estimate
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
Random Sampling, Point Estimation and Maximum Likelihood.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
Lecture 15: Statistics and Their Distributions, Central Limit Theorem
LECTURER PROF.Dr. DEMIR BAYKA AUTOMOTIVE ENGINEERING LABORATORY I.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
The final exam solutions. Part I, #1, Central limit theorem Let X1,X2, …, Xn be a sequence of i.i.d. random variables each having mean μ and variance.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Radiation Detection and Measurement, JU, 1st Semester, (Saed Dababneh). 1 Radioactive decay is a random process. Fluctuations. Characterization.
Summarizing Risk Analysis Results To quantify the risk of an output variable, 3 properties must be estimated: A measure of central tendency (e.g. µ ) A.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
Beginning Statistics Table of Contents HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2008 by Hawkes Learning Systems/Quant Systems, Inc.
Chapter 5 Sampling Distributions. Introduction Distribution of a Sample Statistic: The probability distribution of a sample statistic obtained from a.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Maximum likelihood estimators Example: Random data X i drawn from a Poisson distribution with unknown  We want to determine  For any assumed value of.
CHAPTER 4 ESTIMATES OF MEAN AND ERRORS. 4.1 METHOD OF LEAST SQUARES I n Chapter 2 we defined the mean  of the parent distribution and noted that the.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Theoretical distributions: the other distributions.
Theoretical distributions: the Normal distribution.
Confidence Intervals Cont.
Modeling and Simulation CS 313
Chapter Eight Estimation.
Probability and Statistics for Particle Physics
Sampling Distributions and Estimation
Physics 114: Lecture 7 Probability Density Functions
Sampling Distributions & Point Estimation
Discrete Event Simulation - 4
Counting Statistics and Error Prediction
2.3 Estimating PDFs and PDF Parameters
9. Data Fitting Fitting Experimental Spectrum
Sampling Distributions (§ )
Introduction to Statistics − Day 4
CHAPTER – 1.2 UNCERTAINTIES IN MEASUREMENTS.
Applied Statistics and Probability for Engineers
CHAPTER – 1.2 UNCERTAINTIES IN MEASUREMENTS.
Presentation transcript:

Basic Counting StatisticsBasic Counting StatisticsBasic Counting StatisticsBasic Counting Statistics

W. Udo Schröder, 2009 Statistics 2 Stochastic Nuclear Observables 2 sources of stochastic observables x in nuclear science: 1) Nuclear phenomena are governed by quantal wave functions and inherent statistics 2) Detection of processes occurs with imperfect efficiency (  < 1) and finite resolution distributing sharp events x 0 over a range in x. Stochastic observables x have a range of values with frequencies determined by probability distribution P(x). characterize by set of moments of P = ∫ x n P (x)dx; n 0, 1, 2,… Normalization = 1. First moment (expectation value) of P: E(x) = = ∫xP (x)dx second central moment = “variance” of P(x):  x 2 = >

Uncertainty and Statistics Nuclear systems: quantal wave functions  i (x,…;t) (x,…;t) = degrees of freedom of system, time. Probability density (e.g., for x, integrate over other d.o.f.) 1 2 Partial probability rates  for disappearance (decay of 1  2) can vary over many orders of magnitude  no certainty  statistics

W. Udo Schröder, 2009 Statistics 4 The Normal Distribution Continuous function or discrete distribution (over bins) Normalized probability 2x2x P(x) x x

W. Udo Schröder, 2009 Statistics 5 Experimental Mean Counts and Variance Measured by ensemble sampling expectation values + uncertainties Sample (Ensemble) = Population instant 236 U (0.25mg) source, count  particles emitted during N = 10 time intervals min).  =?? Slightly different from sample to sample

W. Udo Schröder, 2009 Statistics 6 Sample Statistics x P(x) Assume true population distribution for variable x with true (“population”) mean pop = 5.0, x =1.0 Mean arithmetic sample average = ( )/3 = 5.01 Variance of sample averages  s 2 =  2 = [( )2 +2( ) 2 ]/2 = 0.01 s  = pop  5.01 ± 0.05  x 2 = /3 =  x = 0.05  Result: pop  5.01 ±  =5.11  =1.11 =4.96  =  =4.96  = independent sample measurements (equivalent statistics): x

W. Udo Schröder, 2009 Statistics 7 Example of Gaussian Population x m x m Sample size makes a difference (  weighted average) n = 10 n = 50 The larger the sample, the narrower the distribution of x values, the more it approaches the true Gaussian (normal) distribution.

W. Udo Schröder, 2009 Statistics 8 Central-Limit Theorem Increasing size n of samples: Distribution of sample means  Gaussian normal distrib. regardless of form of original (population) distribution. The means (averages) of different samples in the previous example cluster together closely.  general property:  The average of a distribution does not contain information on the shape of the distribution. The average of any truly random sample of a population is already somewhat close to the true population average. Many or large samples narrow the choices: smaller Gaussian width  Standard error of the mean decreases with incr. sample size

W. Udo Schröder, 2009 Statistics 9 Integer random variable m = number of events, out of N total, of a given type, e.g., decay of m (from a sample of N ) radioactive nuclei, or detection of m (out of N ) photons arriving at detector. p = probability for a (one) success (decay of one nucleus, detection of one photon) Choose an arbitrary sample of m trials out of N trials p m = probability for at least m successes (1-p) N-m = probability for N-m failures (survivals, escaping detection) Probability for exactly m successes out of a total of N trials How many ways can m events be chosen out of N ?  Binomial coefficient Total probability (success rate) for any sample of m events: Binomial Distribution

W. Udo Schröder, 2009 Statistics 10 Moments and Limits Distributions for N=30 and p=0.1  p=0.3 Poisson  Gaussian Probability for m “successes” out of N trials, individual probability p

W. Udo Schröder, 2009 Statistics 11 Poisson Probability Distribution Probability for observing m events when average is =  m=0,1,2,…  2 =   = = N·p and  2 =  is the mean, the average number of successes in N trials. Observe N counts (events)  uncertainty is  = √  Unlike the binomial distribution, the Poisson distribution does not depend explicitly on p or N ! For large N, p: Poisson  Gaussian (Normal Distribution) Results from binomial distribution in the limit of small p and large N (N·p > 0)

W. Udo Schröder, 2009 Statistics 12 Moments of Transition Probabilities Small probability for process, but many trials (n 0 = 6.38·10 17 )    0< n 0 · < ∞ Statistical process follows a Poisson distribution: n=“random” Different statistical distributions: Binomial, Poisson, Gaussian

W. Udo Schröder, 2009 Statistics 13 Radioactive Decay as Poisson Process Slow radioactive decay of large sample Sample size N » 1, decay probability p « 1, with 0 < N·p <  137 Cs unstable isotope  decay t 1/2 = 27 years  p = ln2/27 = 0.026/a = 8.2· s -1  0 Sample of 1  g: N = nuclei (=trials for decay) How many will decay(= activity  ) ?  = = N·p = 8.2·10 +5 s -1 Count rate estimate =d /dt = (8.2·10 +5 ± 905) s -1 estimated Probability for m actual decays P ( ,m) =

W. Udo Schröder, 2009 Statistics 14 Functions of Stochastic Variables Random independent variable sets {N 1 }, {N 2 },….,{N n } corresponding variances  1 2,  2 2,….,  n 2 Function f(N 1, N 2,….,N n ) defined for any tuple {N 1, N 2,….,N n } Expectation value (mean) Gauss’ law of error propagation: Further terms if N i not independent (  correlations) Otherwise, individual component variances (  f) 2 add.

W. Udo Schröder, 2009 Statistics 15 Example: Spectral Analysis Background B Peak Area A B1B1 B2B2 Analyze peak in range channels c 1 – c 2 : beginning of background left and right of peak n = c 1 – c Total area = N 12 =A+B N(c 1 )=B 1, N(c 2 )=B 2, Linear background = n(B 1 +B 2 )/2 Peak Area A: Adding or subtracting 2 Poisson distributed numbers N 1 and N 2 : Variances always add

W. Udo Schröder, 2009 Statistics 16 Confidence Level Assume normally distributed observable x: Sample distribution with data set  observed average and std. error  approximate population. Confidence level CL (Central Confidence Interval): With confidence level CL (probability in %), the true value differs by less than  = n  from measured average. Trustworthy exptl. results quote ±3  error bars! Measured Probability

W. Udo Schröder, 2009 Statistics 17 Setting Confidence Limits Example: Search for rare decay with decay rate  observe no counts within time  t. Decay probability law dP/dt=-dN/Ndt= exp {- ·t}. P( t ) = symmetric in  and t normalized P Upper limit Higher confidence levels CL (0  CL  1)  larger upper limits for a given time  t inspected. Reduce limit by measuring for longer period. no counts in  t

W. Udo Schröder, 2009 Statistics 18 Maximum Likelihood Measurement of correlations between observables y and x: {x i, y i | i=1-N} Hypothesis: y(x) =f(c 1,…,c m ; x). Parameters defining f: {c 1,…,c m } n dof =N-m degrees of freedom for a “fit” of the data with f. for every data point Maximize simultaneous probability Minimize chi-squared by varying {c 1,…,c m }: ∂  2 /∂c i = 0 When is   as good as can be?

W. Udo Schröder, 2009 Statistics 19 Minimizing  2 Example: linear fit f(a,b;x) = a + b·x to data set {x i, y i,  i } Minimize: Equivalent to solving system of linear equations

W. Udo Schröder, 2009 Statistics 20 Distribution of Chi-Squareds Distribution of possible  2 for data sets distributed normally about a theoretical expectation (function) with n dof degrees of freedom: (Stirling’s formula) Reduced  2 : ndof =5 u:=  2 Should be P  0.5 for a reasonable fit

W. Udo Schröder, 2009 Statistics 21 CL for  2 -Distributions 1-CL

W. Udo Schröder, 2009 Statistics 22 Correlations in Data Sets Correlations within data set. Example: y i small whenever x i small x y uncorrel. P(x,y) x y correlated P(x,y) 

W. Udo Schröder, 2009 Statistics 23 Correlations in Data Sets uncertainties of deduced most likely parameters c i (e.g., a, b for linear fit) depend on depth/shallowness and shape of the  2 surface cici cjcj uncorrelated  2 surface cici cjcj correlated  2 surface

W. Udo Schröder, 2009 Statistics 24 Multivariate Correlations cici cjcj initial guess search path Smoothed  2 surface Different search strategies: Steepest gradient, Newton method w or w/o damping of oscillations, Biased MC:Metropolis MC algorithms, Simulated annealing (MC derived from metallurgy),  Various software packages LINFIT, MINUIT,….