פרקים נבחרים בפיסיקת החלקיקים אבנר סופר אביב 2007 6.

Slides:



Advertisements
Similar presentations
The Simple Regression Model
Advertisements

Laws of division of casual sizes. Binomial law of division.
Intensity Transformations (Chapter 3)
Use of moment generating functions. Definition Let X denote a random variable with probability density function f(x) if continuous (probability mass function.
Lecture 2 Today: Statistical Review cont’d:
Chapter 2: Probability Random Variable (r.v.) is a variable whose value is unknown until it is observed. The value of a random variable results from an.
DEPARTMENT OF HEALTH SCIENCE AND TECHNOLOGY STOCHASTIC SIGNALS AND PROCESSES Lecture 1 WELCOME.
Review of Basic Probability and Statistics
Probability Densities
Point estimation, interval estimation
G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1 Statistical Data Analysis: Lecture 2 1Probability, Bayes’ theorem 2Random variables and.
Class notes for ISE 201 San Jose State University
Chapter 6 Continuous Random Variables and Probability Distributions
Evaluating Hypotheses
K. Desch – Statistical methods of data analysis SS10 2. Probability 2.3 Joint p.d.f.´s of several random variables Examples: Experiment yields several.
Tch-prob1 Chapter 4. Multiple Random Variables Ex Select a student’s name from an urn. S In some random experiments, a number of different quantities.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
Lecture 4 Probability and what it has to do with data analysis.
Continuous Random Variables and Probability Distributions
Chapter 4: Joint and Conditional Distributions
Review of Probability and Statistics
The joint probability distribution function of X and Y is denoted by f XY (x,y). The marginal probability distribution function of X, f X (x) is obtained.
Random Variable and Probability Distribution
Lecture II-2: Probability Review
5-1 Two Discrete Random Variables Example Two Discrete Random Variables Figure 5-1 Joint probability distribution of X and Y in Example 5-1.
The Multivariate Normal Distribution, Part 2 BMTRY 726 1/14/2014.
5-1 Two Discrete Random Variables Example Two Discrete Random Variables Figure 5-1 Joint probability distribution of X and Y in Example 5-1.
Modern Navigation Thomas Herring
NIPRL Chapter 2. Random Variables 2.1 Discrete Random Variables 2.2 Continuous Random Variables 2.3 The Expectation of a Random Variable 2.4 The Variance.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Standard error of estimate & Confidence interval.
Review of Probability.
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
OUTLINE Probability Theory Linear Algebra Probability makes extensive use of set operations, A set is a collection of objects, which are the elements.
Short Resume of Statistical Terms Fall 2013 By Yaohang Li, Ph.D.
1 Probability and Statistics  What is probability?  What is statistics?
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
CHAPTER 4 Multiple Random Variable
Lecture 3 A Brief Review of Some Important Statistical Concepts.
CPSC 531: Probability Review1 CPSC 531:Probability & Statistics: Review II Instructor: Anirban Mahanti Office: ICT 745
1 G Lect 8b G Lecture 8b Correlation: quantifying linear association between random variables Example: Okazaki’s inferences from a survey.
0 K. Salah 2. Review of Probability and Statistics Refs: Law & Kelton, Chapter 4.
1 G Lect 4a G Lecture 4a f(X) of special interest: Normal Distribution Are These Random Variables Normally Distributed? Probability Statements.
Lab 3b: Distribution of the mean
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
1 Introduction to Statistical Methods for High Energy Physics Glen Cowan 2006 CERN Summer Student Lectures CERN Summer Student Lectures on Statistics Glen.
1 G Lect 2M Examples of Correlation Random variables and manipulated variables Thinking about joint distributions Thinking about marginal distributions:
CS433 Modeling and Simulation Lecture 03 – Part 01 Probability Review 1 Dr. Anis Koubâa Al-Imam Mohammad Ibn Saud University
Lecture V Probability theory. Lecture questions Classical definition of probability Frequency probability Discrete variable and probability distribution.
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
Uniovi1 Some statistics books, papers, etc. G. Cowan, Statistical Data Analysis, Clarendon, Oxford, 1998 see also R.J. Barlow,
CY1B2 Statistics1 (ii) Poisson distribution The Poisson distribution resembles the binomial distribution if the probability of an accident is very small.
Probability (outcome k) = Relative Frequency of k
Continuous Random Variables and Probability Distributions
CHAPTER – 1 UNCERTAINTIES IN MEASUREMENTS. 1.3 PARENT AND SAMPLE DISTRIBUTIONS  If we make a measurement x i in of a quantity x, we expect our observation.
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
G. Cowan Computing and Statistical Data Analysis / Stat 1 1 Computing and Statistical Data Analysis London Postgraduate Lectures on Particle Physics; University.
Pattern Recognition Mathematic Review Hamid R. Rabiee Jafar Muhammadi Ali Jalali.
디지털통신 Random Process 임 민 중 동국대학교 정보통신공학과 1.
Biostatistics Class 3 Probability Distributions 2/15/2000.
Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Chapter 3: Discrete Random Variables and Their Distributions CIS.
Review of Probability Theory
3.1 Expectation Expectation Example
MEGN 537 – Probabilistic Biomechanics Ch.3 – Quantifying Uncertainty
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005
CHAPTER – 1.2 UNCERTAINTIES IN MEASUREMENTS.
Chapter 2. Random Variables
CHAPTER – 1.2 UNCERTAINTIES IN MEASUREMENTS.
Presentation transcript:

פרקים נבחרים בפיסיקת החלקיקים אבנר סופר אביב

Administrative stuff Projects status Other homework problems: –Open questions in HW #1 (questions about the Quantum Universe) and HW #3 (difference between D mixing and Bs mixing analyses) – we will go over them when return from break The plan for the next few weeks: –Statistics (with as many real examples as possible) –Root and RooFit –Practicing statistics and analysis techniques Lecture on Tuesday, April 10 (Mimona) instead of Monday (Passover break)?

Why do we use statistics in EPP? Scientific claims need to be based on solid mathematics –How confident are we of the result? What is the probability that we are wrong? –Especially important when working at the frontier of knowledge: extraordinary claims require extraordinary proof Proving something with high certainty is usually expensive –Many first measurements are made with marginal certainty Statistical standards: –“Evidence” –“Observation”

Probability Set S (sample space) Subset A  S The probability P(A) is a real number that satisfies the axioms 1.P(A)  0 2.If A and B are disjoint subsets, i.e., A  B = 0, then P(A  B) = P(A) + P(B) 3.P(S) = 1

Derived properties P(!A) = 1 – P(A), where !A = S – A P(A  !A) = 1 1  P(A)  0 P(null set) = 0 If A  B, then P(B)  P(A) P(A  B) = P(A) + P(B) – P(A  B)

More definitions Subsets A and B are independent if P(A  B) = P(A) P(B) A random variable x is a variable that has a specific value for each element of the set An element may have more than 1 random variables: x = {x 1, …x n }

Interpretation of Probability in data analysis Limiting relative frequency: –Elements of the sample space S = possible outcomes of a repeatable measurement –The probability of a particular outcome e (= element of S) is (note that the single element e belongs to a subset with one element = an elementary subset) –A non-elementary subset A corresponds to an occurrence of any of the outcomes in the subset, with probability

Example 1 Element e = D mixing parameter y’ measured to be 0.01 Subset A = y’ measured to be in range [0.005, 0.015] P(A) = fraction of experiments in which y’ is measured in [0.005, 0.015], given that its true value is 0.002

Example 2 e = (x’ 2, y’) measured to be ( , 0.01) A = (x’ 2, y’) measured to be anywhere outside the brown (“4  ”) contour P(A) = fraction of experiments in which (x’ 2, y’) are measured outside the contour, given that their true values are the measured ones

Example 3 e = error on CP-violating parameter   measured to be 42  A =   error measured to be 42  or greater P(A) = fraction of experiments in which the   error is measured to be 42  or greater

About the relative frequency interpretation Straightforward when measurements are repeatable: –Particle collisions in an experiment –Radioactive decays of identical nuclei Also works when measurements are repeatable only in principle : –Measurement of the D mixing parameters using all the data we will ever have –Measurement of the average height of all humans Physical laws don’t change

Probability density functions Outcome of an experiment is a continuous random variable x –Applies to most measurements in particle physics Define: the probability density function (PDF) f(x), such that f(x) dx = probability to observe x in [x, x+dx] = fraction of experiments in which x will be measured in [x, x+dx] To satisfy axiom 3: P(S) = 1, normalize the PDF:

The PDF and finite number of observations A set of n meas measurements x m (m=1…n meas ) can be presented as a histogram: n b (b=1…n bins ) = number of measurements for which x falls in bin b n b / n meas = probability for a measurement to be in bin b. –  b n b / n meas = 1 n b / (n meas  x b ) = (discrete) probability density function Continuum limit (infinite number of observations, infinitely fine binning):

Cumulative distribution The cumulative distribution of f(x) is Alternatively: F(x) = probability to obtain measurement whose value is < x f(x) = dF(x)/dx (for differentiable F(x))  -point x  is the value of x such that F(x  ) = , where 1    0. Or: x  = F  (  ) Median = x ½ = value of x such that F(x ½ ) = ½ Mode = x mode such that f(x mode ) > f(all other values of x) –may not be useful or unique if f(x) has multiple local maxima

Extension to multi-variable PDFs For f(x), x = {x 1, … x n }, the  -point turns into an  -contour of dimension n-1 Marginal PDFs: –f x (x) =  f(x,y) dy –f y (y) =  f(x,y) dx x and y are independent variables if f(x,y) = f x (x) f y (y) –Also called uncorrelated variables

Functions of random variables a(x) is a continuous function of random variable x, which has PDF f(x) –E.g., a = x 2, a = log(x), etc. What is the PDF g(a)? Require equal probabilities in corresponding infinitesimal regions: g(a) da = f(x) dx  g(a) = f(x(a)) |dx/da| Assumes a(x) can be inverted Abs value to keep PDF positive x a dx da

Example The CP-violation phases  are not measured directly. We measure cos  or sin  or sin 2 , then transform to the phases:

Multiple-valued x(a) If a(x) is not uniquely invertable, need to add up the different contributions. x a dx 1 da dx 2 dS(a) = sum of 2 regions

Functions of multiple random variables What is g(a) for a(x), x = {x 1, … x n } Example.: z = xy, what is f(z) given g(x) and h(y)? f(z) is the Mellin convolution of g(x) h(y) dS is the hypersurface in x that encloses [a, a+da]

Another example: z = x + y f(z) is the familiar Fourier convolution of g(x) and h(y). Recall from the D mixing analysis: The measured decay time t is the true decay time t’ (distribution P(t’)) + a random detector error  t (distribution  (  t):

Multiple functions of multiple random variables g(a 1, …. a n ) = f(x 1, … x n ) |J|, where the Jacobian is To determine the marginal distribution g i (a i ), need to integrate g(a 1, …. a n ) over the a j  i variables

Expectation values The expectation value of a random variable x distributed according to the PDF f(x): Also called population mean E[x] is the most commonly used location parameter (others are the  - point x  and the mode) The expectation value of a function a(x) is

Moments The nth algebraic moment of f(x): –Note that the population mean  is the special case  ’ 1 The nth central moment In particular, is the population variance of f(x) The standard deviation gives an idea of the spread of f(x)

Mean and variance of functions Take a function of many random variables: a(x). Then

Covariance For 2 random variables x, y, the covariance cov[x,y] or V xy is For 2 functions a(x), b(x), the covariance is Note that V ab = V ba and V aa =  a 2 The dimensionless correlation coefficient is  ab = V ab / (  a  b ) –Note that 1   ab   1

Understanding covariance and correlation V xy =E[(x  x )(y  y )] is the expectation value of the product of the deviations from the means. If having x >  x increases the probability of having y >  y then V xy > 0, x and y are positively correlated If having x >  x increases the probability of having y <  y then V xy < 0, x and y are negatively correlated or anti-correlated. For independent variables (defined as f(x,y) = f x (x) f y (y)), we find E[xy] = E[x] E[y] =  x  y so V xy = 0. Does V xy = 0 necessarily mean that the variables are independent?...

Covariance and correlation …No. E.g.,

Propagation of errors Take n random variables x with unknown PDF f(x), but with E[x] and V ij known (or estimated) Take the function y(x). What are E[y] and V[y]? –Remember: we don’t know f(x). Expand y:

Why is this “error propagation”? Because we often estimate errors from covariances

Special cases Note: These formulae don’t work if y is significantly non-linear within a distance  i around the mean 

Orthogonal transformation of variables It is often useful to work in variables in which the covariance matrix is diagonal: cov[y i,y j ] =  i 2  ij This can always achieved with a linear transformation: where the rows of the transformation matrix A ij are the eigenvectors of cov[x i,x j ]. Then  i 2 are the eigenvalues of cov[x i,x j ]

Visualize for 2 dimensions Recall the definition of the correlation coefficient  ab = V ab / (  a  b ). So we can write Eigenvector 1 Eigenvector 2

More on linear variable transformations The uncorrelated variables y i have a simpler covariance matrix, but may not always correspond to physically interesting quantities E.g., in D mixing, x’ 2 and y’ have a very high correlation coefficient of  =  0.94 But they are the physically interesting variables…