1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

October 1999 Statistical Methods for Computer Science Marie desJardins CMSC 601 April 9, 2012 Material adapted.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #22.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Maximum likelihood (ML) and likelihood ratio (LR) test
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
Maximum likelihood (ML)
Maximum likelihood (ML) and likelihood ratio (LR) test
Estimation of parameters. Maximum likelihood What has happened was most likely.
Basics of Statistical Estimation. Learning Probabilities: Classical Approach Simplest case: Flipping a thumbtack tails heads True probability  is unknown.
Presenting: Assaf Tzabari
CHAPTER 6 Statistical Analysis of Experimental Data
Probability Distributions Random Variables: Finite and Continuous A review MAT174, Spring 2004.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Statistical inference Population - collection of all subjects or objects of interest (not necessarily people) Sample - subset of the population used to.
Maximum likelihood (ML)
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
1 Probability and Statistics  What is probability?  What is statistics?
Random Variables & Probability Distributions Outcomes of experiments are, in part, random E.g. Let X 7 be the gender of the 7 th randomly selected student.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Review and Preview This chapter combines the methods of descriptive statistics presented in.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
LECTURER PROF.Dr. DEMIR BAYKA AUTOMOTIVE ENGINEERING LABORATORY I.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #23.
BINOMIALDISTRIBUTION AND ITS APPLICATION. Binomial Distribution  The binomial probability density function –f(x) = n C x p x q n-x for x=0,1,2,3…,n for.
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Yung-Kyun Noh and Joo-kyung Kim Biointelligence.
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
EAS31116/B9036: Statistics in Earth & Atmospheric Sciences Lecture 3: Probability Distributions (cont’d) Instructor: Prof. Johnny Luo
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
POSC 202A: Lecture 4 Probability. We begin with the basics of probability and then move on to expected value. Understanding probability is important because.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #25.
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Computer simulation Sep. 9, QUIZ 2 Determine whether the following experiments have discrete or continuous out comes A fair die is tossed and the.
Sampling and estimation Petter Mostad
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Section 5-1 Review and Preview.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #24.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Lecture Slides Elementary Statistics Eleventh Edition and the Triola Statistics Series by.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Conditional Expectation
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
Probability Distributions ( 확률분포 ) Chapter 5. 2 모든 가능한 ( 확률 ) 변수의 값에 대해 확률을 할당하는 체계 X 가 1, 2, …, 6 의 값을 가진다면 이 6 개 변수 값에 확률을 할당하는 함수 Definition.
Theoretical distributions: the Normal distribution.
4. Overview of Probability Network Performance and Quality of Service.
Bayesian Estimation and Confidence Intervals Lecture XXII.
Introductory Statistics and Data Analysis
The Maximum Likelihood Method
Bayesian Estimation and Confidence Intervals
Probability and Statistics for Particle Physics
The Maximum Likelihood Method
Basic Probability Theory
The Maximum Likelihood Method
Econometric Models The most basic econometric model consists of a relationship between two variables which is disturbed by a random error. We need to use.
Chapter 5 Sampling Distributions
Statistical NLP: Lecture 4
Lecture Slides Elementary Statistics Twelfth Edition
Lecture Slides Elementary Statistics Twelfth Edition
Computing and Statistical Data Analysis / Stat 7
M248: Analyzing data Block A UNIT A3 Modeling Variation.
Lecture Slides Essentials of Statistics 5th Edition
Mathematical Foundations of BME Reza Shadmehr
Lecture Slides Essentials of Statistics 5th Edition
Presentation transcript:

1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21

Data Analysis Example Simple? Yea, but we just discovered the Z boson –Too bad we are 40 years late, otherwise we would be in line for the Nobel prize

Measurements

Probability and Statistics Important when looking for small signals when you have little opportunity to quickly increase your statistics Unimportant when you do something that has large statistics Unimportant when you do something that you can easily reproducible In the world of small numbers, everything boils down to probability of this versus that A question of “Do I see signal or do I see background?” requires some sort of probabilistic qualification to be a well defined question 4

Probability Fairly intuitive term referring to “inexact” processes Poisson process example: When something happens with a certain rate (frequency over time), how many of these occurences will you get over an hour? Slight simplification but close: you toss a coin every second and count how many times will get tails over an hour If you repeat this experiment 1,000,000 times and plot the outcomes (the number of heads in each experiment) on the graph and then normalize. You will get the Poisson distribution of getting N outcomes out of 1,000 when the expected rate is 500 (given that the probability is p=0.5, right?) 5

Poisson Probability 6

Using Probability Distributions 7

Probability in Statistics The usual “frequentist” definition of probability is: If you had to repeat the same experiment many times, in the limit of infinite number of trials, what would be the frequency of something happening This frequency is your probability Assumes no prior knowledge or beliefs, just statistics in action One can also incorporate subjective belief Could be prior knowledge like previous measurements of the same quantity or physical bounds (like a cross-section can’t be less than zero) This is characteristic of Bayesian statistics These are best discussed in examples 8

Other Statistical Distributions Gaussian or normal: The mean is m 0, variance is  2 The central limit theorem: For x i following any distribution with finite mean and variance, y=  x i follows gaussian distribution in the limit of large n The sum of a large number of fluctuations x i will be distributed as gaussian even if x i themselves are not gaussian 9

Binomial When something has two outcomes (like a coin toss) and the probability of a specific outcome is p (correspondingly the probability of the “other” outcome is 1-p in each trial) The probability to observe r successful outcomes out of N trials given true probability of p: Where: 10

Probability Density Function Consider Poisson distribution: Can read it two ways: The probability to observe particular values n for a fixed Or probability of particular values of, given that the number of observed events is n If you integrate Poisson distribution over all possible lambdas you will get 1 for any fixed n, just as if you sum it up over all possible n for a fixed lambda If this is allowable, you have a joint probability density function (p.d.f.), which you can view either as a function of the data (n) or the parameters (lambda) 11

Measurements and Statistics A typical example is a fit, in which you want to measure some unknown parameters of the “theory” However, when you start thinking of it, there is a lot of conditionality in that, e.g. you assume that at least the parameterization of the “theory” is true – how do you know it to be the case? Therefore, it may well be insufficient to do a simple fit, you often may need to answer many other questions Asking proper questions in statistics is very important as the answer and whether it is meaningful depend on the question Will get back to it later 12

Estimators A typical fit problem: we have a set of N measured quantities x = (x1,..., xN) described by a joint p.d.f. f(x; µ), where µ = (µ1,..., µn) is set of n parameters whose values are unknown Any measurement is an attempt to estimate true parameters of the “theory” using observed data But you are only estimating and your estimates may have biases, so people usually talk about “estimators” There are various ways to build estimators, most (if not all usually used) estimators converge to the true value of the parameters But not all can be equal (the “efficiency”, which is how fast your estimator converges to the true value with more data, can matter) Various fitting techniques are effectively different ways of picking the estimators 13

Common Estimators For a set of data xi following the same pdf with a common mean, the estimators: The mean: The variance: Variance of the estimator of the mean: Variance of the estimator of the variance is more complex (relies on higher moments of the true distribution): 14

Maximum Likelihood Likelihood for N measurements x i : It is essentially a joint p.d.f. seen as a function of the parameters  While xi could be any measurements, it’s easy to visualize it as a histogram of x with N bins; in each bin you calculate how probable it is to see x i for a particular  (or set of  ’s) In ML, the estimators are those values of  that maximize the likelihood function One could use MINUIT to find the position of minimum for -log(L) 15

TO BE CONTINUED 16