Presentation is loading. Please wait.

Presentation is loading. Please wait.

Computing and Statistical Data Analysis / Stat 8

Similar presentations


Presentation on theme: "Computing and Statistical Data Analysis / Stat 8"— Presentation transcript:

1 Computing and Statistical Data Analysis / Stat 8
Computing and Statistical Data Analysis Stat 8: More Parameter Estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway, University of London Course web page: G. Cowan Computing and Statistical Data Analysis / Stat 8

2 Computing and Statistical Data Analysis / Stat 8
ML with binned data Often put data into a histogram: Hypothesis is where If we model the data as multinomial (ntot constant), then the log-likelihood function is: G. Cowan Computing and Statistical Data Analysis / Stat 8

3 ML example with binned data
Previous example with exponential, now put data into histogram: Limit of zero bin width → usual unbinned ML. If ni treated as Poisson, we get extended log-likelihood: G. Cowan Computing and Statistical Data Analysis / Stat 8

4 Relationship between ML and Bayesian estimators
In Bayesian statistics, both q and x are random variables: Recall the Bayesian method: Use subjective probability for hypotheses (q); before experiment, knowledge summarized by prior pdf p(q); use Bayes’ theorem to update prior in light of data: Posterior pdf (conditional pdf for q given x) G. Cowan Computing and Statistical Data Analysis / Stat 8

5 ML and Bayesian estimators (2)
Purist Bayesian: p(q | x) contains all knowledge about q. Pragmatist Bayesian: p(q | x) could be a complicated function, → summarize using an estimator Take mode of p(q | x) , (could also use e.g. expectation value) What do we use for p(q)? No golden rule (subjective!), often represent ‘prior ignorance’ by p(q) = constant, in which case But... we could have used a different parameter, e.g., l = 1/q, and if prior pq(q) is constant, then pl(l) is not! ‘Complete prior ignorance’ is not well defined. G. Cowan Computing and Statistical Data Analysis / Stat 8

6 The method of least squares
Suppose we measure N values, y1, ..., yN, assumed to be independent Gaussian r.v.s with Assume known values of the control variable x1, ..., xN and known variances We want to estimate θ, i.e., fit the curve to the data points. The likelihood function is G. Cowan Computing and Statistical Data Analysis / Stat 8

7 The method of least squares (2)
The log-likelihood function is therefore So maximizing the likelihood is equivalent to minimizing Minimum defines the least squares (LS) estimator Very often measurement errors are ~Gaussian and so ML and LS are essentially the same. Often minimize χ2 numerically (e.g. program MINUIT). G. Cowan Computing and Statistical Data Analysis / Stat 8

8 LS with correlated measurements
If the yi follow a multivariate Gaussian, covariance matrix V, Then maximizing the likelihood is equivalent to minimizing G. Cowan Computing and Statistical Data Analysis / Stat 8


Download ppt "Computing and Statistical Data Analysis / Stat 8"

Similar presentations


Ads by Google