Probability Frequentist versus Bayesian and why it matters Roger Barlow Manchester University 11 th February 2008.

Slides:



Advertisements
Similar presentations
October 1999 Statistical Methods for Computer Science Marie desJardins CMSC 601 April 9, 2012 Material adapted.
Advertisements

27 th March CERN Higgs searches: CL s W. J. Murray RAL.
Practical Statistics for LHC Physicists Bayesian Inference Harrison B. Prosper Florida State University CERN Academic Training Lectures 9 April, 2015.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
1 LIMITS Why limits? Methods for upper limits Desirable properties Dealing with systematics Feldman-Cousins Recommendations.
An Answer and a Question Limits: Combining 2 results Significance: Does  2 give  2 ? Roger Barlow BIRS meeting July 2006.
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Statistics In HEP 2 Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Statistics for Particle Physics: Intervals Roger Barlow Karlsruhe: 12 October 2009.
Slide 1 Statistics for HEP Roger Barlow Manchester University Lecture 1: Probability.
Probability theory Much inspired by the presentation of Kren and Samuelsson.
Results from Small Numbers Roger Barlow Manchester 26 th September 2007.
Slide 1 Statistics for HEP Roger Barlow Manchester University Lecture 3: Estimation.
Developments in Bayesian Priors Roger Barlow Manchester IoP meeting November 16 th 2005.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
I The meaning of chance Axiomatization. E Plurbus Unum.
Statistics for Particle Physics: Limits Roger Barlow Karlsruhe: 12 October 2009.
Slide 1 Statistics Workshop Tutorial 4 Probability Probability Distributions.
1.  Why understanding probability is important?  What is normal curve  How to compute and interpret z scores. 2.
Probability (cont.). Assigning Probabilities A probability is a value between 0 and 1 and is written either as a fraction or as a proportion. For the.
Sensitivity of searches for new signals and its optimization
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Everything You Always Wanted To Know About Limits* Roger Barlow Manchester University YETI06 *But were afraid to ask.
Maximum likelihood (ML)
Statistical Analysis of Systematic Errors and Small Signals Reinhard Schwienhorst University of Minnesota 10/26/99.
Statistical aspects of Higgs analyses W. Verkerke (NIKHEF)
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
STA Lecture 161 STA 291 Lecture 16 Normal distributions: ( mean and SD ) use table or web page. The sampling distribution of and are both (approximately)
Estimation of Statistical Parameters
1 Probability and Statistics  What is probability?  What is statistics?
Statistical Decision Theory
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
16-1 Copyright  2010 McGraw-Hill Australia Pty Ltd PowerPoint slides to accompany Croucher, Introductory Mathematics and Statistics, 5e Chapter 16 The.
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 5 Section 1 – Slide 1 of 33 Chapter 5 Section 1 Probability Rules.
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
Ex St 801 Statistical Methods Probability and Distributions.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 1 – Slide 1 of 34 Chapter 11 Section 1 Random Variables.
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Statistics In HEP Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
Making sense of randomness
Confidence Intervals First ICFA Instrumentation School/Workshop At Morelia, Mexico, November 18-29, 2002 Harrison B. Prosper Florida State University.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #25.
Sampling and estimation Petter Mostad
BAYES and FREQUENTISM: The Return of an Old Controversy 1 Louis Lyons Imperial College and Oxford University CERN Summer Students July 2014.
Practical Statistics for Particle Physicists Lecture 2 Harrison B. Prosper Florida State University European School of High-Energy Physics Parádfürdő,
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #24.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Lecture 4 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Chapter 8: Probability: The Mathematics of Chance Probability Models and Rules 1 Probability Theory  The mathematical description of randomness.  Companies.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
10.1 Estimating with Confidence Chapter 10 Introduction to Inference.
Probability Distributions ( 확률분포 ) Chapter 5. 2 모든 가능한 ( 확률 ) 변수의 값에 대해 확률을 할당하는 체계 X 가 1, 2, …, 6 의 값을 가진다면 이 6 개 변수 값에 확률을 할당하는 함수 Definition.
Frequentist Confidence Limits and Intervals Roger Barlow SLUO Lectures on Statistics August 2006.
Statistics Probability and Likelihood Roger Barlow Manchester University Hadron Collider Physics Summer School, Fermilab 20 th August 2010.
BAYES and FREQUENTISM: The Return of an Old Controversy
Unit 5: Hypothesis Testing
Confidence Intervals and Limits
Lecture 4 1 Probability (90 min.)
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Confidence Intervals for Proportions
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Statistics for HEP Roger Barlow Manchester University
Everything You Always Wanted To Know About Limits*
Presentation transcript:

Probability Frequentist versus Bayesian and why it matters Roger Barlow Manchester University 11 th February 2008

Frequentist and Bayesian Probability Roger Barlow Slide 2 Outline Different definitions of probability: Frequentist and Bayesian Measurements: definitions usually give the same results Differences in dealing with  Nongaussian measurements  Small number counting  Constrained parameters Difficulties for both Conclusions and recommendations

Frequentist and Bayesian Probability Roger Barlow Slide 3 What is Probability? A is some possible event or fact. What is P(A)? Classical: An intrinsic property Frequentist: Limit N  N(A) / N Bayesian: My degree of belief in A What do we mean by P(A)?

Frequentist and Bayesian Probability Roger Barlow Slide 4 Classical (Laplace and others) Symmetry factor Coin – ½ Cards – 1 / 52 Dice – 1 / 6 Roulette – 1 / 32 Equally likely outcomes Extend to more complicated systems of several coins, many cards, etc. The probability of an event is the ratio of the number of cases favourable to it, to the number of all cases possible when nothing leads us to expect that any one of these cases should occur more than any other, which renders them, for us, equally possible. Théorie analytique des probabilités

Frequentist and Bayesian Probability Roger Barlow Slide 5 Classical Probability Breaks down Can’t handle continuous variables Bertrand’s paradox: if we draw a chord at random, what is the probability that it is longer than the side of the triangle? Answer: 1/3 or ½ or 1/4 Cannot enumerate ‘equally likely’ cases in a unique way

Frequentist and Bayesian Probability Roger Barlow Slide 6 Frequentist Probability (von Mises, Fisher) Limit of frequency P(A)= Limit N  N(A)/N This was a property of the classical definition, now promoted to become a definition itself P(A) depends not just on A but on the ensemble – which must be specified. This leads to two surprising features Ensemble of Everything A

Frequentist and Bayesian Probability Roger Barlow Slide 7 Feature 1:There can be many Ensembles Probabilities belong to the event and the ensemble Insurance company data shows P(death) for 40 year old male clients = 1.4% (example due to von Mises) Does this mean a particular 40 year old German has a 98.6% chance of reaching his 41 st Birthday? No. He belongs to many ensembles  German insured males  German males  Insured nonsmoking vegetarians  German insured male racing drivers  … Each of these gives a different number. All equally valid.

Frequentist and Bayesian Probability Roger Barlow Slide 8 Feature 2: Unique events have no ensemble Some events are unique. Consider “It will probably rain tomorrow.” There is only one tomorrow (Tuesday 12 th February). There is NO ensemble. P(rain) is either 0/1 =0 or 1/1 = 1 Strict frequentists cannot say 'It will probably rain tomorrow'. This presents severe social problems.

Frequentist and Bayesian Probability Roger Barlow Slide 9 Circumventing the limitation A frequentist can say: “The statement ‘It will rain tomorrow’ has a 70% probability of being true.” by assembling an ensemble of statements and ascertaining that 70% (say) are true. (E.g. Weather forecasts with a verified track record) Say “It will rain tomorrow” with 70% confidence For unique events, confidence level statements replace probability statements.

Frequentist and Bayesian Probability Roger Barlow Slide 10 Bayesian (Subjective) Probability P(A) is a number describing my degree of belief in A 1=certain belief. 0=total disbelief Can be calibrated against simple classical probabilities. P(A)=0.5 means: I would be indifferent given the choice of betting on A or betting on a coin toss. A can be anything: death, rain, horse races, existence of SUSY… Very adaptable. But no guarantee my P(A) is the same as your P(A). Subjective = unscientific?

Frequentist and Bayesian Probability Roger Barlow Slide 11 General (uncontroversial) form P(A|B)P(B) = P(A & B) = P(B|A) P(A ) P(A|B)=P(B|A) P(A) P(B) P(B) can be written P(B|A) P(A) + P(B|not A) (1-P(A)) Examples: People P(Artist|Beard)=P(Beard|Artist) P(Artist) P(Beard)  /K Cherenkov counter P(  |signal)=P(signal|  ) P(  ) P(signal) Medical diagnosis P(disease|symptom)=P(symptom|disease) P(disease) P(symptom) Bayes’ Theorem 0.9*0.5/(.9*.5+.01*.5)= 0.989

Frequentist and Bayesian Probability Roger Barlow Slide 12 Bayes’ Theorem “Bayesian” form P(Theory|Data)=P(Data|Theory) P(Theory) P(Data) “Theory” may be an event (e.g. rain tomorrow) Or a parameter value (e.g. Higgs Mass): then P(Theory) is a function P(M H |Data)=P(Data|M H ) P(M H ) P(Data) Prior Posterior

Frequentist and Bayesian Probability Roger Barlow Slide 13 Measurements:Bayes at work Result value x Theoretical ‘true’ value  P(  |x)  P(x|  ) P(  ) Prior is generally taken as uniform Ignore normalisation problems Construct theory of measurements – prior of second measurement is posterior of the first P(x|  ) is often Gaussian, but can be anything (Poisson, etc) For Gaussian measurement and uniform prior, get Gaussian posterior = X

Frequentist and Bayesian Probability Roger Barlow Slide 14 Aside: “Objective” Bayesian statistics Attempt to lay down rule for choice of prior ‘Uniform’ is not enough. Uniform in what? Suggestion (Jeffreys): uniform in a variable for which the expected Fisher information is minimum (statisticians call this a ‘flat prior’). Has not met with general agreement – different measurements of the same quantity have different objective priors

Frequentist and Bayesian Probability Roger Barlow Slide 15 Measurement and Frequentist probability M T =174  3 GeV : What does it mean? For true value  the probability (density) for a result x is (for the usual Gaussian measurement) P( x ; ,  )=( 1 /  2  ) exp-[( x -  ) 2 /2  2 ] For a given , the probability that x lies within  is 68%. This does not mean that for a given x, the ‘inverse’ probability that  lies within  is 68% P( x ; ,  ) cannot be used as a probability for . (It is called the likelihood function for  given x.) M T =174  3 GeV Is there a 68% probability that M T lies between 171 and 177 GeV? No. M T is unique. It is either in the range or outside. (Soon we’ll know.) But   3 does bracket x 68% of the time: The statement ‘M T lies between 171 and 177 GeV’ has a 68% probability of being true. M T lies between 171 and 177 GeV with 68% confidence

Frequentist and Bayesian Probability Roger Barlow Slide 16 Pause for breath For Gaussian measurements of quantities with no constraints/objective prior knowledge the same results are given by: Frequentist confidence intervals Bayesian posteriors from uniform priors A frequentist and a simple Bayesian will report the same outcome from the same raw data, except one will say ‘confidence’ and the other ‘probability’. They mean something different but such concerns can be left to the philosophers

Frequentist and Bayesian Probability Roger Barlow Slide 17 Frequentist confidence intervals beyond the simple Gaussian Select Confidence Level value CL and strategy From P(x,  ) choose construction (functions x 1 (  ), x 2 (  ) ) for which P( x  [ x 1 (  ), x 2 (  ) ])  CL for all  Given a measurement X, make statement  [  LO,  HI CL Where X= x 2 (  LO ), X= x 1 (  HI ) (Neyman technique)

Frequentist and Bayesian Probability Roger Barlow Slide 18 Confidence Belt x  Example: proportional Gaussian  = 0.1  Measures with 10% accuracy Result (say)  LO =90.91  HI = Constructed horizontally such that the probability of a result lying inside the belt is 68%(or whatever) Read vertically using the measurement X

Frequentist and Bayesian Probability Roger Barlow Slide 19 Bayesian Proportional Gaussian Likelihood function C exp(- ½(  -100) 2 /(0.1  ) 2 ) Integration gives C= % (central) limits 92.6 and Different techniques give different answers 16% 68% 16%

Frequentist and Bayesian Probability Roger Barlow Slide 20 Small number counting experiments Poisson distribution P(r;  )= e -   r / r! For large  can use Gaussian approx. But not small  Frequentists: Choose CL. Just use one curve to give upper limit Discrete observable makes smooth curves into ugly staircases Observe n. Quote upper limit as  HI from solving  0 n P(r,  HI ) =  0 n e -  HI  HI r /r! = 1-CL Translation. n is small.  can’t be very large. If the true value is  HI (or higher) then the chance of a result this small (or smaller) is only (1-CL) (or less)

Frequentist and Bayesian Probability Roger Barlow Slide 21 Frequentist Poisson Table Upper limits n90%95%99%

Frequentist and Bayesian Probability Roger Barlow Slide 22 Bayesian limits from small number counts P(r,  )=exp(-  )  r /r! With uniform prior this gives posterior for  Shown for various small r results Read off intervals... r=6 r=2 r=1 r=0 P(  ) 

Frequentist and Bayesian Probability Roger Barlow Slide 23 Upper limits Upper limit from n events  0  HI exp(-  )  n /n! d  = CL Repeated integration by parts:  0 n exp(-  HI )  HI r /r! = 1-CL Same as frequentist limit This is a coincidence! Lower Limit formula is not the same

Frequentist and Bayesian Probability Roger Barlow Slide 24 Result depends on Prior Example: 90% CL Limit from 0 events Prior flat in  Prior flat in  X X = =

Frequentist and Bayesian Probability Roger Barlow Slide 25 Which is right? Bayesian Method is generally easier, conceptually and in practice Frequentist method is truly objective. Bayesian probability is personal ‘degree of belief’. This does not worry biologists but should worry physicists. Ambiguity appears in Bayesian results as differences in prior give different answers, though with enough data these differences vanish Check for ‘robustnesss under change of prior’ is standard statistical technique, generally ignored by physicists ‘Uniform priors’ is not a sufficient answer. Uniform in what?

Frequentist and Bayesian Probability Roger Barlow Slide 26 Problems for Frequentists: Add a background  =S+b Frequentist method (b known,  measured, S wanted  1. Find range for  2. Subtract b to get range for S Examples: See 5 events, background % Upper limit: 10.5  9.3  See 5 events, background % Upper limit: 10.5  5.4 ? See 5 events, background % Upper limit: 10.5  -0.1 

Frequentist and Bayesian Probability Roger Barlow Slide 27 S< -0.1? What’s going on? If N<b we know that there is a downward fluctuation in the background. (Which happens…) But there is no way of incorporating this information without messing up the ensemble Really strict frequentist procedure is to go ahead and publish. We know that 5% of 95%CL statements are wrong – this is one of them Suppressing this publication will bias the global results

Frequentist and Bayesian Probability Roger Barlow Slide 28 Similar problems Expected number of events must be non-negative Mass of an object must be non-negative Mass-squared of an object must be non-negative Higgs mass from EW fits must be bigger than LEP2 limit of 114 GeV 3 Solutions Publish a ‘clearly crazy’ result Use Feldman-Cousins technique Switch to Bayesian analysis

Frequentist and Bayesian Probability Roger Barlow Slide 29  =S+b for Bayesians No problem! Prior for  is uniform for S  b Multiply and normalise as before Posterior Likelihood Prior Read off Confidence Levels by integrating posterior = X

Frequentist and Bayesian Probability Roger Barlow Slide 30 Another Aside: Coverage Given P(x;  ) and an ensemble of possible measurements {x i } and some confidence level algorithm, coverage is how often ‘  LO   HI ’ is true. Isn’t that just the confidence level? Not quite. Discrete observables may mean the confidence belt is not exact – move on side of caution Other ‘nuisance’ parameters may need to be taken account of – again erring on side of caution Coverage depends on . For a frequentist it is never less than the CL (‘undercoverage’). It may be more (‘overcoverage’) – this is to be minimised but not crucial For a Bayesian coverage is technically irrelevant – but in practice useful

Frequentist and Bayesian Probability Roger Barlow Slide 31 Bayesian pitfall(Heinrich and others) Observe n events from Poisson with  =  S+b Channel strength S – unknown. Flat prior Efficiency x luminosity  - from sub-experiment with flat prior Background b - from sub-experiment with flat prior Investigated coverage and all OK Partition into classes (e.g. different run periods) Coverage falls!

Frequentist and Bayesian Probability Roger Barlow Slide 32 What’s happening Problem due to efficiency x Lumi priors  =  1 +  2 +  3 +… Uniform density in all N components means P (  )   N-1 ‘Solve’ by taking priors P(  i )  1/  i (Arguments you should have done so in the first place – Jeffreys’ Prior) 22 11

Frequentist and Bayesian Probability Roger Barlow Slide 33 Another example: Unitarity triangle Measure CKM angle  by measuring B  decays (charged and neutral, branching ratios and CP asymmetries). 6 quantities. Many different parametrisations suggested Uniform priors in different parametrisations give different results from each other and from a Frequentist analysis (according to CKMfitter: disputed by UTfit) For a complex number z=x+iy=re i  a flat prior in x and y is not the same as a flat prior in r and 

Frequentist and Bayesian Probability Roger Barlow Slide 34 Loss of ambiguities? Toy example Measure X=(  +  ) 2 =1.00  0.07, Y=  2 =1.10  0.07  Is interesting but  is a nuisance parameter Clearly 4 fold ambiguity  1,  0 or  2 Frequentists stop there Bayesians integrate over  and get a peak at  0 double that at  2 – feature persists whatever the prior used for . Is this valid? (Real example will be more subtle)

Frequentist and Bayesian Probability Roger Barlow Slide 35 Conclusions Frequentist statistics cannot do everything Bayesian statistics can be dangerous. Choose between 1) Never use it 2) Use only if frequentist method has problems 3) Use only with care and expert guidance and always check for robustness under different priors 4) Use as investigative tool to explore possible interpretations 5) Just plug in the package and write down the results But always know what you are doing and say what you are doing.

Frequentist and Bayesian Probability Roger Barlow Slide 36 Backup slides

Frequentist and Bayesian Probability Roger Barlow Slide 37 Incorporating Constraints: Poisson Work with total source strength (s+b) you know is greater than the background b Need to solve Formula not as obvious as it looks.

Frequentist and Bayesian Probability Roger Barlow Slide 38 Feldman Cousins Method Works by attacking what looks like a different problem... * by Feldman and Cousins, mostly

Frequentist and Bayesian Probability Roger Barlow Slide 39 Feldman Cousins:  =s+b b is known. N is measured. s is what we're after This is called 'flip-flopping' and BAD because is wrecks the whole design of the Confidence Belt Suggested solution: 1) Construct belts at chosen CL as before 2) Find new ranking strategy to determine what's inside and what's outside 1 sided 90% 2 sided 90%

Frequentist and Bayesian Probability Roger Barlow Slide 40 Feldman Cousins: Ranking First idea (almost right) Sum/integrate over range of (s+b) values with highest probabilities for this observed N. (advantage that this is the shortest interval) Glitch: Suppose N small. (low fluctuation) P(N;s+b) will be small for any s and never get counted Instead: compare to 'best' probability for this N, at s=N-b or s=0 and rank on that number Such a plot does an automatic ‘flip-flop’ N~b single sided limit (upper bound) for s N>>b 2 sided limits for s

Frequentist and Bayesian Probability Roger Barlow Slide 41 How it works Has to be computed for the appropriate value of background b. (Sounds complicated, but there is lots of software around) As n increases, flips from 1- sided to 2-sided limits – but in such a way that the probability of being in the belt is preserved s n Means that sensible 1-sided limits are quoted instead of nonsensical 2- sided limits!

Frequentist and Bayesian Probability Roger Barlow Slide 42 Arguments against using Feldman Cousins Argument 1 It takes control out of hands of physicist. You might want to quote a 2 sided limit for an expected process, an upper limit for something weird Counter argument: This is the virtue of the method. This control invalidates the conventional technique. The physicist can use their discretion over the CL. In rare cases it is permissible to say ”We set a 2 sided limit, but we're not claiming a signal”

Frequentist and Bayesian Probability Roger Barlow Slide 43 Feldman Cousins: Argument 2 Argument 2 If zero events are observed by two experiments, the one with the higher background b will quote the lower limit. This is unfair to hardworking physicists Counterargument An experiment with higher background has to be ‘lucky’ to get zero events. Luckier experiments will always quote better limits. Averaging over luck, lower values of b get lower limits to report. Example: you reward a good student with a lottery ticket which has a 10% chance of winning $10. A moderate student gets a ticket with a 1% chance of winning $20. They both win. Were you unfair? Example: you reward a good student with a lottery ticket which has a 10% chance of winning £ 10. A moderate student gets a ticket with a 1% chance of winning £ 20. They both win. Were you unfair?

Frequentist and Bayesian Probability Roger Barlow Slide Including Systematic Errors  =aS+b  is predicted number of events S is (unknown) signal source strength. Probably a cross section or branching ratio or decay rate a is an acceptance/luminosity factor known with some (systematic) error b is the background rate, known with some (systematic) error

Frequentist and Bayesian Probability Roger Barlow Slide Full Bayesian Assume priors for S (uniform?) For a (Gaussian?) For b (Poisson or Gaussian?) Write down the posterior P(S,a,b). Integrate over all a,b to get marginalised P(s) Read off desired limits by integration

Frequentist and Bayesian Probability Roger Barlow Slide Hybrid Bayesian Assume priors For a (Gaussian?) For b (Poisson or Gaussian?) Integrate over all a,b to get marginalised P(r,S) Read off desired limits by  0 n P(r,S) =1-CL etc Done approximately for small errors (Cousins and Highland). Shows that limits pretty insensitive to  a,  b Numerically for general errors (RB: java applet on SLAC web page). Includes 3 priors (for a) that give slightly different results

Frequentist and Bayesian Probability Roger Barlow Slide Extend Feldman Cousins Profile Likelihood: Use P(S)=P(n,S,a max,b max ) where a max,b max give maximum for this S,n Empirical Bayes And more… Results being compared as outcome from Banff workshop

Frequentist and Bayesian Probability Roger Barlow Slide 48 Summary Straight Frequentist approach is objective and clean but sometimes gives ‘crazy’ results Bayesian approach is valuable but has problems. Check for robustness under choice of prior Feldman-Cousins deserves more widespread adoption Lots of work still going on This will all be needed at the LHC