Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.

Slides:



Advertisements
Similar presentations
Statistics.  Statistically significant– When the P-value falls below the alpha level, we say that the tests is “statistically significant” at the alpha.
Advertisements

Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
Hypothesis Testing making decisions using sample data.
Hypothesis Testing An introduction. Big picture Use a random sample to learn something about a larger population.
Inference Sampling distributions Hypothesis testing.
Find the Joy in Stats ? ! ? Walt Senterfitt, Ph.D., PWA Los Angeles County Department of Public Health and CHAMP.
QM Spring 2002 Business Statistics Introduction to Inference: Hypothesis Testing.
Hypothesis Testing Steps of a Statistical Significance Test. 1. Assumptions Type of data, form of population, method of sampling, sample size.
Probability (cont.). Assigning Probabilities A probability is a value between 0 and 1 and is written either as a fraction or as a proportion. For the.
Chapter 9 Hypothesis Testing.
BCOR 1020 Business Statistics
Probability Population:
Introduction to Testing a Hypothesis Testing a treatment Descriptive statistics cannot determine if differences are due to chance. A sampling error occurs.
Lecture 9: p-value functions and intro to Bayesian thinking Matthew Fox Advanced Epidemiology.
Inferential Statistics
INFERENTIAL STATISTICS – Samples are only estimates of the population – Sample statistics will be slightly off from the true values of its population’s.
Tests of significance & hypothesis testing Dr. Omar Al Jadaan Assistant Professor – Computer Science & Mathematics.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Fundamentals of Data Analysis Lecture 4 Testing of statistical hypotheses.
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
1 Statistical Inference Greg C Elvers. 2 Why Use Statistical Inference Whenever we collect data, we want our results to be true for the entire population.
Hypothesis Testing: One Sample Cases. Outline: – The logic of hypothesis testing – The Five-Step Model – Hypothesis testing for single sample means (z.
Copyright © 2012 by Nelson Education Limited. Chapter 7 Hypothesis Testing I: The One-Sample Case 7-1.
STA Statistical Inference
Introduction to Probability Theory March 24, 2015 Credits for slides: Allan, Arms, Mihalcea, Schutze.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Psy B07 Chapter 4Slide 1 SAMPLING DISTRIBUTIONS AND HYPOTHESIS TESTING.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Unit 5: Hypothesis Testing.
Agresti/Franklin Statistics, 1 of 122 Chapter 8 Statistical inference: Significance Tests About Hypotheses Learn …. To use an inferential method called.
CHAPTER 17: Tests of Significance: The Basics
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Statistics In HEP Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
S-012 Testing statistical hypotheses The CI approach The NHST approach.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
Chapter 20 Testing Hypothesis about proportions
Chapter 21: More About Test & Intervals
Introduction to Statistical Inference A Comparison of Classical Orthodoxy with the Bayesian Approach.
Not in FPP Bayesian Statistics. The Frequentist paradigm Defines probability as a long-run frequency independent, identical trials Looks at parameters.
CHAPTER 15: Tests of Significance The Basics ESSENTIAL STATISTICS Second Edition David S. Moore, William I. Notz, and Michael A. Fligner Lecture Presentation.
CHAPTER 9 Testing a Claim
Populations III: evidence, uncertainty, and decisions Bio 415/615.
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
Chapter 8 Hypothesis Testing I. Significant Differences  Hypothesis testing is designed to detect significant differences: differences that did not occur.
Stats Lunch: Day 3 The Basis of Hypothesis Testing w/ Parametric Statistics.
AP Statistics Section 11.1 B More on Significance Tests.
Inferential Statistics Inferential statistics allow us to infer the characteristic(s) of a population from sample data Slightly different terms and symbols.
AP Statistics Chapter 11 Notes. Significance Test & Hypothesis Significance test: a formal procedure for comparing observed data with a hypothesis whose.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Unit 5: Hypothesis Testing.
Chapter 13 Understanding research results: statistical inference.
Uncertainty and confidence Although the sample mean,, is a unique number for any particular sample, if you pick a different sample you will probably get.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 9 Testing a Claim 9.1 Significance Tests:
CHAPTER 15: Tests of Significance The Basics ESSENTIAL STATISTICS Second Edition David S. Moore, William I. Notz, and Michael A. Fligner Lecture Presentation.
Section 9.1 First Day The idea of a significance test What is a p-value?
Fundamentals of Data Analysis Lecture 4 Testing of statistical hypotheses pt.1.
Philosophy of science What is a scientific theory? – Is a universal statement Applies to all events in all places and time – Explains the behaviour/happening.
FIXETH LIKELIHOODS this is correct. Bayesian methods I: theory.
+ Testing a Claim Significance Tests: The Basics.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 9 Testing a Claim 9.1 Significance Tests:
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 21 More About Tests and Intervals.
Unit 5: Hypothesis Testing
Hypothesis Testing: Hypotheses
Null Hypothesis Testing
Significance Tests: The Basics
Wellcome Trust Centre for Neuroimaging
Presentation transcript:

Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove that a theory is false by finding contradicting evidence but, cannot confirm a theory (may find a better theory or a falsification) 2) Statistical methods of Fisher – everything you have learned this semester

Bayesian inference: 1) Requires explicit assignment of prior probabilities, based on existing information, to the outcomes of experiments. 2) Allows for assignment of probabilities of specific outcomes

The frequentist assumption: there is a true, fixed value for parameters of interest (eg. ave weight of yellow perch in Lake Erie, ave height of college students) and the expected value of the parameter equals the average value obtained by random sampling (infinitely repeated). Some problem: -populations change constantly, so nothing really fixed -truly random sample hard to obtain (convenience samples) -experiments rarely repeated

Confidence in (frequentist) parameter estimates: -K% confidence interval around Xbar (usually 95%) -Tells us that  will lie within (Xbar  1.96SE) in 95% of the infinite number of possible samples we could collect -It does not state that there is a 95% probability that the true mean does occur in the confidence interval

Frequentist hypothesis testing: --The P-value of a statistical test is the probability of observing the given result conditional on H 0.  = P(x|H 0 ) = probability of x, given H 0 -- The P-value does not say how probable the null is given the data i.e., P(H 0 |x) --”Proving” the null is false, does not prove the alternative to be true.

P (H A |x) Probability of the alternative hypothesis given the observed data Likelihood that the treatment caused the effect (for experiments) What we really want to know!!

Many researchers consider that a P-value < 0.05 means that P (H A |x) is high, in other words they think it likely that the alternative hypothesis is true …… when they really tested P (x|H 0 ).

Bayesian parameter estimation: Begins with the joint probability of two events being equal to the probability of the first event and the conditional probability of the second event, given the first event. It is usually expressed: P (A|B) = P(B|A) * P(A) P(B) Bayes’ Theorem

(b. 1702, London - d. 1761, Tunbridge Wells, Kent), mathematician who first used probability inductively and established a mathematical basis for probability inference (a means of calculating, from the number of times an event has not occurred, the probability that it will occur in future trials). He set down his findings on probability in "Essay Towards Solving a Problem in the Doctrine of Chances" (1763), published posthumously in the Philosophical Transactions of the Royal Society of London. Rev Thomas Bayes

Consider 2 racehorses: --Buttercup and Muffin --Buttercup has won 7 of the last 12 races (53.8%) between the two horses --Muffin has won 5 of the last 12 (41.7%) --So…… there is a greater likelihood of Buttercup winning, right? Unless, it is raining……….

--On 3 of Muffin’s previous 5 wins, it was raining --So, if it is raining, you could estimate Muffin’s chance of a win at 3/5 or 60% --But then you ignore the fact that, overall Buttercup has won more often. --You must combine the two pieces of information --Look at 4 possible situations

RainingNot raining Muffin wins32 Muffin looses16 --If it is raining on the day you place your bet, you want to know the probability of Muffin wining in the rain. -- # times X happened / # time X could have happened --3 / 4 = 75% chance of Muffin winning

P (Muffin wins|rain) = P (rain|Muffin wins) * P (Muffin wins) It was raining on 3 of 5 days Muffin won, therefore, P (rain|Muffin wins) = 3/5=0.6 P (Muffin wins) = 5/12 = P (rain) = 4/12 = P (A|B) = P(B|A) * P(A) P(B) P (rain)

P (Muffin wins|rain) = P (rain|Muffin wins) * P (Muffin wins) P (rain) P (Muffin wins|rain) = 0.6 * / = 0.758

P (A|B) = P(B|A) * P(A) P(B) --In Bayesian terminology P(A) is the prior probability of obtaining a specific parameter --P(A) is the probability of observing A expected by the investigator before the experiment is conducted -- existing data --objective statement without data --subjective measure of belief --Priors in ecology usually based on previous data, information usually reported in the introduction of papers.

Non-mathematical explanation of prior probabilities --You want to estimate the whole season batting average of a Player X based on the first 2 weeks of the season --Player X batted.550 during the first two weeks --Any player is very unlikely to bat >.500 for the entire season -- The players, or the entire league’s batting average from the last year could be used as a prior probability -- Common sense dictates that Player X’s average will tend down from.550 and toward last season’s total average

--prior probabilities can be “non informative” in the absence of any inforamtion --in that case P(A) is a uniform distribution, with all values equally likely --Posterior probability is what you are trying to figure out, the probability of the event in which you are interested.

Bayesian Hypothesis testing: --Can generalize; Bayes theorem can be extended to assess the relative probabilities of alternative hypotheses (alternative prior probability distributions) --Can determine the likelihood of an hypothesis posterior odds = prior odds * Bayes factor

-- different scales of Bayes factor proposed to say whether the data favor H A --More intuitive and satisfying than frequentist statistics, but harder to do! -- Math harder (integration) and little software currently available

?? As N increases, it becomes less likely to accept the Null at a fixed alpha level. What guidelines should we use to adjust alpha as sample size increases? ?? How can policy makers use a rejection of a null hypothesis at the p=0.05 level to make a decision? What does rejection at the 0.05 level mean anyway? ?? Why do researcher act as if they have tested P (H A |x) when they really tested P (x|H 0 )? Do most researchers realize what they are testing? Or do they realize and not care….. They just want to support that ol’ alternative hypothesis? ?? What will motivate statisticians to write easy to use Bayesian software? All students be prepared to answer these questions