Bayesian statistics So far we have thought of probabilities as the long term “success frequency”: #successes / #trails → P(success). In Bayesian statistics.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

A Tutorial on Learning with Bayesian Networks
Lecture XXIII.  In general there are two kinds of hypotheses: one concerns the form of the probability distribution (i.e. is the random variable normally.
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
Flipping A Biased Coin Suppose you have a coin with an unknown bias, θ ≡ P(head). You flip the coin multiple times and observe the outcome. From observations,
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
.. . Parameter Estimation using likelihood functions Tutorial #1 This class has been cut and slightly edited from Nir Friedman’s full course of 12 lectures.
Binomial Distribution & Bayes’ Theorem. Questions What is a probability? What is the probability of obtaining 2 heads in 4 coin tosses? What is the probability.
Parameter Estimation using likelihood functions Tutorial #1
This presentation has been cut and slightly edited from Nir Friedman’s full course of 12 lectures which is available at Changes.
Results 2 (cont’d) c) Long term observational data on the duration of effective response Observational data on n=50 has EVSI = £867 d) Collect data on.
. Maximum Likelihood (ML) Parameter Estimation with applications to reconstructing phylogenetic trees Comput. Genomics, lecture 6b Presentation taken from.
Thanks to Nir Friedman, HU
1 © Lecture note 3 Hypothesis Testing MAKE HYPOTHESIS ©
Additional Slides on Bayesian Statistics for STA 101 Prof. Jerry Reiter Fall 2008.
A quick intro to Bayesian thinking 104 Frequentist Approach 10/14 Probability of 1 head next: = X Probability of 2 heads next: = 0.51.
Class 3 Binomial Random Variables Continuous Random Variables Standard Normal Distributions.
Statistical Decision Theory
.. . Maximum Likelihood (ML) Parameter Estimation with applications to inferring phylogenetic trees Comput. Genomics, lecture 6a Presentation taken from.
BINOMIAL DISTRIBUTION Success & Failures. Learning Goals I can use terminology such as probability distribution, random variable, relative frequency distribution,
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Bayesian Analysis and Applications of A Cure Rate Model.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
1 Lecture note 4 Hypothesis Testing Significant Difference ©
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Statistical Inference Statistical Inference is the process of making judgments about a population based on properties of the sample Statistical Inference.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
Not in FPP Bayesian Statistics. The Frequentist paradigm Defines probability as a long-run frequency independent, identical trials Looks at parameters.
Problem: 1) Show that is a set of sufficient statistics 2) Being location and scale parameters, take as (improper) prior and show that inferences on ……
LECTURE 12 TUESDAY, 10 MARCH STA 291 Spring
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
Bayesian Prior and Posterior Study Guide for ES205 Yu-Chi Ho Jonathan T. Lee Nov. 24, 2000.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Trond Reitan (Division of statistics and insurance mathematics, Department of Mathematics, University of Oslo) Statistical modelling and latent variables.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
Statistics 300: Elementary Statistics Section 11-2.
CHAPTER 9 Inference: Estimation The essential nature of inferential statistics, as verses descriptive statistics is one of knowledge. In descriptive statistics,
Sampling considerations within Market Surveillance actions Nikola Tuneski, Ph.D. Department of Mathematics and Computer Science Faculty of Mechanical Engineering.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
The Uniform Prior and the Laplace Correction Supplemental Material not on exam.
Dirichlet Distribution
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Crash course in probability theory and statistics – part 2 Machine Learning, Wed Apr 16, 2008.
Sampling and Sampling Distributions
Oliver Schulte Machine Learning 726
Bayesian Estimation and Confidence Intervals
MCMC Output & Metropolis-Hastings Algorithm Part I
Presented by: Karen Miller
Ch3: Model Building through Regression
Bayes Net Learning: Bayesian Approaches
Binomial Distribution & Bayes’ Theorem
Tutorial #3 by Ma’ayan Fishelson
Lecture 11 Nonparametric Statistics Introduction
OVERVIEW OF BAYESIAN INFERENCE: PART 1
POINT ESTIMATOR OF PARAMETERS
Computing and Statistical Data Analysis / Stat 8
Statistical NLP: Lecture 4
Reversing Label Switching:
Example Human males have one X-chromosome and one Y-chromosome,
Important Distinctions in Learning BNs
LECTURE 07: BAYESIAN ESTIMATION
CS 594: Empirical Methods in HCC Introduction to Bayesian Analysis
Learning From Observed Data
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions July chonbuk national university.
CS639: Data Management for Data Science
MAS2317- Introduction to Bayesian Statistics
Mathematical Foundations of BME Reza Shadmehr
Presentation transcript:

Bayesian statistics So far we have thought of probabilities as the long term “success frequency”: #successes / #trails → P(success). In Bayesian statistics probabilities are subjective! Examples * Probability that two companies merge * Probability that a stock goes up * Probability that it rains tomorrow We typically want to make inference for a parameter q, for example m, s2 or p. How is this done using subjective probabilities? lecture 8

Bayesian statistics Bayesian idea: We describe our “knowledge” about the parameter of interest, q, in terms of a distribution p(q). This is known as the prior distriubtion (or just prior) – as it describes the situation before we see any data. Example: Assume q is the probability of success. Prior distributions describing what value we think q has: lecture 8

Bayesian statistics Posterior Let x denote our data. The conditional distribution of q given data x is denoted the posterior distribution: Here f(x|q) how data is specified conditional on q. Example: Let x denote the number of successes in n trail. Conditional q, x follows a binomial distribution: lecture 8

Bayesian statistics Posterior – some data We now observe n=10 experiment with x=3 successes, i.e. x/n=0.3 Posterior distributions – our “knowledge” after having seen data. Shaded area: Prior distribution Solid line: Posterior distribution Notice that the posteriors are moving towards 0.3. lecture 8

Bayesian statistics Posterior – some data We now observe n = 100 experiment with x = 30 successes, i.e. x/n = 0.3 Posterior distributions – our “knowledge” after having seen data. Shaded area: Prior distribution Solid line: Posterior distribution Notice that the posteriors are almost identical. lecture 8

Bayesian statistics Mathematical details The prior is given by a so-called Beta distribution with parameters a > 0 and b > 0: The posterior then becomes a Beta distribution with parameters a+x and b+n-x. lecture 8