Maximum Likelihood Find the parameters of a model that best fit the data… Forms the foundation of Bayesian inference Slide 1.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Chapter 2 Describing Contingency Tables Reported by Liu Qi.
Special random variables Chapter 5 Some discrete or continuous probability distributions.
INTRODUCTION TO MACHINE LEARNING Bayesian Estimation.
Log-linear and logistic models Generalised linear model ANOVA revisited Log-linear model: Poisson distribution logistic model: Binomial distribution Deviances.
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
Flipping A Biased Coin Suppose you have a coin with an unknown bias, θ ≡ P(head). You flip the coin multiple times and observe the outcome. From observations,
Descriptive statistics Experiment  Data  Sample Statistics Sample mean Sample variance Normalize sample variance by N-1 Standard deviation goes as square-root.
Basics of Statistical Estimation. Learning Probabilities: Classical Approach Simplest case: Flipping a thumbtack tails heads True probability  is unknown.
Statistics Lecture 11.
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference (Sec. )
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
A random variable that has the following pmf is said to be a binomial random variable with parameters n, p The Binomial random variable.
458 Fitting models to data – III (More on Maximum Likelihood Estimation) Fish 458, Lecture 10.
Chapter 21 Random Variables Discrete: Bernoulli, Binomial, Geometric, Poisson Continuous: Uniform, Exponential, Gamma, Normal Expectation & Variance, Joint.
Chapter Two Probability Distributions: Discrete Variables
Moment Generating Functions 1/33. Contents Review of Continuous Distribution Functions 2/33.
The Triangle of Statistical Inference: Likelihoood
Random Variables & Probability Distributions Outcomes of experiments are, in part, random E.g. Let X 7 be the gender of the 7 th randomly selected student.
Modeling and Simulation CS 313
Statistics for Engineer Week II and Week III: Random Variables and Probability Distribution.
Moment Generating Functions
Probability Distributions and Dataset Properties Lecture 2 Likelihood Methods in Forest Ecology October 9 th – 20 th, 2006.
The Triangle of Statistical Inference: Likelihoood Data Scientific Model Probability Model Inference.
Methodology Solving problems with known distributions 1.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
Probability Refresher COMP5416 Advanced Network Technologies.
1 3. Random Variables Let ( , F, P) be a probability model for an experiment, and X a function that maps every to a unique point the set of real numbers.
Dirichlet Distribution
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 12 Tests of Goodness of Fit and Independence n Goodness of Fit Test: A Multinomial.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
Introduction A probability distribution is obtained when probability values are assigned to all possible numerical values of a random variable. It may.
Chapter 6 – Continuous Probability Distribution Introduction A probability distribution is obtained when probability values are assigned to all possible.
Crash course in probability theory and statistics – part 2 Machine Learning, Wed Apr 16, 2008.
Probability distributions and likelihood
Theme 8. Major probability distributions
MECH 373 Instrumentation and Measurements
Chapter 14 Fitting Probability Distributions
CHAPTER 12 MODELING COUNT DATA: THE POISSON AND NEGATIVE BINOMIAL REGRESSION MODELS Damodar Gujarati Econometrics by Example, second edition.
Statistical Modelling
IEE 380 Review.
Chapter 4. Inference about Process Quality
Dr.MUSTAQUE AHMED MBBS,MD(COMMUNITY MEDICINE), FELLOWSHIP IN HIV/AIDS
Probability for Machine Learning
CS 2750: Machine Learning Density Estimation
Maximum Likelihood Estimation
Generalized Linear Models
CONCEPTS OF HYPOTHESIS TESTING
SA3202 Statistical Methods for Social Sciences
Chapter 3 Discrete Random Variables and Probability Distributions
Goodness-of-Fit Tests
Probability & Statistics Probability Theory Mathematical Probability Models Event Relationships Distributions of Random Variables Continuous Random.
More about Posterior Distributions
Probability Review for Financial Engineers
Econometric Models The most basic econometric model consists of a relationship between two variables which is disturbed by a random error. We need to use.
Discrete Event Simulation - 4
TESTING HYPOTHESES AND ASSESSING GOODNESS OF FIT
Some Discrete Probability Distributions
Statistical NLP: Lecture 4
Categorical Data Analysis
3. Random Variables Let (, F, P) be a probability model for an experiment, and X a function that maps every to a unique point.
Chapter 3 : Random Variables
Random Variables A random variable is a rule that assigns exactly one value to each point in a sample space for an experiment. A random variable can be.
Berlin Chen Department of Computer Science & Information Engineering
Maximum Likelihood We have studied the OLS estimator. It only applies under certain assumptions In particular,  ~ N(0, 2 ) But what if the sampling distribution.
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Inference Concepts 1-Sample Z-Tests.
Maximum Likelihood Estimation (MLE)
CHAPTER 2.1 PROBABILITY DISTRIBUTIONS.
CHAPTER 2.1 PROBABILITY DISTRIBUTIONS.
Presentation transcript:

Maximum Likelihood Find the parameters of a model that best fit the data… Forms the foundation of Bayesian inference Slide 1

Distributions of Discrete Variables Random variables (the observed data) Discrete Are integer values Example: Binomial Multinomial Poisson Negative binomial

Distributions of continuous Variables Random variables are continuous Example: Gaussian (normal) Log normal Gamma Beta

PMF of Poisson Probability mass function (PMF) is a function that gives the probability that a discrete random variable is exactly equal to some value P(Yi = k | rate parameter = r) = 𝑒 −𝑟 𝑟 𝑘 𝑘!

PMF of Poisson In one unit of time we predict that Yi = k P(Yi = k | rate parameter = r) = 𝑒 −𝑟 𝑟 𝑘 𝑘!

Likelihood P(Yi | p) Probability distribution of observing data Yi, given a particular parameter value, p Subscript on Y indicates that there are many possible outcomes but only one possible parameter. Slide 7

Likelihood P(Yi = k | rate parameter = r) = 𝑒 −𝑟 𝑟 𝑘 𝑘! This expression is the probability of “data” given the hypothesis. Data are k events in one unit time Hypothesis is that the rate parameter is r Slide 8

Likelihood P(Yi = k | rate parameter = r) = 𝑒 −𝑟 𝑟 𝑘 𝑘! After collection of the data, the data are known. Alternative hypotheses are different values of r. Given the data, how likely are the possible hypotheses? Slide 9

Likelihood P(Yi = k | rate parameter = r) = 𝑒 −𝑟 𝑟 𝑘 𝑘! Introduce symbol: “L” likelihood L(data | hypothesis), L(Y | pm) Shift in thinking – m alternative parameters… One set of data Slide 10

Likelihood P(Yi = k | rate parameter = r) = 𝑒 −𝑟 𝑟 𝑘 𝑘! Difference in likelihood and probability: Probability: the hypothesis is known, data are unknown Likelihood: data are known, hypothesis is not known Slide 11

Likelihood in practice Generate data Determine range of parameter values that are alternative hypotheses Determine the probability that the data came from a distribution with a given parameter value

Likelihood in practice Generate data

Likelihood in practice Determine range of parameter values that are alternative hypotheses best.guess.mu <- seq(15,25,by = 0.1) best.guess.sig <- 5

Likelihood in practice Determine the probability that the data came from a distribution with a given parameter value