- 1 - 3.7 Example: Bioassay experiment Problem statement –Observations: At each level of dose, 5 animals are tested, and number of death are observed.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

NORMAL OR GAUSSIAN DISTRIBUTION Chapter 5. General Normal Distribution Two parameter distribution with a pdf given by:
Copula Regression By Rahul A. Parsa Drake University &
Bayesian inference of normal distribution
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Sampling Distributions (§ )
Objectives (BPS chapter 24)
Descriptive statistics Experiment  Data  Sample Statistics Sample mean Sample variance Normalize sample variance by N-1 Standard deviation goes as square-root.
Chapter 10 Simple Regression.
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Development of Empirical Models From Process Data
Chapter 11 Multiple Regression.
Computer vision: models, learning and inference
Maximum likelihood (ML)
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Lecture II-2: Probability Review
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
1 As we have seen in section 4 conditional probability density functions are useful to update the information about an event based on the knowledge about.
Probability theory 2 Tron Anders Moger September 13th 2006.
Random Sampling, Point Estimation and Maximum Likelihood.
Bayesian inference review Objective –estimate unknown parameter  based on observations y. Result is given by probability distribution. Bayesian inference.
Exam I review Understanding the meaning of the terminology we use. Quick calculations that indicate understanding of the basis of methods. Many of the.
Basic Probability (Chapter 2, W.J.Decoursey, 2003) Objectives: -Define probability and its relationship to relative frequency of an event. -Learn the basic.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
MEGN 537 – Probabilistic Biomechanics Ch.5 – Determining Distributions and Parameters from Observed Data Anthony J Petrella, PhD.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Multiple Random Variables Two Discrete Random Variables –Joint pmf –Marginal pmf Two Continuous Random Variables –Joint Distribution (PDF) –Joint Density.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Simulation techniques Summary of the methods we used so far Other methods –Rejection sampling –Importance sampling Very good slides from Dr. Joo-Ho Choi.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
Confidence Intervals vs. Prediction Intervals Confidence Intervals – provide an interval estimate with a 100(1-  ) measure of reliability about the mean.
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Generalized Linear Models (GLMs) and Their Applications.
Estimation Method of Moments (MM) Methods of Moment estimation is a general method where equations for estimating parameters are found by equating population.
Sampling and estimation Petter Mostad
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 11: Models Marshall University Genomics Core Facility.
Ch4: 4.3The Normal distribution 4.4The Exponential Distribution.
- 1 - Matlab statistics fundamentals Normal distribution % basic functions mew=100; sig=10; x=90; normpdf(x,mew,sig) 1/sig/sqrt(2*pi)*exp(-(x-mew)^2/sig^2/2)
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.
Unit 4 Review. Starter Write the characteristics of the binomial setting. What is the difference between the binomial setting and the geometric setting?
MEGN 537 – Probabilistic Biomechanics Ch.5 – Determining Distributions and Parameters from Observed Data Anthony J Petrella, PhD.
1 Estimation Chapter Introduction Statistical inference is the process by which we acquire information about populations from samples. There are.
WARM UP: Penny Sampling 1.) Take a look at the graphs that you made yesterday. What are some intuitive takeaways just from looking at the graphs?
Bayesian Estimation and Confidence Intervals Lecture XXII.
Chapter 4 Basic Estimation Techniques
Bayesian Estimation and Confidence Intervals
Probability Theory and Parameter Estimation I
Inference: Conclusion with Confidence
Ch3: Model Building through Regression
Analyzing Redistribution Matrix with Wavelet
Statistical Methods For Engineers
Estimating the Value of a Parameter
Simple Linear Regression
Lecture # 2 MATHEMATICAL STATISTICS
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves.
Parametric Methods Berlin Chen, 2005 References:
Sampling Distributions (§ )
HKN ECE 313 Exam 2 Review Session
Mathematical Foundations of BME Reza Shadmehr
Classical regression review
Presentation transcript:

Example: Bioassay experiment Problem statement –Observations: At each level of dose, 5 animals are tested, and number of death are observed. As dose x is increased, # of death increases. –Objective: estimate unknown probability of death  given dose x. Note that  is not a single parameter but varies w.r.t. x.

- 2 - Establishing posterior distribution Binomial distribution –Let the unknown death probability be  i for x i for ith group. –Then the data y i are binomially distributed: Modeling dose response relation –  is not constant but a function of dose x. Simplest trial is linear function. –This model has flaw that for very low or high x, it approaches ±∞. To meet this condition, introduce logistic transformation Then leads to –Relation is changed to This is called a logistic regression model.

- 3 - Establishing posterior distribution The likelihood –Likelihood of each individual group –Remark: model is characterized by two parameters , , not . Hence, these are the parameters to be estimated. Joint posterior pdf –Prior distribution is assumed uniform or non-informative, i.e., –Remark: can we write down the expression in closed form ? Never seen the pdf like this, nor are there standard pdfs available. Nevertheless, this is probability distribution indeed.

- 4 - Analyzing posterior distribution Rough estimate of the parameters –We can crudely estimate these by least squares minimization (linear regression) of z i on x i where i=1,…,4. The solution for ( ,  ) are (0.1,2.9). –We can also estimate standard error of ( ,  ) The solution for (V , V  ) are (0.3, 0.5). –This gives us rough range of the parameters.  ~ 0.1 ± 3*0.3 and  ~ 2.9 ± 3*0.5

- 5 - Analyzing posterior distribution Contour or 3-D plot of joint posterior density –Define the domain as mean ± 3 std errors, which are [0.1  0.9]=[-1,1] x [2.9  1.5]=[1,5]. –Use computer program to plot contour, and find out the range is wrong indicating that the crude estimation was really crude indeed. –After trial & error, find out much wider range is needed, [-5,10]x[-10,40]. Remarks –How can we analyze this distribution ? i.e., how can we get the mean, variance, confidence intervals, etc ?

- 6 - Sampling from posterior distribution Grid method (inverse cdf method) –Refer section 1.9 of Gelman or of Mahadevan. Procedure –In order to generate samples following pdf f(v), 1.Construct cdf F(v) which is the integral of f(v). 2.Draw random value U from the uniform distribution on [0,1]. 3.let v=F -1 (U). Then the value v will be a random draw from f(v). Practice –Generate samples for N(0.1). –Validate with analytic solution by comparing pdf shape mean, std, 95% conf bounds.

- 7 - Sampling from posterior distribution Factorization approach –Sample the two parameters  in the same way as the two parameters of normal distribution. where p(  |y): marginal pdf of  which is obtained by integration. p(  y): conditional pdf of b given . Procedure 1.Normalize the discrete joint pdf, i.e., make the total probability value 1. 2.Make marginal pdf p(  |y) by simply summing over . 3.Draw  from marginal pdf of  using grid method. 4.Draw  from conditional pdf of . 5.Post-analyze using the samples: As we have obtained samples of  , one can evaluate several characteristics.

- 8 - Sampling from posterior distribution Sample results –Scatter plot & comparison with contour plot –Posterior distribution of LD50, which is the dose level x at which the prob. death is 0.5. Simulating this is trivial.

- 9 - Posterior prediction by sampling Original definition –In this problem, probability of death is predicted at a dose level x. Practical solution –For each sample a & b, generate new y (which is prob. death q here). –By this way, simulation is much more convenient than the analytic (or numerical) integration.

Remarks on grid based method Remarks –There can be difficulty finding correct location & scale for the grid points. If the area is too small, it can miss the important area. If the area is too large, it can miss the importance due to large interval. –When computing posterior pdf, overflow / underflow encountered usually because it is multiplication of individual pdfs. To avoid, introduce log to the posterior density. –Computation grow prohibitly as more parameters are included. 500 grid for a parameter leads to 500^4 = 6.25e10 numbers for the whole grid, at which the function value should be evaluated. Conclusion: this method is not used in practice.