1 Parametric Sensitivity Analysis For Cancer Survival Models Using Large- Sample Normal Approximations To The Bayesian Posterior Distribution Gordon B.

Slides:



Advertisements
Similar presentations
Bayesian inference Lee Harrison York Neuroimaging Centre 01 / 05 / 2009.
Advertisements

Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Pattern Recognition and Machine Learning
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
Bayesian Estimation in MARK
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
BAYESIAN POSTERIOR DISTRIBUTIONS FOR PROBABILISTIC SENSITIVITY ANALYSIS Gordon B. Hazen and Min Huang, IEMS Department, Northwestern University, Evanston.
Visual Recognition Tutorial
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Classification and risk prediction
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 10 Statistical Modelling Martin Russell.
USE OF LAPLACE APPROXIMATIONS TO SIGNIFICANTLY IMPROVE THE EFFICIENCY
Clustering.
Presenting: Assaf Tzabari
Machine Learning CMPT 726 Simon Fraser University
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation X = {
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation Given.
Bayesian Analysis of Dose-Response Calibration Curves Bahman Shafii William J. Price Statistical Programs College of Agricultural and Life Sciences University.
Lecture II-2: Probability Review
Survival Analysis A Brief Introduction Survival Function, Hazard Function In many medical studies, the primary endpoint is time until an event.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Topic Models in Text Processing IR Group Meeting Presented by Qiaozhu Mei.
The Triangle of Statistical Inference: Likelihoood
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
علي اكبر حقدوست، اپيدميولوژيست دانشگاه علوم پزشكي كرمان.
Geo479/579: Geostatistics Ch12. Ordinary Kriging (1)
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Bayesian Analysis and Applications of A Cure Rate Model.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
INTRODUCTION TO Machine Learning 3rd Edition
Week 41 How to find estimators? There are two main methods for finding estimators: 1) Method of moments. 2) The method of Maximum likelihood. Sometimes.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Multilevel and multifrailty models. Overview  Multifrailty versus multilevel Only one cluster, two frailties in cluster e.g., prognostic index (PI) analysis,
Machine Learning 5. Parametric Methods.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
Introduction We consider the data of ~1800 phenotype measurements Each mouse has a given probability distribution of descending from one of 8 possible.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Estimating standard error using bootstrap
Applied statistics Usman Roshan.
Probability Theory and Parameter Estimation I
Parameter Estimation 主講人:虞台文.
Special Topics In Scientific Computing
Latent Variables, Mixture Models and EM
Bayesian Inference for Small Population Longevity Risk Modelling
CONCEPTS OF HYPOTHESIS TESTING
More about Posterior Distributions
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Parametric Estimation
Pattern Recognition and Machine Learning
LECTURE 09: BAYESIAN LEARNING
Topic Models in Text Processing
Parametric Methods Berlin Chen, 2005 References:
CS639: Data Management for Data Science
Applied Statistics and Probability for Engineers
Probabilistic Surrogate Models
Presentation transcript:

1 Parametric Sensitivity Analysis For Cancer Survival Models Using Large- Sample Normal Approximations To The Bayesian Posterior Distribution Gordon B. Hazen, PhD Northwestern University

2 Cancer survival models as components of many analyses Col et al. (2002), “Survival impact of tamoxifen use for breast cancer risk reduction”, Medical Decision Making 22,

3 A simple cancer survival model – the Conditional Cure model p = probability of cure  = mortality rate if not cured Survival function:

4 Fitting CCure Model to DATA

5 Fitting CCure Model to DATA (cont.)

6

7

8

9 Question It is easy to choose p,  to fit a Conditional Cure survival curve to SEER survival data, but … How should we conduct a sensitivity analysis on the resulting estimates

10 The Bayesian approach Treat the unknowns p,  as random variables with specified prior distribution. Use Bayes’ rule to calculate the posterior distribution of p,  given SEER or other data. Use this posterior distribution to guide a sensitivity analysis, or to conduct a probabilistic sensitivity analysis.

11 Bayesian model with censoring Posterior distribution Posterior distribution is analytically awkward

12 Bayesian model with censoring Ovarian Cancer Stage II: Posterior on p, 

13 Bayesian model with censoring Awkward analytical form makes the posterior distribution on p,  difficult to use for sensitivity analysis: –Where is a 95% credible region? –How to generate random p,  for probabilistic sensitivity analysis? Solution: Large-sample Bayesian posterior distributions are approximately normal

14 Large-sample Bayesian posteriors Fundamental result: For large samples, the Bayesian posterior distribution is approximately multivariate normal with –mean equal to the posterior mode (under a uniform prior, this is the maximum likelihood estimate) –covariance matrix equal to the matrix inverse of the Hessian of the log-posterior evaluated at the posterior mode.

15 Hessian of the log posterior…

16 Hessian of the log posterior The Hessian is the matrix of second partial derivatives with respect to p and  :

17 Large-Sample Bayesian Posteriors Using Excel’s Solver to calculate mle and covariance matrix for p,  Value of log posterior at p,  Mle’s Hessian H p,  covariance matrix SEER data

18 Large-sample Bayesian posterior Ovarian Cancer Stage II True posterior density Approximate normal density

19 Two-way sensitivity analysis on p,  Vary p and  simultaneously two standard deviations along the principal component of the approximate normal posterior density.

20 Two-way sensitivity analysis on p,  cont.) The resulting variation in stage II ovarian cancer survival:

21 Two-way sensitivity analysis on p,  (cont.) The resulting variation in survival for a 50-year- old white female with stage II ovarian cancer:

22 Summary The Conditional Cure model for cancer survival. A method for using a large-sample normal approximation to the Bayesian posterior distribution to guide a sensitivity analysis of parameter estimates for this model. Appears to be a useful and practical method.

23 Potential Pitfalls Large-sample normal approximation requires mle to be an interior maximum – estimates p = 0, p = 1, or  = 0 do not yield approximate normal posteriors If sample size is very large, then posterior distribution will be so tight that sensitivity analysis is unnecessary.