October 28, 2008 Model qualification and assumption checking 1 To Validate or not to Validate? If all models are wrong but some.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Hypothesis Testing Steps in Hypothesis Testing:
Inference for Regression
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Correlation and regression
Mitigating Risk of Out-of-Specification Results During Stability Testing of Biopharmaceutical Products Jeff Gardner Principal Consultant 36 th Annual Midwest.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
The Simple Regression Model
Midterm Review Goodness of Fit and Predictive Accuracy
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 41 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Introduction to Probability and Statistics Linear Regression and Correlation.
AP Statistics Section 12.1 A. Now that we have looked at the principles of testing claims, we proceed to practice. We begin by dropping the unrealistic.
Inferences About Process Quality
Role and Place of Statistical Data Analysis and very simple applications Simplified diagram of a scientific research When you know the system: Estimation.
Business Statistics - QBM117 Statistical inference for regression.
Correlation and Regression Analysis
Linear Regression/Correlation
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
V. Rouillard  Introduction to measurement and statistical analysis ASSESSING EXPERIMENTAL DATA : ERRORS Remember: no measurement is perfect – errors.
Introduction to Linear Regression and Correlation Analysis
Inference for regression - Simple linear regression
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
On Model Validation Techniques Alex Karagrigoriou University of Cyprus "Quality - Theory and Practice”, ORT Braude College of Engineering, Karmiel, May.
Understanding Statistics
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
1 1 Slide Simple Linear Regression Part A n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n.
Chem. 31 – 9/23 Lecture Guest Lecture Dr. Roy Dixon.
CJT 765: Structural Equation Modeling Class 7: fitting a model, fit indices, comparingmodels, statistical power.
Two bootstrapping routines for obtaining uncertainty measurement around the nonparametric distribution obtained in NONMEM VI Paul G. Baverel 1, Radojka.
1 Chapter 10 Correlation and Regression 10.2 Correlation 10.3 Regression.
CS433: Modeling and Simulation Dr. Anis Koubâa Al-Imam Mohammad bin Saud University 15 October 2010 Lecture 05: Statistical Analysis Tools.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Measures of central tendency are statistics that express the most typical or average scores in a distribution These measures are: The Mode The Median.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Resampling techniques
BPS - 5th Ed. Chapter 221 Two Categorical Variables: The Chi-Square Test.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
1 Analysis of Variance & One Factor Designs Y= DEPENDENT VARIABLE (“yield”) (“response variable”) (“quality indicator”) X = INDEPENDENT VARIABLE (A possibly.
Reserve Variability – Session II: Who Is Doing What? Mark R. Shapland, FCAS, ASA, MAAA Casualty Actuarial Society Spring Meeting San Juan, Puerto Rico.
Population Pharmacokinetic Characteristics of Levosulpiride and Terbinafine in Healthy Male Korean Volunteers Yong-Bok Lee College of Pharmacy and Institute.
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
Three Frameworks for Statistical Analysis. Sample Design Forest, N=6 Field, N=4 Count ant nests per quadrat.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Case Selection and Resampling Lucila Ohno-Machado HST951.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Linear Regression Models Andy Wang CIS Computer Systems Performance Analysis.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
Bootstrapping James G. Anderson, Ph.D. Purdue University.
Model Comparison. Assessing alternative models We don’t ask “Is the model right or wrong?” We ask “Do the data support a model more than a competing model?”
Stats Methods at IC Lecture 3: Regression.
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
CJT 765: Structural Equation Modeling
Slides by JOHN LOUCKS St. Edward’s University.
Simple Linear Regression
3.2. SIMPLE LINEAR REGRESSION
St. Edward’s University
Presentation transcript:

October 28, 2008 Model qualification and assumption checking 1 To Validate or not to Validate? If all models are wrong but some are useful, why bother? To Qualify or not to Qualify? To Check or not to Check? To Evaluate or not to Evaluate? Pravin Jadhav, Pharmacometrics

October 28, Model qualification and assumption checking Model Qualification or Checking Qualification/Checking can be used to provide the following –feedback on how to improve the current model (learn), and/or –some reassurance that the model can at the least regenerate the data that were used to build the model (confirm).

October 28, Model qualification and assumption checking Model Qualification or Checking Before you go to quantitative qualification methods, (Qualitative)- –Use prior knowledge (nothing better than this) Properties of the drug Parameter estimates from previous studies or analyses (e.g. NonCompartmental analysis) Drugs in the same class In vitro data, etc. –Quality of the experimental design and data used in building the model –Specify model qualification objectives

October 28, Model qualification and assumption checking Precision of parameter estimates Log-likelihood profiling Bootstrap –Parametric –Nonparametric

October 28, Model qualification and assumption checking Log-Likelihood profiling: Principles Obj Fn CL 0 Maximum Likelihood Estimate Derive confidence interval for parameter estimates –Procedure Using the final model run, fix the value of parameter estimate to different values than maximum likelihood estimate Run estimation Plot difference in OBJ from the final model run against parameter –Automated In WFN, use nmllp runname paraname val1 val2…

October 28, Model qualification and assumption checking Log-Likelihood profiling Clearance estimate

October 28, Model qualification and assumption checking Log-Likelihood profiling This method is dependent on –-2*log-likelihood is chi-square distributed with 1 degree of freedom (one reduced parameter compared to full model) For some estimation methods (for example, FO) the assumption might not be accurate –Gobburu and Lawrence Pharm Res Jan;19(1):92-8 –Wahlby et. al. J Pharmacokinet Pharmacodyn Jun;28(3): What does this mean to you? How do you derive this?

October 28, Model qualification and assumption checking What does this mean? For most examples, if you use FO method- change of 3.84 in objective function value for the reduced model versus full model is probably not accurate for 95% significance level? If the exact p-value is needed –Do you need higher or lower change based on the previous graph? ???? How do you derive conditional distribution? If the exact p-value is needed, one will need to generate conditional distribution of the log-lileklihood and not rely on theoretical distribution (using randomization test, which will be not covered today) Generally speaking, the conditional distribution will be different for each combination of full model, reduced model and dataset (dense/sparse etc.) based on the approximations used-- But we don’t do it for every run- why?

October 28, Model qualification and assumption checking Bootstrap Wikipedia: bootstrapping or booting which began in the 1880s as a leather strap and evolved into a group of metaphors that share a common meaning, a self-sustaining process that proceeds without external help. Wikipedia: Bootstrapping is the practice of estimating properties of an estimator (such as its variance) by measuring those properties when sampling from an approximating distribution. –Smooth bootstrap –Parametric bootstrap –Case resampling (Non-parametric bootstrap) –Resampling residuals –Wild bootstrap

October 28, Model qualification and assumption checking Bootstrap Dataset 1000: Sample 100 subjects Dataset 1: Sample 100 subjects Non-parametric –Sample individuals to create several datasets from the original data Example from homework #3 Run E S T I M A T I O N 1000 Population estimates Original Data ID=9 ID=5 ID=45 ID=67 WFN use: nmbs runname 100 will call nmgo with 100 bootstrap sampled data sets taken from dataset supplied in runname

October 28, Model qualification and assumption checking Dataset 1000: using 100 CLi, Vi, ERRij and original data structure Dataset 1: using 100 CLi, Vi, ERRij and original data structure Bootstrap Parametric –Monte Carlo Simulations to create several datasets from the final model and model parameters One compartment example from homework #2 CL V Final model ESTIMATEESTIMATE 1000 Population estimates Error

October 28, Model qualification and assumption checking Bootstrap Parametric or Non-Parametric method will yield N (one for each successfully converged boostrap samples) sets of parameters –For example, N=1000 sets of CL, V,  2 CL,  2 V and  2

October 28, Model qualification and assumption checking Model Qualification or Checking Diagnostic plots (Slide 7 From Dr. Tornoe’s slides) –Observed and predicted concentration vs. time –Observed vs. predicted concentration –Residuals vs. time –Residuals vs. predictions Did you do this ever? –Homework #1, #2, #3

October 28, Model qualification and assumption checking Model Qualification or Checking: Observed vs. predicted concentration Homework #3 Student 1: There is no systemic bias (over-prediction or under-prediction) in individual predictions from the model; however, population predictions appear to be slightly under- predicted (biased). Student 2:The model was able to describe the data very well with no systematic bias. Student 3: Individual Predictions: The observed and predicted concentrations appear to be closely distributed around the line of identity suggesting minimal residual error and hence, the validity of the one-compartment model with first order absorption.

October 28, Model qualification and assumption checking Model Qualification or Checking: Residuals vs. time or predictions Homework #3 Student 1: Weighted residual are homogenously and randomly distributed around the line with zero mean without any trend, suggesting………………………. Student 2: Observation of the weighted residuals vs. time and vs. population predicted shows no heteroscedasticity.

October 28, Model qualification and assumption checking Model Qualification or Checking: Observed and predicted concentration vs. time Student 1: From a visual point of view, observed concentrations are well described by a one- cmpt body model with oral absorption.

October 28, Model qualification and assumption checking Model Qualification or Checking: Other plots There are several other plots one could make to test assumptions etc. We will not discuss these or others but the underlying message is to check assumptions that went into the model

October 28, Model qualification and assumption checking Model Qualification or Checking: Predictive check (Posterior) Predictive check is proposed to check whether the posited model should be excluded, because the model fails to provide a reasonable summary of the data used for modeling. Originally developed for checking fully Bayesian models. –The posterior distribution, a reflection of the uncertainty of a parameter, is influenced by the strength of the prior knowledge. Recollect Session 1 and today’s Q&A –Major question: How close the posterior distribution was to the current data? a summary feature, called a statistic (for example, SSE), calculated from the current data, are compared with the same statistic calculated under the posterior distribution. If this comparison failed to meet a prespecified criterion, the model might be rejected. Makes a lot of sense in Bayesian framework.

October 28, Model qualification and assumption checking Model Qualification or Checking: Predictive check Why (posterior) in parentheses –ML methods do not use priors –The ML approach yields only the point estimates of the parameters (called the ML estimates) and the asymptotic standard errors, and not a posterior distribution Yano Y, Beal SL, Sheiner LB. J Pharmacokinet Pharmacodyn. 2001;28: Jadhav, P. R.; Gobburu, J.V.S.; AAPS Journal, Vol. 7 No. 3 (2005)

October 28, Model qualification and assumption checking Model Qualification or Checking: Predictive check Three steps in PC or PPC –Estimation step y OD : Original data (For example, Drug Concentration)  : Estimated population parameters –What about 1 compartment model for IV administration –Simulation step y 1 rep ….. Y n rep are generated using  –Evaluation step Compare –test statistics T(y OD ) to T(y i rep ): mean concentration at time t and area under the curve (determined empirically) –Discrepancy variable T(y OD,  ) to T(y i rep,  ): sum of squared errors (SSE), determined using the observed and model- predicted variables (eg, concentrations), mean prediction and mean absolute prediction errors

October 28, Model qualification and assumption checking Evaluation step –graphical assessment of the 95% prediction interval (visual PC) –considerable scatter beyond the 95% prediction interval could indicate a poor model, but the converse may not be valid. Model Qualification or Checking: Predictive check

October 28, Model qualification and assumption checking Model Qualification or Checking: Predictive check –predictive p-value (Pp) What does Pp =0, 0.5 and 1 mean? –probability of equivalence (p eqv ),

October 28, Model qualification and assumption checking Predictive check: References Gelman A, Carlin JB, Stern HS, Rubin DB. Model checking and sensitivity analysis. In: Gelman A, ed. Bayesian Data Analysis. London: Chapman & Hall; 1995:

October 28, Model qualification and assumption checking Other approaches Internal validation External validation Sensitivity analysis