A Two Level Monte Carlo Approach To Calculating

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

INTRODUCTION TO MACHINE LEARNING Bayesian Estimation.
Some methodological issues in value of information analysis: an application of partial EVPI and EVSI to an economic model of Zanamivir Karl Claxton and.
Bayesian Estimation in MARK
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Cost-effectiveness models to inform trial design: Calculating the expected value of sample information Alan Brennan and J Chilcott, S Kharroubi, A O’Hagan.
Propagation of Error Ch En 475 Unit Operations. Quantifying variables (i.e. answering a question with a number) 1. Directly measure the variable. - referred.
The uptake of value of information methods Solutions found and challenges to come Alan Brennan Director of Operational Research ScHARR.
What role should probabilistic sensitivity analysis play in SMC decision making? Andrew Briggs, DPhil University of Oxford.
Valuing Trial Designs from a Pharmaceutical Perspective using Value Based Pricing (VBP) Penny Watson 1, Alan Brennan 1 1 Health Economics and Decision.
Visual Recognition Tutorial
Sensitivity Analysis for Complex Models Jeremy Oakley & Anthony O’Hagan University of Sheffield, UK.
USE OF LAPLACE APPROXIMATIONS TO SIGNIFICANTLY IMPROVE THE EFFICIENCY
Methods to Analyse The Economic Benefits of a Pharmacogenetic (PGt) Test to Predict Response to Biologic Therapy in Rheumatoid Arthritis, and to Prioritise.
Chapter 7 Sampling and Sampling Distributions
Basics of Statistical Estimation. Learning Probabilities: Classical Approach Simplest case: Flipping a thumbtack tails heads True probability  is unknown.
Results Doubling the sample size from n=50 to 100, increases the value of a 6 month study by 64% but the value of a 3 years study by only 34%. Similarly.
Evaluating Hypotheses
Results 2 (cont’d) c) Long term observational data on the duration of effective response Observational data on n=50 has EVSI = £867 d) Collect data on.
Value of Information Some introductory remarks by Tony O’Hagan.
1 Transforming the efficiency of Partial EVSI computation Alan Brennan Health Economics and Decision Science (HEDS) Samer Kharroubi Centre for Bayesian.
Non-parametric Bayesian value of information analysis Aim: To inform the efficient allocation of research resources Objectives: To use all the available.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Continuous Random Variables and Probability Distributions
Value of Information for Complex Economic Models Jeremy Oakley Department of Probability and Statistics, University of Sheffield. Paper available from.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Maximum likelihood (ML)
Lecture II-2: Probability Review
RLSELE Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Gaussian process modelling
Calibration and Model Discrepancy Tony O’Hagan, MUCM, Sheffield.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Model Inference and Averaging
Bayesian Inference Ekaterina Lomakina TNU seminar: Bayesian inference 1 March 2013.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Copyright © 2009 Cengage Learning Chapter 10 Introduction to Estimation ( 추 정 )
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Exam I review Understanding the meaning of the terminology we use. Quick calculations that indicate understanding of the basis of methods. Many of the.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
1 ECE-517 Reinforcement Learning in Artificial Intelligence Lecture 7: Finite Horizon MDPs, Dynamic Programming Dr. Itamar Arel College of Engineering.
LECTURER PROF.Dr. DEMIR BAYKA AUTOMOTIVE ENGINEERING LABORATORY I.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Conjugate Priors Multinomial Gaussian MAP Variance Estimation Example.
17 May 2007RSS Kent Local Group1 Quantifying uncertainty in the UK carbon flux Tony O’Hagan CTCD, Sheffield.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
BCS547 Neural Decoding.
Bayesian Prior and Posterior Study Guide for ES205 Yu-Chi Ho Jonathan T. Lee Nov. 24, 2000.
1 Optimizing Decisions over the Long-term in the Presence of Uncertain Response Edward Kambour.
- 1 - Calibration with discrepancy Major references –Calibration lecture is not in the book. –Kennedy, Marc C., and Anthony O'Hagan. "Bayesian calibration.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Statistics Presentation Ch En 475 Unit Operations.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Bayesian Estimation and Confidence Intervals Lecture XXII.
Crash course in probability theory and statistics – part 2 Machine Learning, Wed Apr 16, 2008.
Bayesian Estimation and Confidence Intervals
12. Principles of Parameter Estimation
Ch3: Model Building through Regression
Introduction to Instrumentation Engineering
Akio Utsugi National Institute of Bioscience and Human-technology,
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Lecture 3 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Introduction to Estimation
Learning From Observed Data
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions July chonbuk national university.
12. Principles of Parameter Estimation
Presentation transcript:

A Two Level Monte Carlo Approach To Calculating Expected Value of Sample Information:- How to Value a Research Design. Alan Brennan and Jim Chilcott, Samer Kharroubi, Tony O’Hagan University of Sheffield IHEA June 2003

What are EVPI and EVSI? Adoption Decision Choose policy with greatest expected ‘payoff’ Uncertainty Confidence interval e.g. |--------[]---------| implies Adoption Decision could be wrong Perfect Information perfect knowledge about parameter e.g. [] Value of Information quantify additional payoff obtained by switching adoption decision after obtaining additional data Expected Value of Information quantify average additional payoff obtained for range of possible collected data EVPI Expected value of obtaining perfect knowledge EVSI Expected value of obtaining a specific sample size

Typical Process Model Net benefit of alternative policies Probabilistic Sensitivity Analysis e.g. 1,000 samples … C-E plane CEACs Partial EVPIs identify single or groups of parameters Partial EVSIs propose specific data collections (n) calculate expected value of different samples

 = uncertain model parameters d = set of possible decisions NB(d, ) = net benefit (λ*QALY – Cost) for decision d, parameters  i = parameters of interest – possible data collection -i = other parameters (not of interest, remaining uncertainty) Expected net benefit (1) Baseline decision = (2) Perfect Information on i = two expectations Partial EVPI = (2) – (1) (3) Sample Information on i = Partial EVSI = (3) – (1)

Expected Value of Sample Information 0)Decision model, threshold, priors for uncertain parameters 1) Simulate data collection: sample parameter(s) of interest once ~ prior decide on sample size (ni) (1st level) sample the simulated data | parameter of interest 2) combine prior + simulated data --> simulated posterior 3) now simulate 1000 times parameters of interest ~ simulated posterior unknown parameters ~ prior uncertainty (2nd level) 4) calculate best strategy = highest mean net benefit 5) Loop 1 to 4 say 1,000 times Calculate average net benefits 6) EVSI parameter set = (5) - (mean net benefit | current information) e.g. 1000 * 1000 simulations

Bayesian Updating: Normal Prior 0= mean, 0= uncertainty in mean (standard deviation) = precision of the prior mean 2pop = patient level uncertainty ( needed for update formula) Simulated Data   = sample mean (further data collection e.g. clinical trial ). = sample variance = precision of the sample mean . Simulated Posterior N(1, 1 )

Normal Posterior Variance – Implications . 1 always < 0 If n is very small, 1 = almost 0 If n is very large, 1 = almost 0

Bayesian Updating: Beta / Binomial e.g. % responders to a treatment Prior % responders ~ Beta (a,b) Simulated Data n cases, y successful responders Simulated Posterior % responders ~ Beta (a+y,b+n-y)

Bayesian Updating: Gamma / Poisson e.g. no. of side effects a patient experiences in a year Prior side effects per person ~ Gamma (a,b) Simulated Data n samples, (y1, y2, … yn) from a Poisson distribution Simulated Posterior mean side effects per person ~ Gamma (a+ yi , b+n)

Bayesian Updating without a Formula WinBUGS Put in prior distribution Put in data MCMC gives posterior (‘000s of iterations) Use posterior in model Loop to next data sample Other approximation methods Conjugate Distributions

EVSI Results for Illustrative Model

Common Properties of EVSI curve Fixed at zero if no sample is collected Bounded above by EVPI, monotonic, diminishing returns ? EVSI (n) = EVPI * [1 – exp -a*sqrt(n) ]

Correct 2 level EVPI Algorithm 0)Decision model, threshold, priors for uncertain parameters 1) Simulate data collection: sample parameter(s) of interest once ~ prior decide on sample size (ni) (1st level) sample the simulated data | parameter of interest 2) combine prior + simulated data --> simulated posterior 3) now simulate 1000 times parameters of interest ~ simulated posterior unknown parameters ~ prior uncertainty (2nd level) 4) calculate best strategy = highest mean net benefit 5) Loop 1 to 4 say 1,000 times Calculate average net benefits 6) EVPI parameter set = (5) - (mean net benefit | current information) fixed at sampled value e.g. 1000 * 1000 simulations

Shortcut 1 level EVPI Algorithm 1) Simulate data collection: sample parameter(s) of interest once ~ prior decide on sample size (ni) (1st level) sample the simulated data | parameter of interest 2) combine prior + simulated data --> simulated posterior 3) now simulate 1000 times parameters of interest ~ simulated posterior unknown parameters ~ prior uncertainty (2nd level) 4) calculate best strategy = highest mean net benefit 5) Loop 1 to 4 say 1,000 times Calculate average net benefits 6) EVPI parameter set = (5) - (mean net benefit | current information) 2) fix remaining unknown parameters constant at prior mean value Accurate if .. (a) net benefit functions are linear functions of the -i for all d,i, and (b) i and -i are independent.

How many samples for convergence ? outer level - over 500 samples, converges to within 1% inner level - required 10,000 samples, to converge to within 2%.

How wrong is US 1 level EVPI approach for a non-linear model? E.g – squared every model parameter Adjusted R2 for simple linear regression = 0.86 Linear Model Non-linear Model

Maximum of Monte Carlo Expectations:- Upward Bias simple Monte Carlo estimates are unbiased But, bias occurs when use maximum of Monte Carlo estimates. E{max(X,Y)} > max{E(X),E(Y)}. E.g. X ~ N(0,1), Y ~ N(0,1) E{max(X,Y)} = 1.2 , max{E(X),E(Y)} = 1.0 If variances of X and Y reduces to 0.5, E{max(X,Y)} = 1.08 E{max(X,Y)} continues to fall variance reduces, but stays > 1.0 Illustrative model:- 1000*1000 not enough to eliminate bias. Increasing no. of samples reduces bias, since it reduces the variances of the Monte Carlo estimates.

Computation Issues: Emulators Gaussian Processes (Jeremy Oakley, CHEBS) F() ~ NB(d, ) Bayesian non-linear regression (complicated maths) Assumes only a smooth functional form to NB(d, ) Benefits Can emulate complex time consuming models with formula i.e. speed up each sample Can produce a quick approximation to inner expectation for partial EVPI Similar quick approximation for partial EVSI but only for one parameter

Summary 2 level algorithm for EVPI and EVSI is correct approach Bayesian Updating for Normal, Beta, Gamma OK Others – WinBUGS / approximations There are issues of computation Shortcut 1 level EVPI algorithm is accurate if … net benefit functions are linear and the parameters are independent. Emulators (e.g. Gaussian Processes) can be helpful