Mixed latent Markov models for longitudinal multiple diagnostics data with an application to salmonella in Malawi Marc Henrion, Angeziwa Chirambo, Tonney.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Introduction to Monte Carlo Markov chain (MCMC) methods
MCMC estimation in MlwiN
Online Multi-camera Tracking with a Switiching State-Space Model Wojciech Zajdel, A. Taylan Cemgil, and Ben KrÄose ICPR 2004.
Randomized Complete Block and Repeated Measures (Each Subject Receives Each Treatment) Designs KNNL – Chapters 21,
Latent Variable and Structural Equation Models: Bayesian Perspectives and Implementation. Peter Congdon, Queen Mary University of London, School of Geography.
METHODS FOR HAPLOTYPE RECONSTRUCTION
Maximum Likelihood And Expectation Maximization Lecture Notes for CMPUT 466/551 Nilanjan Ray.
K Means Clustering , Nearest Cluster and Gaussian Mixture
Bayesian statistics – MCMC techniques
Visual Recognition Tutorial
Midterm Review. The Midterm Everything we have talked about so far Stuff from HW I won’t ask you to do as complicated calculations as the HW Don’t need.
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
Visual Recognition Tutorial
Thanks to Nir Friedman, HU
Mixture Modeling Chongming Yang Research Support Center FHSS College.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Modeling Menstrual Cycle Length in Pre- and Peri-Menopausal Women Michael Elliott Xiaobi Huang Sioban Harlow University of Michigan School of Public Health.
The horseshoe estimator for sparse signals CARLOS M. CARVALHO NICHOLAS G. POLSON JAMES G. SCOTT Biometrika (2010) Presented by Eric Wang 10/14/2010.
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Advanced Methods and Models in Behavioral Research – 2010/2011 AMMBR course design CONTENT METHOD Y is 0/1 conjoint analysis logistic regression multi-level.
Lecture 8: Generalized Linear Models for Longitudinal Data.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Conjugate Priors Multinomial Gaussian MAP Variance Estimation Example.
A Comparison of Two MCMC Algorithms for Hierarchical Mixture Models Russell Almond Florida State University College of Education Educational Psychology.
Bayesian statistics Probabilities for everything.
Probability and Measure September 2, Nonparametric Bayesian Fundamental Problem: Estimating Distribution from a collection of Data E. ( X a distribution-valued.
STA 216 Generalized Linear Models Meets: 2:50-4:05 T/TH (Old Chem 025) Instructor: David Dunson 219A Old Chemistry, Teaching.
Extending Group-Based Trajectory Modeling to Account for Subject Attrition (Sociological Methods & Research, 2011) Amelia Haviland Bobby Jones Daniel S.
Latent Class Regression Model Graphical Diagnostics Using an MCMC Estimation Procedure Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University
Simulation Study for Longitudinal Data with Nonignorable Missing Data Rong Liu, PhD Candidate Dr. Ramakrishnan, Advisor Department of Biostatistics Virginia.
Trond Reitan (Division of statistics and insurance mathematics, Department of Mathematics, University of Oslo) Statistical modelling and latent variables.
MPS/MSc in StatisticsAdaptive & Bayesian - Lect 71 Lecture 7 Bayesian methods: a refresher 7.1 Principles of the Bayesian approach 7.2 The beta distribution.
How does Biostatistics at Roche typically analyze longitudinal data
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Jump to first page Bayesian Approach FOR MIXED MODEL Bioep740 Final Paper Presentation By Qiang Ling.
How many iterations in the Gibbs sampler? Adrian E. Raftery and Steven Lewis (September, 1991) Duke University Machine Learning Group Presented by Iulian.
Prediction and Missing Data. Summarising Distributions ● Models are often large and complex ● Often only interested in some parameters – e.g. not so interested.
Crash course in probability theory and statistics – part 2 Machine Learning, Wed Apr 16, 2008.
Markov Chain Monte Carlo in R
Bayesian Semi-Parametric Multiple Shrinkage
Chapter 3: Maximum-Likelihood Parameter Estimation
M.Sc. in Economics Econometrics Module I
ICS 280 Learning in Graphical Models
Model Inference and Averaging
Introduction to the bayes Prefix in Stata 15
Statistical Models for Automatic Speech Recognition
STA 216 Generalized Linear Models
Latent Variables, Mixture Models and EM
Distributions and Concepts in Probability Theory
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
STA 216 Generalized Linear Models
Hidden Markov Models Part 2: Algorithms
More about Posterior Distributions
Multidimensional Integration Part I
'Linear Hierarchical Models'
Issues in Hypothesis Testing in the Context of Extrapolation
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Fixed, Random and Mixed effects
Learning From Observed Data
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions July chonbuk national university.
Bayesian Data Analysis in R
Bayesian Statistics on a Shoestring Assaf Oron, May 2008
Mathematical Foundations of BME Reza Shadmehr
Integration of inconsistent data sources using Hidden Markov Models (HMMs) Paulina Pankowska, Bart Bakker, Daniel Oberski & Dimitris Pavlopoulos.
Presentation transcript:

Mixed latent Markov models for longitudinal multiple diagnostics data with an application to salmonella in Malawi Marc Henrion, Angeziwa Chirambo, Tonney S. Nyirenda, Melita Gordon Malawi – Liverpool - Wellcome Trust Clinical Research Programme Liverpool School of Tropical Medicine mhenrion@mlw.mw JSM 2018 - Vancouver, Canada 30 July 2018 Don’t talk too much here!

Latent Markov Models Latent process, observed outcome variables at each time point Conditional on latent state, outcome variables assumed independent Salmonella data from Malawi: longitudinal data on multiple (binary) diagnostic tests no gold standard D1 D2 y1,1 y2,1 yk,1 … DT y1,2 y2,2 yk,2 y1,T y2,T yk,T TP1 TP2 TPT-1 CRP IP MEASUREMENT MODEL STRUCTURAL MODEL LMM = latent variable model Extends LCA to longitudinal data Structural model + measurement model Completely specified by IP, TP and CRP To keep maths tractable, typically assume independence of outcomes conditional on latent state Application to Salmonella data: 5 tests, 4 molecular tests, reference stool culture Conditional independence unlikely to hold: 4 tests use same PCR technology and 2-by-2 use the same primers

Relaxing the conditional independence assumption 𝑃 𝒀=𝒚;ϑ = 𝑖=1 𝑛 𝑑 𝑖 𝑃( 𝐷 𝑖 (1) = 𝑑 𝑖 (1) ) 𝑡=2 𝑇 1− 𝜏 01,𝑡 (1− 𝑑 𝑖 (𝑡) )(1− 𝑑 𝑖 (𝑡−1) ) 𝜏 01,𝑡 𝑑 𝑖 𝑡 (1− 𝑑 𝑖 (𝑡−1) ) 1− 𝜏 11,𝑡 (1− 𝑑 𝑖 (𝑡) ) 𝑑 𝑖 (𝑡−1) 𝜏 11,𝑡 𝑑 𝑖 (𝑡) 𝑑 𝑖 (𝑡−1) 𝑚=1 𝑘 𝜑 10,𝑘 𝑡 𝑦 𝑖𝑘 (𝑡) 1− 𝑑 𝑖 (𝑡) 1− 𝜑 10,𝑘 𝑡 1− 𝑦 𝑖𝑘 (𝑡) 1− 𝑑 𝑖 (𝑡) 𝜑 11,𝑘 𝑡 𝑦 𝑖𝑘 (𝑡) 𝑑 𝑖 𝑡 1− 𝜑 11,𝑘 𝑡 1− 𝑦 𝑖𝑘 (𝑡) 𝑑 𝑖 (𝑡) Initial state probabilities Conditional response probabilities 𝜑 𝑦𝑑,𝑚 =𝑃 𝑌 𝑚 =𝑦 | 𝑠𝑡𝑎𝑡𝑒=𝑑 Transition probabilities Sum over all possible latent state combinations Product over all patients Basic LMM Now add random effect to CRPs: 𝜑 𝑦𝑑,𝑖,𝑚 = 𝑒𝑥𝑝 𝛼 𝑑,𝑚 + 𝛽 𝑚 𝑍 𝑖 1+𝑒𝑥𝑝 𝛼 𝑑,𝑚 + 𝛽 𝑚 𝑍 𝑖 , 𝑍 𝑖 ~𝑁 0, 𝜎 2 Bayesian approach Frequentist much harder. Principled way of dealing with missing values. Other authors have considered mixed LMMs, but typically consider random effects in the latent process / structural model. Basic LMM likelihood: product over sum of latent states combinations of product of IP, TP and CRPs CRP term further factorises into product over outcome variables – this is what keeps the maths tractable Add subject specific random effects using a logistic function– trivial to add covariates into the CRPs Now you get large integrals to compute: Bayesian approach much easier, also better for missing data Could do similar extension to the transmission probas – but this has been done before

Simulations Assess convergence, identifiability issues, … Converges (up to label switching) Compare performance of basic and mixed LMMs: DIC, … basic LMM mixed LMM Simulations to check: Convergence, identifiability issues Compare performance: DIC, bias of MAP estimates, importance of prior, … Results: Yes, identifiable and converges – up to label switching typical for Bayesian mixture models; easily fixed after running MCMC Less biased MAP (or other) estimates

Application to Malawi salmonella data 4 new molecular PCR tests + stool culture 60 patients observed monthly for 12 months Results TTR primer PCR test achieves best sensitivity/specificity trade-off Data too sparse for time heterogeneous and mixed LMMs despite being biologically more plausible One specific PCR test is best: TTR Figure shows MAP estimates with 95% credible, highest posterior density intervals Unfortunately [embarrassingly?], even though this data motivated the model development, the more complex models: mixed, time heterogeneous do not converge [further work?] COME AND TALK TO ME THIS AFTERNOON!