April 2008UW-Madison © Brian S. Yandell1 Bayesian QTL Mapping Brian S. Yandell University of Wisconsin-Madison www.stat.wisc.edu/~yandell/statgen↑↑ UW-Madison,

Slides:



Advertisements
Similar presentations
Markov Chain Monte Carlo Convergence Diagnostics: A Comparative Review By Mary Kathryn Cowles and Bradley P. Carlin Presented by Yuting Qi 12/01/2006.
Advertisements

Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
METHODS FOR HAPLOTYPE RECONSTRUCTION
Markov-Chain Monte Carlo
DATA ANALYSIS Module Code: CA660 Lecture Block 6: Alternative estimation methods and their implementation.
1 QTL mapping in mice Lecture 10, Statistics 246 February 24, 2004.
Bayesian statistics – MCMC techniques
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
April 2010UW-Madison © Brian S. Yandell1 Bayesian QTL Mapping Brian S. Yandell University of Wisconsin-Madison UW-Madison,
Model SelectionSeattle SISG: Yandell © QTL Model Selection 1.Bayesian strategy 2.Markov chain sampling 3.sampling genetic architectures 4.criteria.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
Inference for regression - Simple linear regression
Bayes Factor Based on Han and Carlin (2001, JASA).
Modeling Menstrual Cycle Length in Pre- and Peri-Menopausal Women Michael Elliott Xiaobi Huang Sioban Harlow University of Michigan School of Public Health.
April 2008UW-Madison © Brian S. Yandell1 Bayesian QTL Mapping Brian S. Yandell University of Wisconsin-Madison UW-Madison,
Model Inference and Averaging
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
QTL 2: OverviewSeattle SISG: Yandell © Seattle Summer Institute 2010 Advanced QTL Brian S. Yandell, UW-Madison
October 2007Jax Workshop © Brian S. Yandell1 Bayesian Model Selection for Multiple QTL Brian S. Yandell University of Wisconsin-Madison
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
October 2005Jax Workshop © Brian S. Yandell1 Bayesian Model Selection for Multiple QTL Brian S. Yandell University of Wisconsin-Madison
Lecture 12: Linkage Analysis V Date: 10/03/02  Least squares  An EM algorithm  Simulated distribution  Marker coverage and density.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
Lecture 24: Quantitative Traits IV Date: 11/14/02  Sources of genetic variation additive dominance epistatic.
BayesNCSU QTL II: Yandell © Bayesian Interval Mapping 1.what is Bayes? Bayes theorem? Bayesian QTL mapping Markov chain sampling18-25.
Tutorial I: Missing Value Analysis
Lecture 1: Basic Statistical Tools. A random variable (RV) = outcome (realization) not a set value, but rather drawn from some probability distribution.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
13 October 2004Statistics: Yandell © Inferring Genetic Architecture of Complex Biological Processes Brian S. Yandell 12, Christina Kendziorski 13,
June 1999NCSU QTL Workshop © Brian S. Yandell 1 Bayesian Inference for QTLs in Inbred Lines II Brian S Yandell University of Wisconsin-Madison
Kevin Stevenson AST 4762/5765. What is MCMC?  Random sampling algorithm  Estimates model parameters and their uncertainty  Only samples regions of.
September 2002Jax Workshop © Brian S. Yandell1 Bayesian Model Selection for Quantitative Trait Loci using Markov chain Monte Carlo in Experimental Crosses.
NSF StatGen: Yandell © NSF StatGen 2009 Bayesian Interval Mapping Brian S. Yandell, UW-Madison overview: multiple.
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
April 2008Stat 877 © Brian S. Yandell1 Reversible Jump Details Brian S. Yandell University of Wisconsin-Madison April.
ModelNCSU QTL II: Yandell © Model Selection for Multiple QTL 1.reality of multiple QTL3-8 2.selecting a class of QTL models comparing QTL models16-24.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Plant Microarray Course
Bayesian Semi-Parametric Multiple Shrinkage
MCMC Output & Metropolis-Hastings Algorithm Part I
Model Selection for Multiple QTL
Bayesian Model Selection for Multiple QTL
Bayesian Inference for QTLs in Inbred Lines I
Graphical Diagnostics for Multiple QTL Investigation Brian S
Gene mapping in mice Karl W Broman Department of Biostatistics
Bayesian Model Selection for Quantitative Trait Loci with Markov chain Monte Carlo in Experimental Crosses Brian S. Yandell University of Wisconsin-Madison.
Inferring Genetic Architecture of Complex Biological Processes Brian S
Jax Workshop © Brian S. Yandell
Bayesian Inference for QTLs in Inbred Lines
Bayesian Interval Mapping
Bayesian Model Selection for Multiple QTL
Bayesian Interval Mapping
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Seattle SISG: Yandell © 2007
Inferring Genetic Architecture of Complex Biological Processes Brian S
Note on Outbred Studies
Brian S. Yandell University of Wisconsin-Madison
Simple Linear Regression
Robust Full Bayesian Learning for Neural Networks
Seattle SISG: Yandell © 2009
Parametric Methods Berlin Chen, 2005 References:
Bayesian Interval Mapping
Presentation transcript:

April 2008UW-Madison © Brian S. Yandell1 Bayesian QTL Mapping Brian S. Yandell University of Wisconsin-Madison UW-Madison, April 2008

April 2008UW-Madison © Brian S. Yandell2 outline 1.What is the goal of QTL study? 2.Bayesian vs. classical QTL study 3.Bayesian strategy for QTLs 4.model search using MCMC Gibbs sampler and Metropolis-Hastings 5.model assessment Bayes factors & model averaging 6.analysis of hyper data 7.software for Bayesian QTLs

April 2008UW-Madison © Brian S. Yandell3 1. what is the goal of QTL study? uncover underlying biochemistry –identify how networks function, break down –find useful candidates for (medical) intervention –epistasis may play key role –statistical goal: maximize number of correctly identified QTL basic science/evolution –how is the genome organized? –identify units of natural selection –additive effects may be most important (Wright/Fisher debate) –statistical goal: maximize number of correctly identified QTL select “elite” individuals –predict phenotype (breeding value) using suite of characteristics (phenotypes) translated into a few QTL –statistical goal: mimimize prediction error

April 2008UW-Madison © Brian S. Yandell4 advantages of multiple QTL approach improve statistical power, precision –increase number of QTL detected –better estimates of loci: less bias, smaller intervals improve inference of complex genetic architecture –patterns and individual elements of epistasis –appropriate estimates of means, variances, covariances asymptotically unbiased, efficient –assess relative contributions of different QTL improve estimates of genotypic values –less bias (more accurate) and smaller variance (more precise) –mean squared error = MSE = (bias) 2 + variance

April 2008UW-Madison © Brian S. Yandell5 Pareto diagram of QTL effects major QTL on linkage map major QTL minor QTL polygenes (modifiers) Intuitive idea of ellipses: Horizontal = significance Vertical = support interval

April 2008UW-Madison © Brian S. Yandell6 check QTL in context of genetic architecture scan for each QTL adjusting for all others –adjust for linked and unlinked QTL adjust for linked QTL: reduce bias adjust for unlinked QTL: reduce variance –adjust for environment/covariates examine entire genetic architecture –number and location of QTL, epistasis, GxE –model selection for best genetic architecture

April 2008UW-Madison © Brian S. Yandell7 2. Bayesian vs. classical QTL study classical study –maximize over unknown effects –test for detection of QTL at loci –model selection in stepwise fashion Bayesian study –average over unknown effects –estimate chance of detecting QTL –sample all possible models both approaches –average over missing QTL genotypes –scan over possible loci

March 2011UW-Madison © Brian S. Yandell8 Reverend Thomas Bayes ( ) –part-time mathematician –buried in Bunhill Cemetary, Moongate, London –famous paper in 1763 Phil Trans Roy Soc London Barnard (1958 Biometrika), Press (1989) Bayesian Statistics Stigler (1986) History of Statistics Carlin Louis (1996); Gelman et al. (1995) books –Was Bayes the first with this idea? (Laplace) billiard balls on rectangular table –two balls tossed at random (uniform) on table –where is first ball if the second is to its right (left)? Who was Bayes? first ball

March 2011UW-Madison © Brian S. Yandell9 Where is the first ball? priorpr(  ) = 1 likelihoodpr(Y |  )=  Y (1-  ) 1-Y posteriorpr(  |Y)= ?  Y=0 Y=1 first ball second ball (now throw second ball n times)

March 2011UW-Madison © Brian S. Yandell10 What is Bayes Theorem? before and after observing data –prior:pr(  ) = pr(parameters) –posterior:pr(  |Y)= pr(parameters|data) posterior = likelihood * prior / constant –usual likelihood of parameters given data –normalizing constant pr(Y) depends only on data constant often drops out of calculation

March 2011UW-Madison © Brian S. Yandell11 What is Probability? Frequentist analysis chance over many trials –long run average estimates confidence intervals –long term frequency hypothesis tests p-values Type I error rate –reject null when true –chance of extreme result Bayesian analysis uncertainty of true value prior –uncertainty before data –incorporate prior knowledge/experience posterior –uncertainty after analyzing current data –balance prior and data

March 2011UW-Madison © Brian S. Yandell12 Likelihood and Posterior Example

March 2011UW-Madison © Brian S. Yandell13 Frequentist or Bayesian? Frequentist approach –fixed parameters range of values –maximize likelihood ML estimates find the peak –confidence regions random region invert a test –hypothesis testing 2 nested models Bayesian approach –random parameters distribution –posterior distribution posterior mean sample from dist –credible sets fixed region given data HPD regions –model selection/critique Bayes factors

March 2011UW-Madison © Brian S. Yandell14 Frequentist or Bayesian? Frequentist approach –maximize over mixture of QT genotypes –locus profile likelihood max over effects –HPD region for locus natural for locus –1-2 LOD drop work to get effects –approximate shape of likelihood peak Bayesian approach –joint distribution over QT genotypes –sample distribution joint effects & loci –HPD regions for joint locus & effects use density estimator

March 2011UW-Madison © Brian S. Yandell15 Choice of Bayesian priors elicited priors –higher weight for more probable parameter values based on prior empirical knowledge –use previous study to inform current study weather prediction, previous QTL studies on related organisms conjugate priors –convenient mathematical form –essential before computers, helpful now to simply computation –large variances on priors reduces their influence on posterior non-informative priors –may have “no” information on unknown parameters –prior with all parameter values equally likely may not sum to 1 (improper), which can complicate use always check sensitivity of posterior to choice of prior

April 2008UW-Madison © Brian S. Yandell16 QTL model selection: key players observed measurements –y = phenotypic trait –m = markers & linkage map –i = individual index (1,…,n) missing data –missing marker data –q = QT genotypes alleles QQ, Qq, or qq at locus unknown quantities – = QT locus (or loci) –  = phenotype model parameters –A = QTL model/genetic architecture pr(q|m,,A) genotype model –grounded by linkage map, experimental cross –recombination yields multinomial for q given m pr(y|q, ,A) phenotype model –distribution shape (assumed normal here) –unknown parameters  (could be non-parametric) after Sen Churchill (2001) y q  m A

April 2008UW-Madison © Brian S. Yandell17 likelihood and posterior

April 2008UW-Madison © Brian S. Yandell18 Bayes posterior vs. maximum likelihood (genetic architecture A = single QTL at ) LOD: classical Log ODds –maximize likelihood over effects µ –R/qtl scanone/scantwo: method = “em” LPD: Bayesian Log Posterior Density –average posterior over effects µ –R/qtl scanone/scantwo: method = “imp”

April 2008UW-Madison © Brian S. Yandell19 LOD & LPD: 1 QTL n.ind = 100, 1 cM marker spacing

April 2008UW-Madison © Brian S. Yandell20 Simplified likelihood surface 2-D for BC locus and effect locus and effect Δ = µ 2 – µ 1 profile likelihood along ridge –maximize likelihood at each for Δ –symmetric in Δ around MLE given weighted average of posterior –average likelihood at each with weight p(Δ) –how does prior p(Δ) affect symmetry?

April 2008UW-Madison © Brian S. Yandell21 LOD & LPD: 1 QTL n.ind = 100, 10 cM marker spacing

April 2008UW-Madison © Brian S. Yandell22 likelihood and posterior likelihood relates “known” data (y,m,q) to unknown values of interest ( ,,A) –pr(y,q|m, ,,A) = pr(y|q, ,A) pr(q|m,,A) –mix over unknown genotypes (q) posterior turns likelihood into a distribution –weight likelihood by priors –rescale to sum to 1.0 –posterior = likelihood * prior / constant

April 2008UW-Madison © Brian S. Yandell23 marginal LOD or LPD What is contribution of a QTL adjusting for all others? –improvement in LPD due to QTL at locus –contribution due to main effects, epistasis, GxE? How does adjusted LPD differ from unadjusted LPD? –raised by removing variance due to unlinked QTL –raised or lowered due to bias of linked QTL –analogous to Type III adjusted ANOVA tests can ask these same questions using classical LOD –see Broman’s newer tools for multiple QTL inference

April 2008UW-Madison © Brian S. Yandell24 LPD: 1 QTL vs. multi-QTL marginal contribution to LPD from QTL at 2 nd QTL 1 st QTL 2 nd QTL

April 2008UW-Madison © Brian S. Yandell25 substitution effect: 1 QTL vs. multi-QTL single QTL effect vs. marginal effect from QTL at 2 nd QTL 1 st QTL

April 2008UW-Madison © Brian S. Yandell26 3. Bayesian strategy for QTLs augment data (y,m) with missing genotypes q build model for augmented data –genotypes (q) evaluated at loci ( ) depends on flanking markers (m) –phenotypes (y) centered about effects (  ) depends on missing genotypes (q) – and  depend on genetic architecture (A) How complicated is model? number of QTL, epistasis, etc. sample from model in some clever way infer most probable genetic architecture –estimate loci, their main effects and epistasis –study properties of estimates

April 2008UW-Madison © Brian S. Yandell27 do phenotypes help to guess genotypes? posterior on QTL genotypes q what are probabilities for genotype q between markers? all recombinants AA:AB have 1:1 prior ignoring y what if we use y? M41M214

April 2008UW-Madison © Brian S. Yandell28 posterior on QTL genotypes q full conditional of q given data, parameters –proportional to prior pr(q | m, ) weight toward q that agrees with flanking markers –proportional to likelihood pr(y|q,  ) weight toward q with similar phenotype values –posterior balances these two this is the E-step of EM computations

April 2008UW-Madison © Brian S. Yandell29

March 2011UW-Madison © Brian S. Yandell30 Y = G + Eposterior for single individual environE ~ N(0,  2 ),  2 known likelihoodpr(Y | G,  2 ) = N(Y | G,  2 ) priorpr(G |  2, ,  ) = N(G | ,  2 /  ) posteriorN(G |  +B 1 (Y-  ), B 1  2 ) Y i = G + E i posterior for sample of n individuals shrinkage weights B n go to 1 Bayes for normal data

March 2011UW-Madison © Brian S. Yandell31 effect of prior variance on posterior normal prior, posterior for n = 1, posterior for n = 5, true mean

April 2008UW-Madison © Brian S. Yandell32 where are the genotypic means? (phenotype mean for genotype q is  q ) data meandata meansprior mean posterior means

April 2008UW-Madison © Brian S. Yandell33 prior & posteriors: genotypic means  q prior for genotypic means –centered at grand mean –variance related to heritability of effect hyper-prior on variance (details omitted) posterior –shrink genotypic means toward grand mean –shrink variance of genotypic mean prior: posterior: shrinkage:

March 2011UW-Madison © Brian S. Yandell34 How do we choose hyper-parameters ,  ? Empirical Bayes:marginalize over prior estimate ,  from marginal posterior likelihoodpr(Y i | Q i,G,  2 ) = N(Y i | G(Q i ),  2 ) priorpr(G Q |  2, ,  ) = N(G Q | ,  2 /  ) marginalpr(Y i |  2, ,  ) = N(Y i | ,  2 (  + 1)/  ) estimates EB posterior Empirical Bayes: choosing hyper-parameters

March 2011UW-Madison © Brian S. Yandell35 What if variance  2 is unknown? recall that sample variance is proportional to chi-square –pr(s 2 |  2 ) =  2 (ns 2 /  2 | n) –or equivalently, ns 2 /  2 |  2 ~  n 2 conjugate prior is inverse chi-square –pr(  2 |,  2 ) = inv-  2 (  2 |,  2 ) –or equivalently,  2 /  2 |,  2 ~  2 –empirical choice:  2 = s 2 /3, =6 E(  2 |,  2 ) = s 2 /2, Var(  2 |,  2 ) = s 4 /4 posterior given data –pr(  2 | Y,,  2 ) = inv-  2 (  2 | +n, (  2 +ns 2 )/( +n) )

April 2008UW-Madison © Brian S. Yandell36 phenotype affected by genotype & environment E(y|q) =  q =  0 + sum {j in H}  j (q) number of terms in QTL model H  2 nqtl (3 nqtl for F 2 ) partition genotypic mean into QTL effects  q =  0 +  1 (q 1 ) +  2 (q 2 ) +  12 (q 1,q 2 )  q = mean+ main effects + epistatic interactions partition prior and posterior (details omitted) multiple QTL phenotype model

April 2008UW-Madison © Brian S. Yandell37 QTL with epistasis same phenotype model overview partition of genotypic value with epistasis partition of genetic variance & heritability

April 2008UW-Madison © Brian S. Yandell38 partition genotype-specific mean into QTL effects µ q = mean + main effects + epistatic interactions µ q =  +  q =  + sum j in A  qj priors on mean and effects  ~ N(  0,  0  2 )grand mean  q ~ N(0,  1  2 )model-independent genotypic effect  qj ~ N(0,  1  2 /|A|)effects down-weighted by size of A determine hyper-parameters via empirical Bayes partition of multiple QTL effects

March 2011UW-Madison © Brian S. Yandell39 Recombination and Distance assume map and marker distances are known useful approximation for QTL linkage –Haldane map function: no crossover interference –independence implies crossover events are Poisson all computations consistent in approximation –rely on given map with known marker locations –1-to-1 relation of distance to recombination –all map functions are approximate anyway

March 2011UW-Madison © Brian S. Yandell40 recombination model pr (Q|X, ) locus is distance along linkage map –identifies flanking marker region flanking markers provide good approximation –map assumed known from earlier study –inaccuracy slight using only flanking markers extend to next flanking markers if missing data –could consider more complicated relationship but little change in results pr (Q|X, ) = pr(geno | map, locus)  pr(geno | flanking markers, locus) Q?

April 2008UW-Madison © Brian S. Yandell41 Where are the loci on the genome? prior over genome for QTL positions –flat prior = no prior idea of loci –or use prior studies to give more weight to some regions posterior depends on QTL genotypes q pr( | m,q) = pr( ) pr(q | m, ) / constant –constant determined by averaging over all possible genotypes q over all possible loci on entire map no easy way to write down posterior

March 2011UW-Madison © Brian S. Yandell42 prior & posterior for QT locus no prior information on locus –uniform prior over genome –use framework map choose interval proportional to length then pick uniform position within interval prior information from other studies concentrate on credible regions use posterior of previous study as new prior posterior prior

April 2008UW-Madison © Brian S. Yandell43 model fit with multiple imputation (Sen and Churchill 2001) pick a genetic architecture –1, 2, or more QTL fill in missing genotypes at ‘pseudomarkers’ –use prior recombination model use clever weighting (importance sampling) compute LPD, effect estimates, etc.

April 2008UW-Madison © Brian S. Yandell44 4. QTL Model Search using MCMC construct Markov chain around posterior –want posterior as stable distribution of Markov chain –in practice, the chain tends toward stable distribution initial values may have low posterior probability burn-in period to get chain mixing well sample QTL model components from full conditionals –sample locus given q,A (using Metropolis-Hastings step) –sample genotypes q given, ,y,A (using Gibbs sampler) –sample effects  given q,y,A (using Gibbs sampler) –sample QTL model A given, ,y,q (using Gibbs or M-H)

March 2011UW-Madison © Brian S. Yandell45 EM-MCMC duality EM approaches can be redone with MCMC –EM estimates & maximizes –MCMC draws random samples simulated annealing: gradually cool towards peak –both can address same problem sometimes EM is hard (impossible) to use MCMC is tool of “last resort” –use exact methods if you can –try other approximate methods –be clever! (math, computing tricks) –very handy for hard problems in genetics

March 2011UW-Madison © Brian S. Yandell46 Why not Ordinary Monte Carlo? independent samples of joint distribution chaining (or peeling) of effects pr(  |Y,Q)=pr(G Q | Y,Q,  2 ) pr(  2 | Y,Q) possible analytically here given genotypes Q Monte Carlo: draw N samples from posterior –sample variance  2 –sample genetic values G Q given variance  2 but we know markers X, not genotypes Q! –would have messy average over possible Q –pr(  |Y,X) = sum Q pr(  |Y,Q) pr(Q|Y,X)

March 2011UW-Madison © Brian S. Yandell47 What is a Markov chain? future given present is independent of past update chain based on current value –can make chain arbitrarily complicated –chain converges to stable pattern π() we wish to study 01 1-q q p 1-p

March 2011UW-Madison © Brian S. Yandell48 Markov chain idea 01 1-q q p 1-p Mitch’s other pubs

March 2011UW-Madison © Brian S. Yandell49 Markov chain Monte Carlo can study arbitrarily complex models –need only specify how parameters affect each other –can reduce to specifying full conditionals construct Markov chain with “right” model –joint posterior of unknowns as limiting “stable” distribution –update unknowns given data and all other unknowns sample from full conditionals cycle at random through all parameters –next step depends only on current values nice Markov chains have nice properties –sample summaries make sense –consider almost as random sample from distribution –ergodic theorem and all that stuff

April 2008UW-Madison © Brian S. Yandell50 MCMC sampling of (,q,µ ) Gibbs sampler –genotypes q –effects µ –not loci Metropolis-Hastings sampler –extension of Gibbs sampler –does not require normalization pr( q | m ) = sum pr( q | m, ) pr( )

April 2008UW-Madison © Brian S. Yandell51 Gibbs sampler idea toy problem –want to study two correlated effects –could sample directly from their bivariate distribution instead use Gibbs sampler: –sample each effect from its full conditional given the other –pick order of sampling at random –repeat many times

April 2008UW-Madison © Brian S. Yandell52 Gibbs sampler samples:  = 0.6 N = 50 samples N = 200 samples

April 2008UW-Madison © Brian S. Yandell53 Metropolis-Hastings idea want to study distribution f( ) –take Monte Carlo samples unless too complicated –take samples using ratios of f Metropolis-Hastings samples: –propose new value * near (?) current value from some distribution g –accept new value with prob a Gibbs sampler: a = 1 always f( ) g( – * )

April 2008UW-Madison © Brian S. Yandell54 Metropolis-Hastings samples N = 200 samples N = 1000 samples narrow gwide gnarrow gwide g histogram

April 2008UW-Madison © Brian S. Yandell55 MCMC realization added twist: occasionally propose from whole domain

April 2008UW-Madison © Brian S. Yandell56 What is the genetic architecture A? components of genetic architecture –how many QTL? –where are loci ( )? how large are effects (µ)? –which pairs of QTL are epistatic? use priors to weight posterior –toward guess from previous analysis –improve efficiency of sampling from posterior increase samples from architectures of interest

March 2011UW-Madison © Brian S. Yandell57 Markov chain for number m add a new locus drop a locus update current model 01mm-1m+1... m

March 2011UW-Madison © Brian S. Yandell58 jumping QTL number and loci

April 2008UW-Madison © Brian S. Yandell59 Whole Genome Phenotype Model E(y) = µ +  (q) = µ + X  –y = n phenotypes –X = n  L design matrix in theory covers whole genome of size L cM X determined by genotypes and model space only need terms associated with q = n  n QTL genotypes at QTL –  = diag(  ) = genetic architecture  = 0,1 indicators for QTLs or pairs of QTLs |  | =  = size of genetic architecture = loci determined implicitly by  –  = genotypic effects (main and epistatic) –µ = reference

April 2008UW-Madison © Brian S. Yandell60 Methods of Model Search Reversible jump (transdimensional) MCMC –sample possible loci ( determines possible  ) –collapse to model containing just those QTL –bookkeeping when model dimension changes Composite model with indicators –include all terms in model:  and  –sample possible architecture (  determines ) can use LASSO-type prior for model selection Shrinkage model –set  = 1 (include all loci) –allow variances of  to differ (shrink coefficients to zero)

April 2008UW-Madison © Brian S. Yandell61 sampling across QTL models A action steps: draw one of three choices update QTL model A with probability 1-b(A)-d(A) –update current model using full conditionals –sample QTL loci, effects, and genotypes add a locus with probability b(A) –propose a new locus along genome –innovate new genotypes at locus and phenotype effect –decide whether to accept the “birth” of new locus drop a locus with probability d(A) –propose dropping one of existing loci –decide whether to accept the “death” of locus 0L 1 m+1 m 2 …

April 2008UW-Madison © Brian S. Yandell62 reversible jump MCMC consider known genotypes q at 2 known loci –models with 1 or 2 QTL M-H step between 1-QTL and 2-QTL models –model changes dimension (via careful bookkeeping) –consider mixture over QTL models H

April 2008UW-Madison © Brian S. Yandell63 Gibbs sampler with loci indicators partition genome into intervals –at most one QTL per interval –interval = 1 cM in length –assume QTL in middle of interval use loci to indicate presence/absence of QTL in each interval –  = 1 if QTL in interval –  = 0 if no QTL Gibbs sampler on loci indicators –see work of Nengjun Yi (and earlier work of Ina Hoeschele)

April 2008UW-Madison © Brian S. Yandell64 Bayesian shrinkage estimation soft loci indicators –strength of evidence for j depends on variance of  j –similar to  > 0 on grey scale include all possible loci in model –pseudo-markers at 1cM intervals Wang et al. (2005 Genetics) –Shizhong Xu group at U CA Riverside

April 2008UW-Madison © Brian S. Yandell65 epistatic interactions model space issues –Fisher-Cockerham partition vs. tree-structured? –2-QTL interactions only? –general interactions among multiple QTL? –retain model hierarchy (include main QTL)? model search issues –epistasis between significant QTL check all possible pairs when QTL included? allow higher order epistasis? –epistasis with non-significant QTL whole genome paired with each significant QTL? pairs of non-significant QTL? Yi et al. (2005, 2007)

March 2011UW-Madison © Brian S. Yandell66 Reversible Jump Details reversible jump MCMC details –can update model with m QTL –have basic idea of jumping models –now: careful bookkeeping between models RJ-MCMC & Bayes factors –Bayes factors from RJ-MCMC chain –components of Bayes factors

March 2011UW-Madison © Brian S. Yandell67 reversible jump choices action step: draw one of three choices (m = number of QTL in model) update step with probability 1-b(m+1)-d(m) –update current model –loci, effects, genotypes as before add a locus with probability b(m+1) –propose a new locus –innovate effect and genotypes at new locus –decide whether to accept the “birth” of new locus drop a locus with probability d(m) –pick one of existing loci to drop –decide whether to accept the “death” of locus

March 2011UW-Madison © Brian S. Yandell68 RJ-MCMC updates effects  loci traits Y genos Q add locus drop locus b(m+1) d(m) 1-b(m+1)-d(m) map X

March 2011UW-Madison © Brian S. Yandell69 propose to drop a locus choose an existing locus –equal weight for all loci ? –more weight to loci with small effects? “drop” effect & genotypes at old locus –adjust effects at other loci for collinearity –this is reverse jump of Green (1995) check acceptance … –do not drop locus, effects & genotypes –until move is accepted 12 m+1 3 …

March 2011UW-Madison © Brian S. Yandell70 propose a new locus –uniform chance over genome –actually need to be more careful (R van de Ven, pers. comm.) choose interval between loci already in model (include 0,L) –probability proportional to interval length ( 2 – 1 )/L uniform chance within this interval 1/( 2 – 1 ) –need genotypes at locus & model effect innovate effect & genotypes at new locus –draw genotypes based on recombination (prior) no dependence on trait model yet –draw effect as in Green’s reversible jump adjust for collinearity: modify other parameters accordingly check acceptance... propose to add a locus 0L 1 m+1 m 2 …

March 2011UW-Madison © Brian S. Yandell71 acceptance of reversible jump accept birth of new locus with probability min(1,A) accept death of old locus with probability min(1,1/A)

March 2011UW-Madison © Brian S. Yandell72 move probabilities birth & death proposals Jacobian between models –fudge factor –see stepwise regression example acceptance of reversible jump mm+1

March 2011UW-Madison © Brian S. Yandell73 reversible jump idea expand idea of MCMC to compare models adjust for parameters in different models –augment smaller model with innovations –constraints on larger model calculus “change of variables” is key –add or drop parameter(s) –carefully compute the Jacobian consider stepwise regression –Mallick (1995) & Green (1995) –efficient calculation with Hausholder decomposition

March 2011UW-Madison © Brian S. Yandell74 model selection in regression known regressors (e.g. markers) –models with 1 or 2 regressors jump between models –centering regressors simplifies calculations

March 2011UW-Madison © Brian S. Yandell75 slope estimate for 1 regressor recall least squares estimate of slope note relation of slope to correlation

March 2011UW-Madison © Brian S. Yandell76 2 correlated regressors slopes adjusted for other regressors

March 2011UW-Madison © Brian S. Yandell77 Gibbs Sampler for Model 1 mean slope variance

March 2011UW-Madison © Brian S. Yandell78 Gibbs Sampler for Model 2 mean slopes variance

March 2011UW-Madison © Brian S. Yandell79 updates from 2->1 drop 2nd regressor adjust other regressor

March 2011UW-Madison © Brian S. Yandell80 updates from 1->2 add 2nd slope, adjusting for collinearity adjust other slope & variance

March 2011UW-Madison © Brian S. Yandell81 model selection in regression known regressors (e.g. markers) –models with 1 or 2 regressors jump between models –augment with new innovation z

March 2011UW-Madison © Brian S. Yandell82 change of variables change variables from model 1 to model 2 calculus issues for integration –need to formally account for change of variables –infinitessimal steps in integration (db) –involves partial derivatives (next page)

March 2011UW-Madison © Brian S. Yandell83 Jacobian & the calculus Jacobian sorts out change of variables –careful: easy to mess up here!

March 2011UW-Madison © Brian S. Yandell84 geometry of reversible jump a1a1 a1a1 a2a2 a2a2

March 2011UW-Madison © Brian S. Yandell85 QT additive reversible jump a1a1 a1a1 a2a2 a2a2

March 2011UW-Madison © Brian S. Yandell86 credible set for additive 90% & 95% sets based on normal regression line corresponds to slope of updates a1a1 a2a2

April 2008UW-Madison © Brian S. Yandell87 collinear QTL = correlated effects effect 2 effect 1 effect 2 effect 1 linked QTL = collinear genotypes  correlated estimates of effects (negative if in coupling phase)  sum of linked effects usually fairly constant

March 2011UW-Madison © Brian S. Yandell88 multivariate updating of effects more computations when m > 2 avoid matrix inverse –Cholesky decomposition of matrix simultaneous updates –effects at all loci accept new locus based on –sampled new genos at locus –sampled new effects at all loci also long-range positions updates before after

March 2011UW-Madison © Brian S. Yandell89 8 QTL simulation (Stevens Fisch 1998) n = 200, h 2 =.5 –SF detected 3 QTL Bayesian IM n h 2 detect

March 2011UW-Madison © Brian S. Yandell90 posterior number of QTL geometric prior with mean 0.5 seems to have no influence on posterior here

March 2011UW-Madison © Brian S. Yandell91 posterior genetic architecture Chromosome count vector m Count

April 2008UW-Madison © Brian S. Yandell92 Bayesian model averaging average summaries over multiple architectures avoid selection of “best” model focus on “better” models examples in data talk later

March 2011UW-Madison © Brian S. Yandell93 model averaging for 8 QTL

March 2011UW-Madison © Brian S. Yandell94 model averaging: focus on chr. 8 all samples samples with 8 QTL

April 2008UW-Madison © Brian S. Yandell95 5. Model Assessment balance model fit against model complexity smaller modelbigger model model fitmiss key featuresfits better predictionmay be biasedno bias interpretationeasiermore complicated parameterslow variancehigh variance information criteria: penalize likelihood by model size –compare IC = – 2 log L( model | data ) + penalty(model size) Bayes factors: balance posterior by prior choice –compare pr( data | model)

April 2008UW-Madison © Brian S. Yandell96 Bayes factors ratio of model likelihoods –ratio of posterior to prior odds for architectures –average over unknown effects (µ) and loci ( ) roughly equivalent to BIC –BIC maximizes over unknowns –BF averages over unknowns

April 2008UW-Madison © Brian S. Yandell97 issues in computing Bayes factors BF insensitive to shape of prior on A –geometric, Poisson, uniform –precision improves when prior mimics posterior BF sensitivity to prior variance on effects  –prior variance should reflect data variability –resolved by using hyper-priors automatic algorithm; no need for user tuning easy to compute Bayes factors from samples –apply Bayes’ rule and solve for pr(y | m, A) pr(A | y, m) = pr(y | m, A) pr(A | m) / constant pr(data|model) = constant * pr(model|data) / pr(model) –posterior pr(A | y, m) is marginal histogram

April 2008UW-Madison © Brian S. Yandell98 Bayes factors and genetic model A |A| = number of QTL –prior pr(A) chosen by user –posterior pr(A|y,m) sampled marginal histogram shape affected by prior pr(A) pattern of QTL across genome gene action and epistasis geometric

April 2008UW-Madison © Brian S. Yandell99 BF sensitivity to fixed prior for effects

April 2008UW-Madison © Brian S. Yandell100 BF insensitivity to random effects prior

March 2011UW-Madison © Brian S. Yandell101 How sensitive is posterior to choice of prior? simulations with 0, 1 or 2 QTL –strong effects (additive = 2, variance = 1) –linked loci 36cM apart differences with number of QTL –clear differences by actual number –works well with 100,000, better with 1M effect of Poisson prior mean –larger prior mean shifts posterior up –but prior does not take over

March 2011UW-Madison © Brian S. Yandell102 simulation study: prior 2 QTL at 15, 65cM n = 100, 200; h 2 = 40% vary prior mean from 1 to 10 QTL –Poisson prior 10 independent simulations examine posterior mean, probability

March 2011UW-Madison © Brian S. Yandell103 posterior m depends on prior

March 2011UW-Madison © Brian S. Yandell104 cumulative posterior as prior mean changes

March 2011UW-Madison © Brian S. Yandell105 effect of prior mean on posterior m

March 2011UW-Madison © Brian S. Yandell106 effect of prior shape on posterior

April 2008UW-Madison © Brian S. Yandell107 marginal BF scan by QTL compare models with and without QTL at –average over all possible models –estimate as ratio of samples with/without QTL scan over genome for peaks –2log(BF) seems to have similar properties to LPD

April 2008UW-Madison © Brian S. Yandell analysis of hyper data marginal scans of genome –detect significant loci –infer main and epistatic QTL, GxE infer most probable genetic architecture –number of QTL –chromosome pattern of QTL with epistasis diagnostic summaries –heritability, unexplained variation

April 2008UW-Madison © Brian S. Yandell109 marginal scans of genome LPD and 2log(BF) “tests” for each locus estimates of QTL effects at each locus separately infer main effects and epistasis –main effect for each locus (blue) –epistasis for loci paired with another (purple) identify epistatic QTL in 1-D scan infer pairing in 2-D scan

April 2008UW-Madison © Brian S. Yandell110 hyper data: scanone

April 2008UW-Madison © Brian S. Yandell111 2log(BF) scan with 50% HPD region

April 2008UW-Madison © Brian S. Yandell112 2-D plot of 2logBF: chr 6 & 15

April 2008UW-Madison © Brian S. Yandell113 1-D Slices of 2-D scans: chr 6 & 15

April 2008UW-Madison © Brian S. Yandell114 1-D Slices of 2-D scans: chr 6 & 15

April 2008UW-Madison © Brian S. Yandell115 What is best genetic architecture? How many QTL? What is pattern across chromosomes? examine posterior relative to prior –prior determined ahead of time –posterior estimated by histogram/bar chart –Bayes factor ratio = pr(model|data) / pr(model)

April 2008UW-Madison © Brian S. Yandell116 How many QTL? posterior, prior, Bayes factor ratios prior strength of evidence MCMC error

April 2008UW-Madison © Brian S. Yandell117 most probable patterns nqtl posterior prior bf bfse 1,4,6,15,6: e ,4,6,6,15,6: e ,1,4,6,15,6: e ,1,4,5,6,15,6: e ,4,6,15,15,6: e ,4,4,6,15,6: e ,2,4,6,15,6: e ,4,5,6,15,6: e ,2,4,5,6,15,6: e , e ,1,2, e ,2, e ,1, e ,4, e

April 2008UW-Madison © Brian S. Yandell118 what is best estimate of QTL? find most probable pattern –1,4,6,15,6:15 has posterior of 3.4% estimate locus across all nested patterns –Exact pattern seen ~100/3000 samples –Nested pattern seen ~2000/3000 samples estimate 95% confidence interval using quantiles chrom locus locus.LCL locus.UCL n.qtl

April 2008UW-Madison © Brian S. Yandell119 how close are other patterns? size & shade ~ posterior distance between patterns –sum of squared attenuation –match loci between patterns –squared attenuation = (1-2r) 2 –sq.atten in scale of LOD & LPD multidimensional scaling –MDS projects distance onto 2-D –think mileage between cities

April 2008UW-Madison © Brian S. Yandell120 how close are other patterns?

April 2008UW-Madison © Brian S. Yandell121 diagnostic summaries

April 2008UW-Madison © Brian S. Yandell Software for Bayesian QTLs R/qtlbim publication –CRAN release Fall 2006 –Yandell et al. (2007 Bioinformatics) properties –cross-compatible with R/qtl –epistasis, fixed & random covariates, GxE –extensive graphics

April 2008UW-Madison © Brian S. Yandell123 R/qtlbim: software history Bayesian module within WinQTLCart –WinQTLCart output can be processed using R/bim Software history –initially designed (Satagopan Yandell 1996) –major revision and extension (Gaffney 2001) –R/bim to CRAN (Wu, Gaffney, Jin, Yandell 2003) –R/qtlbim total rewrite (Yandell et al. 2007)

April 2008UW-Madison © Brian S. Yandell124 other Bayesian software for QTLs R/bim*: Bayesian Interval Mapping –Satagopan Yandell (1996; Gaffney 2001) CRAN –no epistasis; reversible jump MCMC algorithm –version available within WinQTLCart (statgen.ncsu.edu/qtlcart) R/qtl* –Broman et al. (2003 Bioinformatics) CRAN –multiple imputation algorithm for 1, 2 QTL scans & limited mult-QTL fits Bayesian QTL / Multimapper –Sillanpää Arjas (1998 Genetics) –no epistasis; introduced posterior intensity for QTLs (no released code) –Stephens & Fisch (1998 Biometrics) –no epistasis R/bqtl –C Berry (1998 TR) CRAN –no epistasis, Haley Knott approximation * Jackson Labs (Hao Wu, Randy von Smith) provided crucial technical support

April 2008UW-Madison © Brian S. Yandell125 many thanks Karl Broman Jackson Labs Gary Churchill Hao Wu Randy von Smith U AL Birmingham David Allison Nengjun Yi Tapan Mehta Samprit Banerjee Ram Venkataraman Daniel Shriner Michael Newton Hyuna Yang Daniel Sorensen Daniel Gianola Liang Li my students Jaya Satagopan Fei Zou Patrick Gaffney Chunfang Jin Elias Chaibub W Whipple Neely Jee Young Moon USDA Hatch, NIH/NIDDK (Attie), NIH/R01 (Yi, Broman) Tom Osborn David Butruille Marcio Ferrera Josh Udahl Pablo Quijada Alan Attie Jonathan Stoehr Hong Lan Susie Clee Jessica Byers Mark Keller