Summarizing Variation Michael C Neale PhD Virginia Institute for Psychiatric and Behavioral Genetics Virginia Commonwealth University
Overview )Mean )Variance )Covariance )Not always necessary/desirable
Computing Mean Formula E (x i )/N )Can compute with 5Pencil 5Calculator 5SAS 5SPSS 5Mx
One Coin toss 2 outcomes HeadsTails Outc ome Probab ility
Two Coin toss 3 outcomes HHHT/THTT Outc ome Probab ility
Four Coin toss 5 outcomes HHHHHHHTHHTTHTTTTTTT Outc ome Probab ility
Ten Coin toss 9 outcomes Outc ome Probab ility
Pascal's Triangle Pascal's friendChevalier de Mere 1654; Huygens 1657; Cardan /1 1/2 1/4 1/8 1/16 1/32 1/64 1/128 Frequency Probability
Fort Knox Toss Heads-Tails Gauss 1827 Series 1 Infinite outcomes
Variance )Measure of Spread )Easily calculated )Individual differences
Average squared deviation Normal distribution : xixi didi Variance = G d i 2 /N
Measuring Variation )Absolute differences? )Squared differences? )Absolute cubed? )Squared squared? Weighs & Means
Measuring Variation )Squared differences Ways & Means Fisher (1922) Squared has minimum variance under normal distribution
Covariance ) Measure of association between two variables ) Closely related to variance ) Useful to partition variance
Deviations in two dimensions :x:x :y:y
:x:x :y:y dx dy
Measuring Covariation )A square, perimeter 4 )Area 1 Area of a rectangle 1 1
Measuring Covariation )A skinny rectangle, perimeter 4 )Area.25*1.75 =.4385 Area of a rectangle
Measuring Covariation )Points can contribute negatively )Area -.25*1.75 = Area of a rectangle
Measuring Covariation Covariance Formula F = E (x i - : x )(y i - : y ) xy (N-1)
Correlation )Standardized covariance )Lies between -1 and 1 r = F xy 2 2 y x F * F
Summary Formulae : = ( E x i )/N F x = E (x i - :)/(N-1) 2 2 r = F xy 2 2 y x F * F F xy = E (x i -: x )(y i -: y )/(N-1)
Variance covariance matrix Several variables Var(X) Cov(X,Y) Cov(X,Z) Cov(X,Y) Var(Y) Cov(Y,Z) Cov(X,Z) Cov(Y,Z) Var(Z)
Conclusion )Means and covariances )Conceptual underpinning )Easy to compute )Can use raw data instead
Biometrical Model of QTL m d +a-a
Biometrical model for QTL Diallelic locus A/a with p as frequency of a
Classical Twin Studies )Summary: rmz & rdz )Basic model: A C E )rmz = A + C )rdz =.5A + C )var = A + C + E )Solve equations Information and analysis
Contributions to Variance )Additive QTL variance 5VA = 2p(1-p) [ a - d(2p-1) ]2 )Dominance QTL variance 5VD = 4p2 ( 1- p) 2 d2 )Total Genetic Variance due to locus VQ = V A + VD Single genetic locus
Origin of Expectations )P = aA + cC + eE )Standardize A C E )V P = a 2 + c 2 + e 2 )Assumes A C E independent Regression model
Path analysis )Two sorts of variable 5Observed, in boxes 5Latent, in circles )Two sorts of path 5Causal (regression), one-headed 5Correlational, two-headed Elements of a path diagram
Rules of path analysis )Trace path chains between variables )Chains are traced backwards, then forwards, with one change of direction at a double headed arrow )Predicted covariance due to a chain is the product of its paths )Predicted total covariance is sum of covariance due to all possible chains
ACE model MZ twins reared together
ACE model DZ twins reared together
ACE model DZ twins reared apart
Model fitting )Takes care of replicate statistics )Maximum likelihood estimates )Confidence intervals on parameters )Overall fit of model )Comparison of nested models
Fitting models to covariance matrices )MZ covariances 53 statistics V1 CMZ V2 )DZ covariances 53 statistics V1 CDZ V2 )Parameters: a c e )Df = nstat - npar = = 3
Model fitting to covariance matrices )Inherently compares fit to saturated model )Difference in fit between A C E model and A E model gives likelihood ratio test with df = difference in number of parameters
Confidence intervals )Two basic forms 5covariance matrix of parameters 5likelihood curve )Likelihood-based has some nice properties; squares of CIs on a give CI's on a 2 Meeker & Escobar 1995; Neale & Miller, Behav Genet 1997
Multivariate analysis )Comorbidity 5Partition into relevant components 5Explicit models 5One disorder or two or three )Longitudinal data analysis 5Partition into new/old 5Explicit models 5Markov 5Growth curves
Cholesky Decomposition )Provides a way to model covariance matrices )Always fits perfectly )Doesn't predict much else Not a model
Perverse Universe A E.7 P NOT!
Perverse Universe A E Y X r(X,Y)=0; Problem for almost any multivariate method
Analysis of raw data )Awesome treatment of missing values )More flexible modeling 5Moderator variables 5Correction for ascertainment 5Modeling of means )QTL analysis
Technicolor Likelihood Function For raw data in Mx j=1 ln L i = f i 3 ln [w j g(x i,: ij, G ij )] m x i - vector of observed scores on n subjects :ij - vector of predicted means Gij - matrix of predicted covariances - functions of parameters
Pihat Linkage Model for Siblings Each sib pair i has different COVARIANCE
Mixture distribution model Each sib pair i has different set of WEIGHTS p(IBD=2) x P(LDL1 & LDL2 | rQ = 1 ) p(IBD=1) x P(LDL1 & LDL2 | rQ =.5 ) p(IBD=0) x P(LDL1 & LDL2 | rQ = 0 ) Total likelihood is product of weighted likelihoods rQ=1rQ=.5 rQ=.0 weight j x Likelihood under model j
Conclusion )Model fitting has a number of advantages )Raw data can be analysed with greater flexibility )Not limited to continuous normally distributed variables
Conclusion II )Data analysis requires creative application of methods )Canned analyses are of limited use )Try to answer the question!