1 Statistics 262: Intermediate Biostatistics Mixed models; Modeling change.

Slides:



Advertisements
Similar presentations
A. The Basic Principle We consider the multivariate extension of multiple linear regression – modeling the relationship between m responses Y 1,…,Y m and.
Advertisements

Lesson 10: Linear Regression and Correlation
Logistic Regression I Outline Introduction to maximum likelihood estimation (MLE) Introduction to Generalized Linear Models The simplest logistic regression.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Simple Logistic Regression
Topic 15: General Linear Tests and Extra Sum of Squares.
EPI 809/Spring Probability Distribution of Random Error.
CHAPTER 24: Inference for Regression
GEE and Mixed Models for longitudinal data
April 25 Exam April 27 (bring calculator with exp) Cox-Regression
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Multiple regression analysis
Regression Analysis. Unscheduled Maintenance Issue: l 36 flight squadrons l Each experiences unscheduled maintenance actions (UMAs) l UMAs costs $1000.
Chapter 11 Survival Analysis Part 3. 2 Considering Interactions Adapted from "Anderson" leukemia data as presented in Survival Analysis: A Self-Learning.
Analysis – Regression The ANOVA through regression approach is still the same, but expanded to include all IVs and the interaction The number of orthogonal.
Chapter 11 Survival Analysis Part 2. 2 Survival Analysis and Regression Combine lots of information Combine lots of information Look at several variables.
EPI 809/Spring Multiple Logistic Regression.
Logistic Regression Biostatistics 510 March 15, 2007 Vanessa Perez.
Predictive Analysis in Marketing Research
Incomplete Block Designs
This Week Continue with linear regression Begin multiple regression –Le 8.2 –C & S 9:A-E Handout: Class examples and assignment 3.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
SAS Lecture 5 – Some regression procedures Aidan McDermott, April 25, 2005.
Chapter Outline 5.0 PROC GLM
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
© 2002 Prentice-Hall, Inc.Chap 14-1 Introduction to Multiple Regression Model.
Biostatistics Case Studies 2005 Peter D. Christenson Biostatistician Session 4: Taking Risks and Playing the Odds: OR vs.
GROUP DIFFERENCES: THE SEQUEL. Last time  Last week we introduced a few new concepts and one new statistical test:  Testing for group differences 
1 Experimental Statistics - week 10 Chapter 11: Linear Regression and Correlation.
Topic 14: Inference in Multiple Regression. Outline Review multiple linear regression Inference of regression coefficients –Application to book example.
1 Experimental Statistics - week 10 Chapter 11: Linear Regression and Correlation Note: Homework Due Thursday.
Regression Examples. Gas Mileage 1993 SOURCES: Consumer Reports: The 1993 Cars - Annual Auto Issue (April 1993), Yonkers, NY: Consumers Union. PACE New.
Regression For the purposes of this class: –Does Y depend on X? –Does a change in X cause a change in Y? –Can Y be predicted from X? Y= mX + b Predicted.
Correlation and Linear Regression. Evaluating Relations Between Interval Level Variables Up to now you have learned to evaluate differences between the.
Biostatistics Case Studies 2007 Peter D. Christenson Biostatistician Session 3: Incomplete Data in Longitudinal Studies.
April 6 Logistic Regression –Estimating probability based on logistic model –Testing differences among multiple groups –Assumptions for model.
Multilevel Linear Models Field, Chapter 19. Why use multilevel models? Meeting the assumptions of the linear model – Homogeneity of regression coefficients.
Different Distributions David Purdie. Topics Application of GEE to: Binary outcomes: – logistic regression Events over time (rate): –Poisson regression.
Linear correlation and linear regression + summary of tests
April 4 Logistic Regression –Lee Chapter 9 –Cody and Smith 9:F.
6-1 Introduction To Empirical Models Based on the scatter diagram, it is probably reasonable to assume that the mean of the random variable Y is.
+ Chapter 12: More About Regression Section 12.1 Inference for Linear Regression.
Multigroup Models Byrne Chapter 7 Brown Chapter 7.
BUSI 6480 Lecture 8 Repeated Measures.
Chapter 16 Data Analysis: Testing for Associations.
Right Hand Side (Independent) Variables Ciaran S. Phibbs June 6, 2012.
PSYC 3030 Review Session April 19, Housekeeping Exam: –April 26, 2004 (Monday) –RN 203 –Use pencil, bring calculator & eraser –Make use of your.
Right Hand Side (Independent) Variables Ciaran S. Phibbs.
Statistical Analysis Topic – Math skills requirements.
Chapter 10 The t Test for Two Independent Samples
1 Experimental Statistics - week 12 Chapter 12: Multiple Regression Chapter 13: Variable Selection Model Checking.
General Linear Model.
Smith/Davis (c) 2005 Prentice Hall Chapter Fifteen Inferential Tests of Significance III: Analyzing and Interpreting Experiments with Multiple Independent.
1 Statistics 262: Intermediate Biostatistics Regression Models for longitudinal data: Mixed Models.
1 Modeling change Kristin Sainani Ph.D. Stanford University Department of Health Research and Policy
Biostatistics Case Studies Peter D. Christenson Biostatistician Session 3: Missing Data in Longitudinal Studies.
G Lecture 71 Revisiting Hierarchical Mixed Models A General Version of the Model Variance/Covariances of Two Kinds of Random Effects Parameter Estimation.
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
1 Experimental Statistics - week 12 Chapter 11: Linear Regression and Correlation Chapter 12: Multiple Regression.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
1 Experimental Statistics - week 11 Chapter 11: Linear Regression and Correlation.
Regression. Why Regression? Everything we’ve done in this class has been regression: When you have categorical IVs and continuous DVs, the ANOVA framework.
Inferential Statistics Psych 231: Research Methods in Psychology.
Applied Biostatistics: Lecture 2
Multiple Regression.
CHAPTER 29: Multiple Regression*
I. Statistical Tests: Why do we use them? What do they involve?
What’s the plan? First, we are going to look at the correlation between two variables: studying for calculus and the final percentage grade a student gets.
Presentation transcript:

1 Statistics 262: Intermediate Biostatistics Mixed models; Modeling change

2 Within vs. Between subject effects… Solution for Fixed Effects Standard Effect Estimate Error DF t Value Pr > |t| Intercept time chem Example 1: last week’s example. CHEM: A significant chemical effect (time-dependent predictor) could either represent a between-subjects effect or a within-subjects effect. Because these data were made-up, we happened to know that there was primarily a within-subjects effect…

3 Example 1..from last time… id time1 time2 time3 time4 chem1 chem2 chem3 chem patients with depression are given a drug that increases levels of a “happy chemical” in the brain. At baseline, all 6 patients have similar levels of this happy chemical and scores >=14 on a depression scale. Researchers measure depression score and brain-chemical levels at three subsequent time points: at 2 months, 3 months, and 6 months post-baseline. Here are the data in broad form:

4 Example 2… Same as example 1, but made up to have more between-subjects effect than within-subjects effect. id time1 time2 time3 time4 chem1 chem2 chem3 chem

5 Example 3… Same as example 1, but made up to have ONLY between-subjects effect id time1 time2 time3 time4 chem1 chem2 chem3 chem

Example 2 (more between-subjects effects): showing two subjects… id=1: score id=2: chem id=1: chem id=2: score

Example 2:

Example 3:all between-subjects effects

Example 3:

19 Results, example 1… Solution for Fixed Effects Standard Effect Estimate Error DF t Value Pr > |t| Intercept chem time Example 1: significant chem effect. CHEM: proc mixed data=hrp262.long2; model score=chem time / solution; random int/subject=id; run;

20 Results, example 2… Solution for Fixed Effects Standard Effect Estimate Error DF t Value Pr > |t| Intercept <.0001 chem time Example 2: significant chem effect CHEM: proc mixed data=hrp262.long2; model score=chem time / solution; random int/subject=id; run;

21 Results, example 3… The Mixed Procedure Solution for Fixed Effects Standard Effect Estimate Error DF t Value Pr > |t| Intercept <.0001 chem <.0001 time Example 3: significant chem effect CHEM: proc mixed data=hrp262.long3; model score=chem time / solution; random int/subject=id; run;

22 All three examples give the same result… There’s no way to tell if change in chemical levels is causing change in depression score (which is what we probably care about). What to do?...

23 Options Examine graphs! Evaluate baseline relationship of chemical1 and score1 using regular linear regression Here, you will find a strong relationship only in examples 2 and 3, suggesting strong between- subjects effects. Drawback: you still can’t rule out within-subjects effects (could have both) Use only baseline value of chemical (chem1) as a predictor in GEE or Mixed and add a time*chem1 interaction to the model to evaluate change. Here, you will find that chem1 main effect is significant but chem1*time is not in examples 2 and 3 Drawback: A significant time*chem1 interaction would indicate that baseline chemical levels predict change in depression score over time, which is slightly different than saying that change in chemical level predicts change in depression score. Correlate change in time-dependent predictor with change in repeated-measures outcome… Calculate overall change or percent change in outcome and regress this on overall change or percent change in the predictor: see chapter 8 of Twisk (2 time points only) OR model all the changes together (vector of changes)…

24 The change model

25 SAS code to change data… data hrp262.change; set hrp262.broad; time=0; ctime=2; cscore=time2-time1; cchem=chem2-chem1; output; time=1; ctime=1; cscore=time3-time2; cchem=chem3-chem2; output; time=2; ctime=3; cscore=time4-time3; cchem=chem4-chem3; output; label cchem='change in chemical'; label cscore='change in depression score'; run;

id cscore cchem time id time1 time2 time3 time4 chem1 chem2 chem3 chem So, look at change in depression score as your outcome variable. All time- dependent predictors also get a change score.

27 Example 1: naïve linear regression cscore = *cchem

28 Example 2: naïve linear regression cscore = *cchem

29 Example 3: naïve linear regression cscore = *cchem

30 Modeling changes (mixed)… EXAMPLE 1 : Solution for Fixed Effects Standard Effect Estimate Error DF t Value Pr > |t| Intercept time cchem EXAMPLE 2: Solution for Fixed Effects Standard Effect Estimate Error DF t Value Pr > |t| Intercept time cchem EXAMPLE 3: Solution for Fixed Effects Standard Effect Estimate Error DF t Value Pr > |t| Intercept time cchem

31 Modeling changes (GEE)… EXAMPLE 1: Analysis Of GEE Parameter Estimates Standard 95% Confidence Parameter Estimate Error Limits Z Pr > |Z| Intercept time cchem <.0001 EXAMPLE 2: Analysis Of GEE Parameter Estimates Standard 95% Confidence Parameter Estimate Error Limits Z Pr > |Z| Intercept time cchem EXAMPLE 3: Analysis Of GEE Parameter Estimates Standard 95% Confidence Parameter Estimate Error Limits Z Pr > |Z| Intercept cchem time

32 The change model In fact, it often turns out that the changes are not correlated within subjects (we’ve essentially already corrected for between subject variability by using change scores). e.g., just regular old linear regression, where each person contributes three “independent” observations…

33 The change model If time intervals are unequal and unbalanced, you might choose to include the change in time as a predictor in the model. Otherwise, you could face confounding by the time between measurements (which might be correlated with the time-dependent predictor)…

34 SAS code proc genmod data=hrp262.change; class id; model cscore=time cchem; repeated subject=id / type=ind corrw; run; proc mixed data=hrp262.change; class id; model cscore=time cchem/ solution; run; proc reg data=hrp262.change; model cscore=time cchem; run; Which are basically equivalent to good old linear regression…

35 Another example… In the runners study, does weight gain increase BMD?

36 Weight (time-dependent predictor)… Solution for Fixed Effects Standard Effect Estimate Error DF t Value Pr > |t| Intercept <.0001 time pounds <.0001 Weight has a highly significant relationship with spine BMD… Is this a between-subjects or within-subjects effect??

Between vs. within subjects… Strong between-subjects effect (baseline comparison) Weak within-subjects effect (comparison of changes) Deletion of just a few subjects who gained large amounts of weight makes the line much less steep.

38 Modeling changes… proc mixed data=hrp262.bfitchange; class id ; model cspine= ctime cpounds /s; random int /subject=id; run; It makes more sense here to model time as the interval (in months) between DXA (bone) measurements rather than regular old time. The change in BMD between DXAs is correlated to the length of time between DXAs (which was supposed to be 1 year, but ended up being highly variable). It’s not unreasonable to believe that weight gain might also be related to the amount of time that’s passed. If, instead, dxa (clinical visit 1,2,3) or time (continuous) were put in the model, what would a significant effect mean? data hrp262.change; set hrp262.broad; dxa=1; ctime=(dxaday2-dxaday1)*12/365.25; cspine=spine2-spine1; cpounds=pounds2-pounds1; output; dxa=2; ctime=(dxaday3-dxaday2)*12/365.25; cspine=spine2-spine1; cpounds=pounds2-pounds1; output; run;