Download presentation
Presentation is loading. Please wait.
1
Statistical Control Regression & research designs Statistical Control is about –underlying causation –under-specification –measurement validity –“Correcting” the bivariate relationship Regression & Residual formulas Correlations among y, x, y’, & residuals Control for Design & Measurement problems “Controlling” proxy variables Limitations & advantages of statistical control
2
Experimental “vs.” Statistical Control In the research design course (Psyc 941) we emphasized “experimental control” control produced by the actions of the researcher randomizing, balancing, holding constant, IV manipulation,etc. Another important form of control is “Statistical control” control produced by applying the proper statistical analyses usually used as an attempted substitute for “holding constant” Statistical control is generally considered inferior to experimental control, and often involves assumptions or changes in the research question that are “iffy”. But, the results from statistical control are often the first step toward an investment in acquiring experimental control over a “potential confounding variable”.
3
Originally, linear regression was developed for use with agricultural data collected as part of true experiments. The IV was usually some quantitative manipulation, such as amount of fertilizer, water, or seed and the DV was usually some assessment of “yield”. This was a manipulated IV, so the value of the predictor was known with great precision (“no error”) This was a true experiment, so causal interpretations were made of the IV - DV relationship -- so much of the IV could be counted on to produce so much of the DV. tons manure @ acre Bushel-yield @ acre RA & manipulate which plots receive how much manure
4
In behavioral research -- linear correlation ®ression are most often applied to natural groups designs (often called correlational designs - but be careful to keep in mind that the important relationship is between the type of design and causal interpretability, not between the type of stat and causal interpretability) A measured IV is used, so the value of the predictor is known with “more or less” precision (there is certainly some error) This is usually not a true experiment, so causal interpretations are not made of the IV - DV relationship -- so much IV leads to a prediction of so much DV. Linear prediction is still dependent upon a linear relationship (if two things aren’t linearly related, a linear model can’t be used to predict one from the other) Two big differences between the non- & experimental uses of regression: 1)precision of IV values 2) design & causal interp Each of these will show up as part of “statistical control”
5
Reviewing Variations of statistical control via residualization partial correlation -- correlation between two variables (x & y) controlling both for some 3 rd variable (z) -- r yx.z semi-partial correlation -- correlation between two variables (part correlation) (x & y) controlling one of the variables for some 3 rd variable (z) -- r y(x.z) & r x(y.z) Remember – the “3 rd variable” can be about causality, collinearity or measurement error
6
Control of multiple variables… multiple partial correlation -- like partial, but with “multiple 3 rd variables” -- r yx.zabc multiple semi-partial correlation -- like semi-partial, but with “multiple 3 rd variables” -- r y(x.zabc) vs. r x(y.zabc) ANCOVA -- ANalysis of COVAriance -- a kind of semi-partial or multiple semi-partial corr -- test of the DV mean difference between IV conditions that is independent of some 3 rd variable (called the “covariate”) -- r DV(IV.cov)
7
Statistical control is about the underlying causal model of the variables … X Y r Y,X Z 3 rd variable problem … relationship between x & y exists because a 3 rd variable is influencing both Z might be a causal or a psychometric influence X Y Z r Y,X βx βz Combining the two… Including X changes apparent Z-Y relationship because Z influences both X & Y X Y r Y,X Z βz βx Collinearity problem … by ignoring Z, the bivariate relationship mis-estimates X-Y relationship
8
Statistical control is about correctiong for under-specification… When we take a bivariate look at part of a multivariate picture … we’ll underestimate how much we can know about the criterion we’ll likely mis-estimate how that predictor relates to the criterion leaving predictors out usually leads to over-estimation r > β leaving predictors out changes the collinearity structure, and so, might cause us to miss suppressor effects r < β Statistical control in an attempt to improve this … what “would be” the bivariate relationship between these variables, in a population for which the control variable(s) is a constant (and so is not collinear with these variables) ? it is very much like looking at the β for that predictor in a multiple regression, but is in the form of a “corrected” simple correlation r y(x.z) ≈ β x from y’ = β x X + β z Z ≈ R 2 y.X,Z – R 2 y.X “What is the relationship between y and the part of x that is independent of Z?
9
Statistical control is about “correcting” the bivariate correlation to take the control variable(s) “into account” What do we get from this? Here’s where opinions differ … 1.A substitute for experimantal control ? 2.A better estimate of the causal relationship of the 2 variables ? 3.A substitute for construct validity ? 4.Solves under-specification problem ? 5.Probably a better description of the relationship of these 2 variables than is the bivariate analysis ? Rejection of 1-4 (probably for good reasons) has led some to reject the 5 th as well …but then what are we to do? Only perform bivariate analyses (known to be flawed) ? Expect to construct the “full story” from convergent research? (which is already the answer for “what’s the correct study”!!!) More complex models are, on average, more likely to be accurate!
10
Apparent Bivariate Correlation - 0 + Bivariate Correlation after Correction for 3 rd Variable - 0 + Some examples of taking the 3 rd variable into account
11
Of course, it can be uglier than that… Here, what “correction” you get depends upon which variable you control for – remember, larger models are only more accurate “on average” Here, the addition of the “Z” variable will help, but there is also a 3 rd – the interaction of X & Z
12
Here are two formula that contain “all you need to know” y’ = bx + a residual = y - y’ y = the criterionx = the predictor y’ = the predicted criterionvalue -- that part of the criterion that is related to the predictor ( Note: r yx = r yy’ ) residual = difference between criterion and predicted criterion values -- the part of the criterion not related to the predictor (Note: r x res = 0 or r x(y-y’) = 0 )
13
Summary of Interrelationships among the key values… XYY’Y - Y’ X--- y’ “is” x Y r yx--- so, y is related to y’ as to x Y’1.0 r yx--- Y - Y’0.0 (1 - r yx 2 )0.0 --- the residual is the part of the criterion not related to the y or y’ (which “is” x)
14
Examples of Statistical Control A design problem … We want to assess the correlation between the amount of homework completed and performance on exams. Since we can not RA participants to amount of homework, we are concerned that motivation will be a confound. Specifically, we think that motivation will influence both the number of homeworks each participant completes and how well they prepare for the exam. So, the question we want to ask can be phrased as, “What is the correlation between amount of homework completed and test performance that is independent of (not related to) motivation?” To do this we want to correlate the part of homeworks completed that is not related to motivation, with the part of test scores that is not related to motivation. E H M r H,E ?
15
Design control using Partial Correlation by residualization Step 1a predict # homeworks completed from motivation #hwks’ = b(motivation) + a Step 1b find the residual (part of #hwks not related to motivation) residual #hwks = #hwks - #hwks’ ( rem: r x (y-y’) = 0 ) Step 2a predict test scores from motivation test’ = b(motivation) + a Step 2b find the residual (part of test scores not related to motivation) residual test = test - test’ ( again: r x (y-y’) = 0 ) Step 3 correlate the residuals (the part of # hwks not related to motivation and the part of test scores not related to motivation) to find the relationship between # hwks and test scores that is independent of motivation
16
Design control using Semi-Partial Correlation by residualization Consider a version of this question for which we want to control only homework scores for motivation Step 1a predict # homeworks completed from motivation #hwks’ = b(motivation) + a Step 1b find the residual (part of #hwks not related to motivation) residual #hwks = #hwks - #hwks’ ( rem: r x (y-y’) = 0 ) Step 2 correlate the residual of homework (the part of # hwks not related to motivation) and original test scores to find the relationship between test scores and that part of homework scores that is independent of motivation. Notice that the difference between partial and semi-partial (part) correlations is that when doing the semi-partial we residualize only one of the variables we want to correlate.
17
Examples of Statistical Control A measurement problem … We want to assess the correlation between performance on quizzes and performance on the final exam. But we know that both of these variables include “test taking speed”. So, we are concerned that the correlation between the two types of performance will be “tainted” or “inflated”. To do this we want to correlate that part of quiz scores that is not related to test taking speed, with that part of final exam scores that is not related to test taking speed. So, the question we want to ask can be phrased as, “What is the correlation between quiz performance and final exam performance that is independent of (not related to) test taking speed?” Q E r Q,E S Knowledge at time of Quiz Knowledge at time of Exam ?
18
Measurement control using Partial Correlation by residualization Step 1a predict quiz scores from test taking speed quiz’ = b(speed) + a Step 1b find the residual (part of quiz scores not related to test taking speed) residual quiz = quiz - quiz’ Step 2a predict final exam scores from test taking speed final’ = b(speed) + a Step 2b find the residual (part of final exam scores not related to test taking speed) residual final = final - final’ Step 3 correlate the residuals (the part of quiz scores not related to speed and the part of final scores not related to speed) to find the relationship between quiz scores and final scores that is independent of test taking speed
19
Measurement control using Semi-Partial Correlation by residualization Consider a version of this question that involves controlling only quiz scores for test taking speed. Step 1a predict quiz scores from test taking speed quiz’ = b(speed) + a Step 1b find the residual (part of quiz scores not related to test taking speed) residual quiz = quiz - quiz’ ( rem: r x (y-y’) = 0 ) Step 2 correlate the residual of quiz scores (the part of quiz scores not related to speed) and original final scores to find the relationship between final scores that part of quiz scores that is independent of test taking speed Notice that the difference between partial and semi-partial (part) correlations is that when doing the semi-partial we residualize only one of the variables we want to correlate.
20
“Customized” Statistical Control… It is possible to “control” the X and Y variables for different variables, for example… controlling Final scores for test taking speed while controlling Homework scores for motivation & #practices. final’ = b(speed) + a residual final = final - final’ #hwks’ = b(motivation) b(#pract) + a residual #hwks = #hwks - #hwks’ Correlate residual final & residual #hwks Please note: 1) There is no MReg equivalent, 2) There must be a theoretical reason for doing this! Step 1 Step 2 Step 3
21
Statsitical control is an obvious help with under-specification, but it can also be help with proxy variables … Say we find r experf, hwperf =.4 we might ask for what is “homework performance” a proxy? Consider other things that are related to hwperf & add those to the model, to see if “the part of hwperf that isn’t them” is still related to experf… lots of possible results each with a different interp!!! r experf,(hwperf.mot) =.1 part of hwperf is “really” motivation (usual collinearity result) r experf,(hwperf.mot) =.4 none of hwper is motivation (mot and hwperf not collineary) r experf,(hwperf.mot) =.6 part of hwperf not related to mot is more corr with experf than hwperf (suppressor) Should continue with other variables can’t control everything, so we should focus on most important variables to consider as confounds or measurement confounds.
22
More about “Problems” with statistical control In analyses such as the last two examples, we have statistically changed the research question, for example. 1st e.g. -- Instead of asking, “What is the relationship between quiz scores and final exam scores that is independent of test taking speed in the population represented by the sample?” We are asking, “What would be the correlation between quiz scores and final exam scores in the population, if all the members of that population had the same test taking speed?” 2nd e.g. -- Instead of asking, “What is the relationship between the amount of homework completed and test performance that is independent of motivation?” We are asking, “What would be the correlation between amount of homework completed and test performance in the population, if all members of that population had the same motivation?” Populations such as these are unlikely to be representative !!
23
Advantages of Statistical Control 1st one -- Sometimes experimental control is impossible. Sometimes “intrusion free” measures aren’t available 2nd one -- Statistical control is often less expensive (might even be possible with available data, if the control variables have been collected). Doing the analysis with statistical control can help us decide whether or not to commit the resources to perform the experimental control or create better measures 3rd one -- often the only form of experimental control available is post-hoc matching (because you’re studying natural or intact groups). Preliminary analyses that explore the most effective variables for “statistical control” can help you target the most appropriate variables to match on. Can also retain sample size (hoc matching of mis-matched groups can decrease sample size dramatically)
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.