Research Designs Review of a few things Demonstrations vs. Comparisons Experimental & Non-Experimental Designs “IVs” and “DVs” Between Group vs. Within-Group.

Slides:



Advertisements
Similar presentations
Making Inferences about Causality In general, children who watch violent television programs tend to behave more aggressively toward their peers and siblings.
Advertisements

Validity (cont.)/Control RMS – October 7. Validity Experimental validity – the soundness of the experimental design – Not the same as measurement validity.
Non-experimental Designs
FUNDAMENTAL RESEARCH ISSUES © 2012 The McGraw-Hill Companies, Inc.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Slides to accompany Weathington, Cunningham & Pittenger (2010), Chapter 4: An Overview of Empirical Methods 1.
Experimental Design: Single factor designs Psych 231: Research Methods in Psychology.
Design Conditions & Variables Explicating Design Variables Kinds of “IVs” Identifying potential confounds Why control “on the average” is sufficient Characteristics.
Non-Experimental designs: Developmental designs & Small-N designs
Group Comparison Research
Non-Experimental designs: Developmental designs & Small-N designs
Prelude to the Research Validity Lecture A RH: is a guess about the relationships between behaviors In order to test our RH: we have to decide on a research.
Statistical Analysis of Factorial Designs z Review of Interactions z Kinds of Factorial Designs z Causal Interpretability of Factorial Designs z The F-tests.
Introduction to Path Analysis & Mediation Analyses Review of Multivariate research & an Additional model Structure of Regression Models & Path Models Direct.
Multivariate Analyses & Programmatic Research Re-introduction to Programmatic research Factorial designs  “It Depends” Examples of Factorial Designs Selecting.
Research Process Choices & Combinations of research attributes
Multivariate Analyses & Programmatic Research Re-introduction to Multivariate research Re-introduction to Programmatic research Factorial designs  “It.
Sampling & Experimental Control Psych 231: Research Methods in Psychology.
Experimental Control & Design Psych 231: Research Methods in Psychology.
2x2 ANOVA BG & MG Models Kinds of Factorial Designs ANOVA for BG, MG & WG designs Causal Interpretation of Factorial Results.
Prelude to Research Designs Review of a few things Demonstrations vs. Comparisons Experimental & Non-Experimental Designs “IVs” and “DVs” Between Group.
Multivariate Analyses & Programmatic Research Re-introduction to Multivariate research Re-introduction to Programmatic research Factorial designs  “It.
Review of Factorial Designs 5 terms necessary to understand factorial designs 5 patterns of factorial results for a 2x2 factorial designs Descriptive &
Non-Experimental designs
Design Conditions & Variables Limitations of 2-group designs “Kinds” of Treatment & Control conditions Kinds of Causal Hypotheses Explicating Design Variables.
Prelude to Internal Validity Types Of the four types of validity (Measurement, External, Internal & Statistical Conclusion), we will be most concerned.
Research Designs Review of a few things Demonstrations vs. Comparisons Experimental & Non-Experimental Designs “IVs” and “DVs” Can all causal RH: be tested.
Internal Validity Measured & Manipulated Variables & Constants Causes, Effects, Controls & Confounds Components of Internal Validity “Creating” initial.
Experiments and Observational Studies.  A study at a high school in California compared academic performance of music students with that of non-music.
1 Psych 5500/6500 Confounding Variables Fall, 2008.
WG ANOVA Analysis of Variance for Within- groups or Repeated measures Between Groups and Within-Groups ANOVA WG ANOVA & WG t-tests When to use each Similarities.
Copyright © 2010 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Experiments and Observational Studies. Observational Studies In an observational study, researchers don’t assign choices; they simply observe them. look.
Copyright © 2010 Pearson Education, Inc. Slide
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 13 Experiments and Observational Studies.
Experiment Basics: Variables Psych 231: Research Methods in Psychology.
ANOVA Analysis of Variance Picking the correct bivariate statistical test When, Why to Use ANOVA Summarizing/Displaying Data -- 1 qual & 1 quant var How.
Final Study Guide Research Design. Experimental Research.
Chapter 10 Experimental Research: One-Way Designs.
Types of validity we will study for the Next Exam... internal validity -- causal interpretability external validity -- generalizability statistical conclusion.
Slide 13-1 Copyright © 2004 Pearson Education, Inc.
1 Women’s Studies. 2 First Wave Feminism Mary Wollstonecraft’s (1792) A Vindication of the Rights of Women is one of the first written works to be called.
Today: Our process Assignment 3 Q&A Concept of Control Reading: Framework for Hybrid Experiments Sampling If time, get a start on True Experiments: Single-Factor.
Causal inferences During the last two lectures we have been discussing ways to make inferences about the causal relationships between variables. One of.
Between groups designs (2) – outline 1.Block randomization 2.Natural groups designs 3.Subject loss 4.Some unsatisfactory alternatives to true experiments.
Thinking About Psychology: The Science of Mind and Behavior.
STUDYING BEHAVIOR © 2009 The McGraw-Hill Companies, Inc.
An Introduction to Statistics and Research Design
Chapter 4 – Research Methods Different methods to answer different questions A.Does one factor cause another? The Scientific Method - obtain reliable information.
Introduction section of article
Research Methods It is actually way more exciting than it sounds!!!!
Research Strategies. Why is Research Important? Answer in complete sentences in your bell work spiral. Discuss the consequences of good or poor research.
Factorial Designs: Programmatic Research, RH: Testing & Causal Inference Applications of Factorial designs in Programmatic Research Research Hypotheses.
1.) *Experiment* 2.) Quasi-Experiment 3.) Correlation 4.) Naturalistic Observation 5.) Case Study 6.) Survey Research.
Chapter 10 Experimental Research Gay, Mills, and Airasian 10th Edition
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
Sources of Variance & ANOVA BG ANOVA –Partitioning Variation –“making” F –“making” effect sizes Things that “influence” F –Confounding –Inflated within-condition.
Non-Experimental designs
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
Finishing up: Statistics & Developmental designs Psych 231: Research Methods in Psychology.
Experimental and Ex Post Facto Designs
Varieties of Research Designs 3x3 Structure for single-IV designs –(3) Design differences & causal interpretability –(3) Design differences & statistical.
Causal inferences This week we have been discussing ways to make inferences about the causal relationships between variables. One of the strongest ways.
Experiments Textbook 4.2. Observational Study vs. Experiment Observational Studies observes individuals and measures variables of interest, but does not.
Psychological Experimentation The Experimental Method: Discovering the Causes of Behavior Experiment: A controlled situation in which the researcher.
EXPERIMENTAL RESEARCH
CHAPTER 4 Designing Studies
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
Chapter 4 Studying Behavior
Research Methods for the Behavioral Sciences
Presentation transcript:

Research Designs Review of a few things Demonstrations vs. Comparisons Experimental & Non-Experimental Designs “IVs” and “DVs” Between Group vs. Within-Group Designs

Reviewing a few things… Kinds of bivariate research hypotheses (and evidence to support) Kinds of Validity Two ways we “show” our studies have the validity we hope for... Associative research hypothesis Measurement Validity Statistical Conclusion Validity External Validity Internal Validity replication (same study) & convergence (variations) show a statistical relationship between the variables Causal research hypothesis temporal precedence statistical relationship between the variables no alternative explanation of the relationship - no confounds

Reviewing a few more things… What kind of validity relates to the “generalizability” of the results? What are the components of this type of validity? What validity relates to the “causal interpretability” of the results? What are the components of this type of validity & what type of variable is each involved with ? External Validity Population Setting Task/Stimulus Social/Temporal Internal Validity Initial Equivalence -- subject or measured variables Ongoing Equivalence -- procedural or manipulated variables

What are the three types of variable at the beginning of a study??? Causal variable Effect Variable Potential Confounds What are the five “types” at the end of the study??? Tell which are “good” and which are “bad” when testing causal RH: Causal variable Effect Variable Confound Variable Control Variable Constant To test a causal research hypothesis, a design must provide: manipulation of the causal variable measurement of the effect variable elimination of confounds/alternative hypotheses (I.e., everything that isn’t the causal or effect variable is either a constant or is a control variable)

For practice... Study purpose: to compare two different ways of teaching social skills (role playing vs. watching a videotape). Causal Variable? Effect Variable? Potential Confounds? Teaching methodSocial skillsAll other variables Study procedure: 10 pairs of 6th grade girls role-played an “initial meeting” while 20 8th grade girls watched a video about “meeting new people”. Then all the participants took a social skills test. Any controls (var or const.) ? Any confounding variables? How do you know what variables to control, so that they don’t become confounds? Can we causally interpret the results ? Gender -- constant Age/grade difference Any variable not the causal variable must be controlled Nope -- confounds!

There are two basic ways of providing evidence to support a RH: -- a “demonstration” and a “comparison” a demonstration involves using the treatment and showing that the results are “good” a comparison (an experiment) involves showing the difference between the results of the treatment and a “control” lots of commercials use demonstrations We washed these dirty clothes in Tide -- see how clean !!! After taking Tums her heartburn improved !!! He had a terrible headache. After taking Tylenol he’s dancing with his daughter! The evidence from a demonstration usually meets with the response -- “Compared to what ??” a single demonstration is a “implicit” comparison “doesn’t this wash look better then yours ?” “did you last heartburn improve this fast ?” “didn’t your last headache last longer than this ?” explicit comparisons are preferred !!!

When testing causal RH: we must have a “fair comparison” or a “well-run Experiment” that provides init eq of subject variables & ongoing eq of procedural variables For example what if our experiment intended to show that Tide works better compared… Really dirty light-colored clothes washed in a small amount of cold water for 5 minutes with a single rinse -- using Brand-X Barely dirty dark-colored clothes washed in a large amount of hot water for 25 minutes with a double rinse -- using Tide vs. Can you separate the initial and ongoing equivalence confounds ? “dirtyness” of clothes color of clothes amount of water length of washing single vs. double rinse What is supposed to be the “causal variable” that produces the difference in the cleanness of the two loads of clothes? Initial Equivalence confounds Ongoing Equivalence confounds

True Experiment random assignment of individual participants by researcher before IV manip (provides initial equivalence - subject variables - internal validity) treatment/manipulation performed by researcher (provides temporal precedence & ongoing equivalence - internal validity) good control of procedural variables during task completion & DV measurement (provides ongoing equivalence - internal validity) Quasi-Experiment no random assignment of individuals (but perhaps random assignment of intact groups) treatment/manipulation performed by researcher poor or no control of procedural variables during task, etc. Natural Groups Design also called Concomitant Measures or Correlational Design no random assignment of individuals (already in “IV groups”) no treatment manipulation performed by researcher (all variables are measured) -- a comparison among participants already in groups no control of procedural variables during task, etc. Research Designs True Experiments If “well-done,” can be used to test causal RH: -- alternative hyp. are ruled out because there are no confounds !!! Non-Experiments No version can be used to test causal RH: -- can’t rule out alternative hyp. Because there are confounds !!

Words of Caution About the terms “IVs”, “DVs” & causal RH:s... You might have noticed that we’ve not yet used these terms.. Instead we’ve talked about “causal variables” and “effect variables” -- as you probably remember.. –the Independent Variable (IV) is the “causal variable” –the Dependent Variable (DV) is the “effect variable ” However, from the last slide, you have know that we can only say the IV causes the DV if we have a true experiment (and the internal validity it provides) –initial equivalence (control of subject variables) random assignment of participants –ongoing equivalence (control of procedural variables) experimenter manipulates IV, measures DV and controls all other procedural variables

The problem seems to come from there being at least three different meanings or uses of the term “IV”... 1 “the variable manipulated by the researcher” it’s the “IV” because it is “independent” of any naturally occurring contingencies or relationships between behaviors the researcher, and the researcher alone, determines the value of the IV for each participant 2 “the grouping, condition, or treatment variable” 3 “the presumed causal variable in the cause-effect relationship” In these last two both the “IV” & “DV” might be measured !!! So… you don’t have a True Experiment... no IV manipulation to provide temporal precedence no random assignment to provide init. eq. for subject vars no “control” to provide onging eq. for procedural variables … and can’t test a causal RH:

This is important stuff -- so here’s a different approach... It is impossible to have sufficient internal validity to infer cause when studying some IV-DV relationships Say we wanted to test the idea that attending private colleges CAUSES people to be more politically conservative than does attending public universities. –We wouldn’t be able to randomly assign folks to the type of college they attend (no initial eq.) –We wouldn’t be able to control all the other things that happen during those 4 years (no ongoing equivalence) Here are some other categories of “IV”s with the same problem… –gender, age, # siblings –ethnic background, race, neighborhood –characteristics/behaviors of your parents –things that happened earlier in your life

IVs “vs” Confounds Both IVs and Confounds are “causal variables” !!! variables that may cause (influence, etc. ) scores on the DVs What’s the difference ??? The IV is the intended causal variable in the study! We are trying to study if & how & how much the IV influences the DV ! A confound interferes with our ability to study the causal relationship between the IV & the DV, because it is another causal variable that might be influencing the DV. If the IV difference between the conditions is confounded, then if there is a DV difference between the conditions, we don’t know if that difference was caused by the IV, the confound or a combination of both !!!!

Between Groups vs. Within-Groups Designs Between Groups also called Between Subjects or Cross-sectional each participant is in one (& only one) of the treatments/conditions different groups of participants are in each treatment/condition typically used to study “differences” -- when, in application, a participant will usually be in one treatment/condition or another Within-Groups Designs also called Within-Subjects, Repeated Measures, or Longitudinal each participant is in all (every one) of the treatment/conditions one group of participants, each one in every treatment/condition typically used to study “changes” -- when, in application, a participant will usually be moving from one condition to another

Between Groups Design Within-Groups Design Experimental Traditional Tx Tx Pat Sam Kim Lou Todd Bill Glen Sally Kishon Phil Rae Kris Experimental Traditional Tx Tx Pat Sam Kim Lou Todd Bill Pat Sam Kim Lou Todd Bill Different participants in each treatment/condition All participants in each treatment/condition

True Experiment w/ “proper” RA/CB - init eqiv manip of IV by researcher Between Groups (dif parts. in each IV condition) Within-Groups (each part. in all IV conditions) Results might be causally interpreted -- if good ongoing equivalence Research Designs Putting this all together -- here’s a summary of the four types of designs we’ll be working with... Results can not be causally interpreted Results might be causally interpreted -- if good ongoing equivalence Results can not be causally interpreted Non-experiment no or poor RA/CB may have IV manip

Four versions of the same study … which is which? Each participant in our “object identification study” was asked to select whether they wanted to complete the “visual” or the “auditory” condition. Each participant in our “object identification study” was randomly assigned to complete either the “visual” or the “auditory” condition. Each participant in our “object identification study” completed both the “visual” and the “auditory” conditions in a randomly chosen order for each participant. Each participant in our “object identification study” completed first the “visual” and the the “auditory” condition. BG Non WG Exp BG Exp. WG Non

So, you gotta have a True Experiment for the results to be causally interpretable? But, does running a “True Experiment” guarantee that the results will be causally interpretable? What are the elements of a True Experiment?? Random Assignment if Individuals to IV conditions by the researcher before manipulation of the IV Supposed to give us initial equivalence of measured/subject variables. Manipulation of the IV by the researcher Supposed to give us temporal precedence & help control ongoing equivalence of manipulated/procedural variables Please note: A “true experiment” is defined by these two elements! BUT  there is “an asymmetry” between “true exp” and “causal interp” Huh? True Exp is necessary, but not sufficient, for causal interpretability!

What could possibly go wrong …. ??? Random Assignment “might not take” RA is a “probabilistic process”  there’s no guarantee that the groups will be equivalent on all subject variables! Might introduce a confound when doing the IV manipulation might treat the conditions differently other than the IV May “miss” or even “cause” other ongoing equivalence confounds often, especially for younger researchers or newer research topics, we don’t really know what to “control” we may know what to control and just not get it done…

If only True Experiments can be causally interpreted, why even bother running non-experiments? 1 st Remember that we can’t always run a true experiment ! Lots of variables we care about can’t be RA & manip – gender, family background, histories and experiences, personality, etc. Even if we can RA & manip, lots of studies require long-term or field research that makes ongoing equivalence (also required for causal interp) very difficult or impossible. We would greatly limit the information we could learn about how variables are related to each other if we only ran studies that could be causally interpreted.

If only True Experiments can be causally interpreted, why even bother running non- experiments? Cont… 2nd We get very useful information from non-experiments ! True, if we don’t run a True Experiment, we are limited to learning predictive information and testing associative RH: But associative information is the core of our understanding about what variables relate to each other and how they relate Most of the information we use in science, medicine, education, politics, and everyday decisions are based on only associative information – and things go pretty well! Also, designing and conducting True Experiments is made easier if we have a rich understanding of what variables are potential causes and confounds of the behavior we are studying

Between Groups True Experiment Untreated Population Treated Population participant pool to-be-treated group “control group” “experimental group” participant selection random participant assignment treatment not-to-be- treated group no treatment Rem -- samples & “groups” are intended to represent populatioins

Within-Groups True Experiment Untreated Population Treated Population participant selection participant pool random participant assignment 1/2 of subjects untreated treated Each participant represents each target population, in a counterbalanced order

The design has the external validity advantage that each subject REALLY is a member of the population of interest (but we still need a representative sample) The design has the internal validity disadvantages that... we don’t know how participants “end up” in the populations no random participant assignment (no initial equivalence) we don’t know how the populations differ in addition to the treatment per se no control of procedural variables (no ongoing equivalence) Between Groups Non-experiment Untreated Population Treated Population participant selection “control group” “experimental group”

Within-Groups Non-experiment Untreated Population Treated Population treatment occurs to the whole population “control group” treatment group The design has the external validity advantage that each subject REALLY is a member of each population of interest (but we still need a representative sample) The design has the internal validity disadvantages that... we don’t know how the populations differ in addition to the treatment per se no control of procedural variables (no ongoing equivalence) participant selection

There is always “just one more thing”... Sometimes there is no counterbalancing in a Within-groups design, but there can still be causal interpretation… A good example is when the IV is “amount of practice” with “10 practice” and a “50 practice” conditions. There is no way a person can be in the 50 practice condition, and then be in the 10 practice condition Under these conditions (called a “seriated IV”), what matters is whether or not we can maintain “ongoing equivalence” so that the only reason for a change in performance would be the increased practice The length of time involved is usually a very important consideration Which of these would you be more comfortable giving a causal interpretation? When we gave folks an initial test, 10 practice and then the test again, we found that at their performance went up! When we gave folks an initial assessment, 6 months of once-a-week therapy and then the assessment again, their depression went down!