Experimental Design.

Slides:



Advertisements
Similar presentations
Experimental and Ex Post Facto Designs
Advertisements

Experimental Design and the struggle to control threats to validity.
Experimental and Quasi-Experimental Research
Experimental Design I. Definition of Experimental Design
Increasing your confidence that you really found what you think you found. Reliability and Validity.
Copyright © Allyn & Bacon (2007) Hypothesis Testing, Validity, and Threats to Validity Graziano and Raulin Research Methods: Chapter 8 This multimedia.
Defining Characteristics
GROUP-LEVEL DESIGNS Chapter 9.
Experimental Research Designs
EXPERIMENTAL DESIGNS Criteria for Experiments
Who are the participants? Creating a Quality Sample 47:269: Research Methods I Dr. Leonard March 22, 2010.
EXPERIMENTAL DESIGNS What Is Required for a True Experiment? What Are the Independent and Dependent Variables? What Is a Confounding Variable? What Are.
Statistics Micro Mini Threats to Your Experiment!
Sampling and Experimental Control Goals of clinical research is to make generalizations beyond the individual studied to others with similar conditions.
RESEARCH METHODS Lecture 35. EXPERIMENTAL RESEARCH [CONTINUED]
Chapter 9 Experimental Research Gay, Mills, and Airasian
Experimental Research
Experimental Research
Experimental Research Take some action and observe its effects Take some action and observe its effects Extension of natural science to social science.
Chapter 8 Experimental Research
Experimental Design The Gold Standard?.
Applying Science Towards Understanding Behavior in Organizations Chapters 2 & 3.
V ALIDITY IN Q UALITATIVE R ESEARCH. V ALIDITY How accurate are the conclusions you make based on your data analysis? A matter of degree Non-reification.
Design Experimental Control. Experimental control allows causal inference (IV caused observed change in DV) Experiment has internal validity when it fulfills.
EXPERIMENTAL DESIGN Experimental and Pre (Quasi) Experimental Designs.
Techniques of research control: -Extraneous variables (confounding) are: The variables which could have an unwanted effect on the dependent variable under.
Group Quantitative Designs First, let us consider how one chooses a design. There is no easy formula for choice of design. The choice of a design should.
Understanding Research Design Can have confusing terms Research Methodology The entire process from question to analysis Research Design Clearly defined.
Independent vs Dependent Variables PRESUMED CAUSE REFERRED TO AS INDEPENDENT VARIABLE (SMOKING). PRESUMED EFFECT IS DEPENDENT VARIABLE (LUNG CANCER). SEEK.
Experimental Design Chapter 1 Research Strategies and the Control of Nuisance Variables.
Research in Communicative Disorders1 Research Design & Measurement Considerations (chap 3) Group Research Design Single Subject Design External Validity.
Introduction section of article
Experimental Research
Chapter 6 Research Validity. Research Validity: Truthfulness of inferences made from a research study.
Chapter 10 Experimental Research Gay, Mills, and Airasian 10th Edition
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
Chapter 8 – Lecture 6. Hypothesis Question Initial Idea (0ften Vague) Initial ObservationsSearch Existing Lit. Statement of the problem Operational definition.
Introduction to Validity True Experiment – searching for causality What effect does the I.V. have on the D.V. Correlation Design – searching for an association.
Journalism 614: Experimental Methods Experimental Research  Take some action and observe its effects –Extension of natural science to social science.
CHAPTER 8 EXPERIMENTS.
Experimental Design and the struggle to control threats to validity.
Experimental Research Design Causality & Validity Threats to Validity –Construct (particular to experiments) –Internal –External – already discussed.
Criminal Justice and Criminology Research Methods, Second Edition Kraska / Neuman © 2012 by Pearson Higher Education, Inc Upper Saddle River, New Jersey.
Research designs Research designs Quantitative Research Designs.
Educational Research Experimental Research Chapter 9 (8 th Edition) Chapter 13 (7 th Edition) Gay and Airasian.
Experimental and Quasi-Experimental Research
Issues in Evaluating Educational Research
Experimental Research
Experiments Why would a double-blind experiment be used?
Experimental Research Designs
Experimental Design-Chapter 8
Hypothesis Testing, Validity, and Threats to Validity
Designing an Experiment
The Experiment Chapter 7.
CHAPTER 4 Designing Studies
Experiments and Quasi-Experiments
Social Research Methods Experimental Research
Quantitative Research
Hypothesis Testing, Validity &
Experiments and Quasi-Experiments
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
Chapter 18: Experimental and Quasi-Experimental Research
CHAPTER 4 Designing Studies
Experimental Design I. Definition of Experimental Design
Chapter 11 EDPR 7521 Dr. Kakali Bhattacharya
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
Presentation transcript:

Experimental Design

Experimental Design and the struggle to control threats to validity

LOW NATURALISTIC CASE-STUDY INCREASINGLY CONSTRAINED CORRELATIONAL DIFFERENTIAL EXPERIMENTAL HIGH

Experimental design is a planned interference in the natural order of events by the researcher.

What is Experimental Design? Experimental research allows us to test hypotheses and infer causality under controlled conditions designed to maximize internal validity. The high control and internal validity often mean a reduction of external validity. That is, the more precise, constrained, and artificial we become in the experiment, the less natural are the procedures and findings. The result is that we sometimes have difficulty generalizing experimentation to the natural environment, or a larger population.

Experimental Design Comparisons are made under different and controlled conditions. Subjects are assigned to each type of condition in an unbiased manner, usually matched or random. Although causality can often be inferred, results may not be applicable outside of the experimental setting Causality: If X, then Y AND If not X, then not Y

Experimental TIME 0: PRE-TEST. Collect baseline data. CONTROL TREATMENT #1 TREATMENT #2 TREATMENT #3 Experimental

TIME 1: TREATMENT GIVEN CONTROL JOLT COKE COFFEE Experimental

TIME 3: POST-TEST. Collect data on the effects of the treatment and compare to pretest, and to each treatment. CONTROL JOLT COKE COFFEE Experimental

Hypotheses NULL (Statistical) mean JOLT= mean COKE = mean COFFEE = mean CONTROL RESEARCH (Causal) Caffeine causes people to grow tall and nervous. CONFOUNDING (Rival) The differences are due to the amount of citric acid in the drinks.

A simple 2-group, posttest-only design Outfitters given low- impact training Outfitters given NO low-impact training Measure impacts caused by their clients Compare Scores

Hypotheses NULL (Statistical) mean Trained= mean Un-trained RESEARCH (Causal) The clients of outfitters trained in low impact methods will cause fewer impacts than clients of outfitters who did not receive such training. CONFOUNDING (Rival) Prior knowledge may have caused the observed differences in response.

Threats to Validity

Reliability Repeated use of the measure with identical subjects yields identical and consistent results. It is improved by: Clear conceptualization Precise measurement Multiple indicators Pilot-testing

Validity Internal Validity– design and measurement concerns that reduces chances for internal errors. External Validity– describes our ability and intent to generalize to subjects beyond our study sample. Largely an issue of design and sampling.

Validity and Reliability Reliable, NOT valid Valid, NOT Reliable Valid AND Reliable NOT Valid, NOT Reliable NOT Valid but Reliable Valid but UNReliable NOT Valid and UNReliable Valid and Reliable

Internal Validity Specifically, measurement validity Measures are valid for a single purpose Three types of validity: Face—as judged by others or by logic Content—captures the entire meaning of the experience Criterion—agrees with a validates, reliable external source: Concurrent, agrees with a preexisting measure Predictive, agrees with a future behavior or outcome

Validity Because of the confounding Hypothesis we are not 100% sure that our conclusions are valid. Did we indeed measure the effects of new knowledge? Independent -- Dependent? Other Variable/s -- Dependent?

Road map for research success Anticipate all threats to validity. Take plans to eliminate them. Statistical validity Construct validity External validity Internal validity.

Statistical validity The variations in the dependent variable are not due to variation in the subjects, but due to variations in the measuring instrument. The instrument is unreliable. Some statistical assumptions have been violated (e.g., non-normal data treated with parametric statistics; means of ordinal data, etc.).

Construct validity How well the observed data support the theory, and not a rival theory.

External validity The degree to which the findings of the study are valid for subjects outside the present study. The degree to which they are generalizable. Unbiased, complex sampling procedures; many studies, mid-constraint approaches help strengthen external validity. < external = > internal

Internal Validity Threats to causality (our ability to infer). Did the independent variable cause the dependent to change (were they related), or did some confounding variable intervene? Maturation, history, testing, instrumentation, regression to the mean, selection, attrition, diffusion and sequencing effects.

Maturation If the time between pre- and posttest is great enough to allow the subjects to mature, they will! Subjects may change over the course of the study or between repeated measures of the dependent variable due to the passage of time per se. Some of these changes are permanent (e.g., biological growth), while others are temporary (e.g., fatigue).

History Outside events may influence subjects in the course of the experiment or between repeated measures of the dependent variable. Eg., a dependent variable is measured twice for a group of subjects, once at Time A and again at Time B, and that the independent variable (treatment) is introduced between the two measurements. Suppose also that Event X occurs between Time A and Time B. If scores on the dependent measure differ at these two times, the discrepancy may be due to the independent variable or to Event X.

Testing Subjects gain proficiency through repeated exposure to one kind of testing. Scores will naturally increase with repeated testing. If you take the same test (identical or not) 2 times in a row, over a short period of time, you increase your chances of improving your score.

Instrumentation Changes in the dependent variable are not due to changes in the independent variable, but to changes in the instrument (human or otherwise). Measurement instruments and protocols must remain constant and be calibrated. Human observers become better, mechanical instruments become worse!

Regression to the mean If you select people based on extreme scores (High or low), in subsequent testing they will have scores closer to the mean (they would have regressed to the center).

Selection When random assignment or selection is not possible the two groups are not equivalent in terms of the independent variable/s. For example, males=treatment; females=control. Highest threats in naturalistic, case study and differential approaches.

Attrition When subjects are lost from the study. If random it may be OK. Confounding attrition is when the loss is in one group or because of the effects of the independent variable. (Jolt killed off 2 people!)

Diffusion of treatment When subjects communicate with each other (within and between groups) about the treatment) they diffuse the effects of the independent variable.

Sequencing effects The effects caused by the order in which you apply the treatment. A B C A C B B A C, etc.

Subject effects Subjects “know” what is expected of them, and behave accordingly (second guessing). Social desirability effect. Placebo effect. A placebo is a dummy independent effect. Some people react to it.

Experimenter effects Forcing the study to produce the desired outcome. Expectations of an outcome by persons running an experiment may significantly influence that outcome.

Single- and double-blind procedures Single blind –subjects don’t know which is treatment, which is not. Double blind—experimenter is also blind.

Designs One shot: One Group Pre-Post: Static Group: G1 T------O2 G1 O1------T------O2 Static Group: G1 T------O2 G2 (C) ------O2 T=Treatment O=Observation (measurement) C=Control

More designs Random Group: Pretest-Posttest, Randomized Group: RG1 T-------O RG2 (C) O Pretest-Posttest, Randomized Group: RG1 O1------T------O2 RG2 (C) O1------- ------O2 R=Random

Yet another design: Solomon four-group: RG1 O1------T------O2 RG2 O1-------------O2 RG3 T------O2 RG4 O2

One last one ! Latin Square: RG1 O T1 O T2 O T3 O RG2 O T2 O T3 O T1 O