Download presentation
Presentation is loading. Please wait.
Published byAdrianna Janeway Modified over 10 years ago
1
Chapter 7 Flashcards
2
overall plan that describes all of the elements of a research or evaluation study, and ideally the plan allows the researcher or evaluator to reach valid conclusions (e.g., questions or hypotheses to be addressed, number and types of participants to be included, number and types of variables to be studied, collection and analysis of data) Evaluation/research design
3
family of research and evaluation designs characterized by the systematic repeated measurement of a clients outcome(s) at regular, frequent, predesignated intervals under different conditions (baseline and intervention), and the evaluation of outcomes over time and under different conditions in order to monitor client progress, identify intervention effects, and more generally, learn when, why, how, and the extent to which client change occurs. Also known as single-subject designs, single- system designs, N = 1 designs, or sometimes time series or interrupted time series designs Single-case design
4
period of time during which an outcome is measured repeatedly in the absence of an intervention in order to (1) describe the naturally occurring pattern of outcome data (e.g., level, trend, variability) and (2) determine the effect of an intervention on that outcome. Typically symbolized by the letter A Baseline phase
5
period of time during which an intervention is implemented while an outcome is measured repeatedly Intervention phase
6
period of time after an intervention has ended during which outcome data are collected to determine the extent to which a clients progress has been maintained. Also known as a maintenance phase Follow-up phase
7
a variable (e.g., intervention) that produces an effect or is responsible for events or results (e.g., outcome) Cause
8
change in one variable (e.g., outcome) that occurred at least in part as the result of another variable (e.g., intervention) Effect
9
measure of the strength of the relationship between variables (e.g., effect of an intervention on an outcome, as quantified by any one of a number of different statistics). Effect size
10
conclusion based on evidence and reasoning that one variable (e.g., intervention) causes another (e.g., outcome) Causal inference
11
accuracy of conclusions based on evidence and reasoning about the presence, direction, and strength of relationships between variables (e.g., outcome is different during baseline than intervention). The ability to establish that one variable (e.g., intervention) is related to another (e.g., outcome) is a requirement for inferring that one variable caused another Statistical conclusion validity
12
accuracy of conclusions based on evidence and reasoning about causal relationships between variables (e.g., extent to which an intervention, as opposed to other factors, caused a change in an outcome). Internal validity
13
uncertainty about which of several events or processes caused an outcome Causal ambiguity
14
variable that is associated with the independent variable inadvertently influences the outcome, and consequently makes it difficult to determine the effect of the intervention on the outcome (e.g., an unknown event that occurs during intervention but not baseline and causes the pattern of outcome data to change from baseline to intervention). Also known as a confound or confounding variable, and such a result is said to be confounded Extraneous variable
15
plausible reasons for a relationship between an intervention and an outcome, other than that the intervention caused the outcome. Also known as alternative hypotheses. The ability to rule out alternative hypotheses is a requirement for inferring that one variable (intervention) caused another (outcome) Alternative explanations
16
reasons why it might be partly or completely wrong (i.e., invalid) to conclude that one variable (e.g., an intervention) caused another (e.g., an outcome). See also Ambiguous temporal precedence, History effect, Instrumentation effect, Maturation effect, Regression effect, and Testing effect Threats to internal validity
17
potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by an external event that occurs at the same time as the intervention (e.g., a student who has trouble completing his homework for a new teacher improves his performance as he becomes accustomed to her and her expectations, and the improvement is misinterpreted as being due to the rewards he earns working with a social worker). History effect
18
potential threat to internal validity in which an apparent change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by a change in how the outcome is measured (e.g., an older mans weight stabilizes when he begins weighing at his physicians office rather than at home). Instrumentation effect
19
potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by naturally occurring changes in clients over time (e.g., a toddlers tantrums diminish not in response to an intervention, but due to maturing out of the terrible twos, or a child outgrows enuresis). Maturation effect
20
potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by repeated measurement of the outcome (e.g., a pretest about health behaviors may sensitize a client to the need to make changes in diet and exercise). See also Fatigue effect and Practice effect. Testing effect
21
deterioration in an outcome caused by fatigue associated with repeated measurement of the outcome (e.g., a mother who is self-recording each instance of time-out with her preschooler reduces those time-outs to avoid recording the behavior) Fatigue effect
22
improvement in an outcome caused by repeated measurement of the outcome (e.g., taking multiple practice exams may improve a students score on the Graduate Record Exam simply because he or she becomes familiar with the format of the exam) Practice effect
23
potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by the tendency of an individual with unusually high or low scores on a measure to subsequently have scores closer to the mean (e.g., clients who are depressed frequently seek help when they have hit bottom, and their scores are likely to improve somewhat in the following weeks, even without intervention). Also known as regression toward the mean Regression effect
24
result due to the order in which different interventions are administered (e.g., a couple may be more successful in an intervention designed to increase their pleasant time together each day if they first complete an intervention designed to increase their reflective listening). Also known as a sequence effect Order effect
25
single-case design (arguably) consisting of an intervention phase (B) during which the outcome is measured repeatedly B-only design
26
single-case design (arguably, since there isnt repeated measurement during baseline) consisting of one pre-intervention outcome measurement followed by an intervention phase (B) during which the outcome is measured repeatedly B+ design
27
two-phase single-case design consisting of a pre-intervention baseline phase (A) followed by an intervention phase (B) A-B design
28
three-phase single-case design consisting of a pre-intervention baseline phase (A 1 ); an intervention phase (B); and a second baseline phase (A 2 ) in which the intervention is withdrawn to determine if the outcome reverses to the initial baseline pattern A-B-A design
29
three-phase single-case design beginning with the intervention phase (B 1 ), followed by the withdrawal of the intervention (A) to determine if the outcome changes in the absence of the intervention, and reintroduction of the intervention (B 2 ) to see whether the initial intervention effects are replicated B-A-B design
30
four-phase single-case design consisting of a pre- intervention baseline phase (A 1 ); an intervention phase (B 1 ); a second baseline phase (A 2 ) in which the intervention is withdrawn to determine if the outcome reverses to the initial baseline pattern; and a reintroduction of the intervention (B 2 ) to see whether the initial intervention effects are replicated. Also known as a reversal or withdrawal design A-B-A-B design
31
single-case design that begins with a baseline during which the same problem is measured for a single client in two or more settings at the same time. Baseline is followed by the application of the intervention in one setting while baseline conditions remain in effect for other settings, then the intervention is applied sequentially across the remaining settings to see whether intervention effects are replicated across different settings Multiple baseline across settings design
32
single-case design that begins with a baseline during which the same problem is measured for two or more clients at the same time in a particular setting. Baseline is followed by the application of the intervention to one client while baseline conditions remain in effect for other clients, then the intervention is applied sequentially to remaining clients to see whether intervention effects are replicated across different clients Multiple baseline across subjects (clients) design
33
single-case design that begins with a baseline during which two or more problems are measured at the same time for a single client in a particular setting. Baseline is followed by the application of the intervention to one problem with baseline conditions remaining in effect for other problems, then the intervention is applied sequentially to the remaining problems to see whether intervention effects are replicated across different problems Multiple baseline across behaviors (problems) design
34
three-phase single-case design consisting of a pre-intervention baseline (A); an intervention phase (B); and a second intervention phase (C) in which a new intervention is introduced in response to the failure of the first intervention to produce sufficient improvement in the outcome A-B-C design
35
three-phase single-case design consisting of a pre-intervention baseline (A); an intervention phase (B); and a second intervention phase in which a new intervention (C) is added to the first intervention in response to the failure of the first intervention to produce sufficient improvement in the outcome A-B-BC design
36
accuracy of conclusions based on evidence and reasoning about how well a causal relationship applies across or beyond the people, settings, treatment variables, and measurement variables that were studied (e.g., extent to which a causal relationship between an intervention and outcome is the same with different people) External validity
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.