SOCI 4466 PROGRAM & POLICY EVALUATION LECTURE #8 1. Evaluation projects 2. Take-home final 3. Questions?

Slides:



Advertisements
Similar presentations
Postgraduate Course 7. Evidence-based management: Research designs.
Advertisements

Agenda Group Hypotheses Validity of Inferences from Research Inferences and Errors Types of Validity Threats to Validity.
Validity (cont.)/Control RMS – October 7. Validity Experimental validity – the soundness of the experimental design – Not the same as measurement validity.
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
Reliability for Teachers Kansas State Department of Education ASSESSMENT LITERACY PROJECT1 Reliability = Consistency.
Defining Characteristics
Measurement Reliability and Validity
GROUP-LEVEL DESIGNS Chapter 9.
Experimental Research Designs
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
©2005, Pearson Education/Prentice Hall CHAPTER 5 Experimental Strategies.
Assessing Program Impact Chapter 8. Impact assessments answer… Does a program really work? Does a program produce desired effects over and above what.
Research Design and Validity Threats
Validity, Sampling & Experimental Control Psych 231: Research Methods in Psychology.
Reliability and Validity in Experimental Research ♣
Common Designs and Quality Issues in Quantitative Research Research Methods and Statistics.
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Impact and outcome evaluation involve measuring the effects of an intervention, investigating the direction and degree of change Impact evaluation assesses.
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
PSYC512: Research Methods PSYC512: Research Methods Lecture 11 Brian P. Dyre University of Idaho.
Psych 231: Research Methods in Psychology
PPA 415 – Research Methods in Public Administration Lecture 1 – Research Design.
PPA 502 – Program Evaluation Lecture 3c – Strategies for Impact Assessment.
Rosnow, Beginning Behavioral Research, 5/e. Copyright 2005 by Prentice Hall Ch. 6: Reliability and Validity in Measurement and Research.
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Formulating the research design
Methodology: How Social Psychologists Do Research
Studying treatment of suicidal ideation & attempts: Designs, Statistical Analysis, and Methodological Considerations Jill M. Harkavy-Friedman, Ph.D.
EDRS6208 Lecture Three Instruments and Instrumentation Data Collection.
RESEARCH DESIGNS FOR QUANTITATIVE STUDIES. What is a research design?  A researcher’s overall plan for obtaining answers to the research questions or.
Research Design for Quantitative Studies
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
I/O Psychology Research Methods. What is Science? Science: Approach that involves the understanding, prediction, and control of some phenomenon of interest.
Measurement Neuman and Robson Ch. 6. What is it? The process of creating measurable concrete variables from abstract concepts Extends the senses (empirical)
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
Final Study Guide Research Design. Experimental Research.
CRJS 4466 PROGRAM & POLICY EVALUATION LECTURE #3 Evaluation projects Resume preparation Job hunting Questions? In-class test #1 – next week!
Systematic Review Module 7: Rating the Quality of Individual Studies Meera Viswanathan, PhD RTI-UNC EPC.
Techniques of research control: -Extraneous variables (confounding) are: The variables which could have an unwanted effect on the dependent variable under.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Enhancing Rigor in Quantitative Research.
URBDP 591 A Lecture 8: Experimental and Quasi-Experimental Design Objectives Basic Design Elements Experimental Designs Comparing Experimental Design Example.
Chapter Four Experimental & Quasi-experimental Designs.
URBDP 491 A Lecture 7: Research Approaches Objectives How to compare alternative approaches Experimental vs. non-experimental approaches Cross-sectional.
Validity Is the Test Appropriate, Useful, and Meaningful?
Understanding Research Design Can have confusing terms Research Methodology The entire process from question to analysis Research Design Clearly defined.
Research Process Parts of the research study Parts of the research study Aim: purpose of the study Aim: purpose of the study Target population: group whose.
Independent vs Dependent Variables PRESUMED CAUSE REFERRED TO AS INDEPENDENT VARIABLE (SMOKING). PRESUMED EFFECT IS DEPENDENT VARIABLE (LUNG CANCER). SEEK.
Validity RMS – May 28, Measurement Reliability The extent to which a measurement gives results that are consistent.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
CDIS 5400 Dr Brenda Louw 2010 Validity Issues in Research Design.
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
Experiment Basics: Variables Psych 231: Research Methods in Psychology.
WWC Standards for Regression Discontinuity Study Designs June 2010 Presentation to the IES Research Conference John Deke ● Jill Constantine.
SOCW 671 # 8 Single Subject/System Designs Intro to Sampling.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Copyright © 2016 Wolters Kluwer All Rights Reserved Chapter 7 Experimental Design I— Independent Variables.
Evaluating VR Systems. Scenario You determine that while looking around virtual worlds is natural and well supported in VR, moving about them is a difficult.
Introduction to Validity True Experiment – searching for causality What effect does the I.V. have on the D.V. Correlation Design – searching for an association.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Experiments.  Labs (update and questions)  STATA Introduction  Intro to Experiments and Experimental Design 2.
How do you know your product “works”? And what does it mean for a product to “work”?
Research Methods & Design Outline
CRITICALLY APPRAISING EVIDENCE Lisa Broughton, PhD, RN, CCRN.
Chapter 9 Scrutinizing Quantitative Research Design.
Research Designs, Threats to Validity and the Hierarchy of Evidence and Appraisal of Limitations (HEAL) Grading System.
Ron Sterr Kim Sims Heather Cruz aka “The Carpool”
Introduction to Experimental Design
Monitoring and Evaluating FGM/C abandonment programs
Rest of lecture 4 (Chapter 5: pg ) Statistical Inferences
Presentation transcript:

SOCI 4466 PROGRAM & POLICY EVALUATION LECTURE #8 1. Evaluation projects 2. Take-home final 3. Questions?

2. Strategies for Impact Assessment impact: the net effects of a program - the effects that can be uniquely attributed to the program intervention, controlling for the confounding effects of other variables/sources of change impact assessments can be carried out at virtually any stage of the program - piloting, program design, implementation, monitoring, outcome evaluation all impact assessments are comparative - comparing the net effect on those who got the program as compared to some other group - either themselves earlier, a control group, those in an alternative program, etc.

strongest approach to assessing impact is the use of the randomized experimental model Exp -R0X0 Con -R00

pre-requisites for assessing impacts: 1. clearly defined goals and objectives that can be operationalized 2. proper implementation of the intervention note here the considerable difficulties evaluators face in ensuring the above two criteria are met

the three criteria of causality: 1. correlation 2. temporal asymmetry 3. non-spuriousness note the difficulty in demonstrating that a program intervention is the “cause” of a specific outcome - the issue of causation versus correlation - bias in selection of targets - “history” - intervention (Hawthorne) effects - poor measurement

Campbell versus Cronbach: perfect versus good enough impact assessments - lack of experimental control - inability to randomize - “history” - time/money restraints - balancing the importance and impact of the program against practicality gross versus net outcomes Gross= Effects of + Effects of+ Design outcomeintervention other processes Effects (net effect) (extraneous factors)

extraneous confounding factors: - uncontrolled selection (selection bias) - both agency/self selection - “deselection” processes - the drop-out problem - endogenous change (naturally occurring change processes, like healing, learning) - secular drift - interfering effects (history) - maturational and developmental effects

design effects: - stochastic effects: chance fluctuations - the difference between real change and random change - the importance of sampling here, allowing the use of inferential statistics - statistical significance and statistical power: alpha: Type I error (false positive) beta: Type II error (false negative) - significance here of cell sizes and sample size - note differential concern with Type I or II error depending on program type

design effects (continued) - measurement reliability (qualitative/quantitative) - measurement validity (domain, internal consistency, predictive, concurrent) - experimenter/evaluator effects - missing data - sampling biases

choice of outcome measures - back to the measurement model, and reliability and validity - must be feasible to employ, responsive, exhaustive mutually exclusive and, ideally, quantitative - multiple measures best - direct versus indirect

isolating the effects of extraneous factors: - randomized controls - regression-discontinuity controls (pre-determined selection variables) - matched constructed controls - statistically-equated controls - reflexive controls (pre-post) - repeated measures reflexive controls (e.g. panel) - time series reflexive controls - generic controls (established norms, standards)

Full versus partial-coverage programs - if program is delivered to virtually all targets (full coverage), more difficult to find a design to assess impact (e.g. government-funded pension plans; OHIP) - partial coverage programs are not delivered to all targets, so there is opportunity to identify reasonable control/comparison groups

EXHIBIT 7 - F - HERE

judgmental impact assessments: - expert or “connoisseurial” assessments - administrator assessments - participants’ judgments the use of qualitative versus quantitative data

inference validity issues: - reproducibility of the evaluation design + results - generalizability - pooling evaluations - meta analysis