Nature gives us correlations… Evaluation Research (8521) Prof. Jesse Lecy Lecture 0 1.

Slides:



Advertisements
Similar presentations
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Advertisements

Designing an impact evaluation: Randomization, statistical power, and some more fun…
Surviving Your Freshman Year Frequently Asked Questions about Semesters, Averages, and Survival Skills Class of 2016! Sweet 16!
Class 18 – Thursday, Nov. 11 Omitted Variables Bias
Evidence-Based Education (EBE) Grover J. (Russ) Whitehurst Assistant Secretary Educational Research and Improvement United States Department of Education.
1 Title I Faculty Presentation Department of Federal and State Programs or PX
A Guide to Education Research in the Era of NCLB Brian Jacob University of Michigan December 5, 2007.
Lecture 28 Categorical variables: –Review of slides from lecture 27 (reprint of lecture 27 categorical variables slides with typos corrected) –Practice.
Testing Theories: Three Reasons Why Data Might not Match the Theory.
Adapting Designs Professor David Torgerson University of York Professor Carole Torgerson Durham University.
Math 161 Spring 2008 What Is a Confidence Interval?
QUASI-EXPERIMENTS w Compare subjects in different conditions on a DV w Lacks one or more criteria for an experiment (cause, comparison, control) w Interpreted.
Model and Variable Selections for Personalized Medicine Lu Tian (Northwestern University) Hajime Uno (Kitasato University) Tianxi Cai, Els Goetghebeur,
Multiple Regression Involves the use of more than one independent variable. Multivariate analysis involves more than one dependent variable - OMS 633 Adding.
How This Course Promotes “Learning Efficiencies” The following slide show provides an explanation of the assessment and grading process used in the Online.
Chapter 2 – Tools of Positive Analysis
TOOLS OF POSITIVE ANALYSIS
Stat 112: Lecture 9 Notes Homework 3: Due next Thursday
Quasi-Experimental Designs Whenever it is not possible to establish cause-and-effect relations because there is not complete control over the variables.
Statistics for Social and Behavioral Sciences Part IV: Causality Association and Causality Session 22 Prof. Amine Ouazad.
STAT 401 EXPERIMENTAL DESIGN AND ANALYSIS Assist.Prof.Dr. R. Serkan Albayrak Department of Business Administration Yaşar University.
Funding Opportunities at the Institute of Education Sciences Elizabeth R. Albro, Ph.D. Associate Commissioner Teaching and Learning Division National Center.
League of Women Voters® of Colorado Supports Amendment 66.
Common Questions What tests are students asked to take? What are students learning? How’s my school doing? Who makes decisions about Wyoming Education?
-- Preliminary, Do Not Quote Without Permission -- VALUE-ADDED MODELS AND THE MEASUREMENT OF TEACHER QUALITY Douglas HarrisTim R. Sass Dept. of Ed. LeadershipDept.
Back In Stride Again! Room 19 – Ms. Colette Smith, Teacher Phone Number –  Here are the following changes for the School Year  1.
T-Tests and Chi2 Does your sample data reflect the population from which it is drawn from?
Funding Opportunities at the Institute of Education Sciences Elizabeth R. Albro, Ph.D. Associate Commissioner Teaching and Learning Division National Center.
Understanding Statistics
Lecture 22 Dustin Lueker.  The sample mean of the difference scores is an estimator for the difference between the population means  We can now use.
Lecturer:Prof. Elizabeth A. Ritchie, ATMO TAs:Mr. Adrian Barnard Ms. Anita Annamalai NATS 101 Introduction to Weather and Climate Section 14: T/R 2:00.
Reclassification of English Learner Students in California
G Lecture 41 Panel Designs Panel Path Analyses Examples Lord’s Paradox Extensions.
Quantitative methods for politics: an introduction Massimo De Angelis.
© 2014, Florida Department of Education. All Rights Reserved Annual District Assessment Coordinator Meeting VAM Update.
Government and Public Policy
Statistics and Quantitative Analysis U4320 Segment 12: Extension of Multiple Regression Analysis Prof. Sharyn O’Halloran.
COMM 250 Agenda - Week 10 Housekeeping C2 – Returned to You Today RP1 – Due Today (IM Surveys) TP3a – Due Tomorrow Lecture RAT4 Review RP1 Experiments.
POSC 202A: Lecture 12/10 Announcements: “Lab” Tomorrow; Final ed out tomorrow or Friday. I will make it due Wed, 5pm. Aren’t I tender? Lecture: Substantive.
In the 2004, presidential election, of those Texans who voted for either Kerry or Bush, 62% voted for Bush and 38% for Kerry. Of the Massachusetts residents.
OSPI CHANGES AND PRIORITIES January OSPI agency priorities and organization chart.
OCM BOCES SLOs Workshop. Race To The Top: Standards Data Professional Practice Culture APPR.
CAMPBELL SCORES: ELIMINATING COMPETING HYPOTHESES.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Please turn off cell phones, pagers, etc. The lecture will begin shortly.
Chapter 1 Statistical Thinking What is statistics? Why do we study statistics.
Methods of Economic Investigation: Lent Term Radha Iyengar Office Hour: Monday Office: R425.
SW388R6 Data Analysis and Computers I Slide 1 Multiple Regression Key Points about Multiple Regression Sample Homework Problem Solving the Problem with.
Using Regression Discontinuity Analysis to Measure the Impacts of Reading First Howard S. Bloom
Fixed Effects Models Evaluation Research (8521) Prof. Jesse Lecy 1.
Omitted Variable Bias Evaluation Research (8521) Prof. Jesse Lecy 1.
Research Design ED 592A Fall Research Concepts 1. Quantitative vs. Qualitative & Mixed Methods 2. Sampling 3. Instrumentation 4. Validity and Reliability.
Public Finance and Public Policy Jonathan Gruber Third Edition Copyright © 2010 Worth Publishers 1 of 24 Empirical Tools of Public Finance 3.1 The Important.
SOCW 671: #5 Measurement Levels, Reliability, Validity, & Classic Measurement Theory.
Quasi Experimental and single case experimental designs
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
How to Develop Balanced Biliteracy in Language Minority Children Education Fall 2010 Romina G. Ladner.
ECON742: Empirical microeconomics Lecture 1 January 2014.
Seven Sins of Regression Evaluation Research (8521) Prof. Jesse Lecy Lecture 7 1.
Public Finance and Public Policy Jonathan Gruber Third Edition Copyright © 2010 Worth Publishers 1 of 24 Copyright © 2010 Worth Publishers.
Review Design of experiments, histograms, average and standard deviation, normal approximation, measurement error, and probability.
CHAPTER 13: THE NUTS AND BOLTS OF QUASI- EXPERIMENTS.
Statistics in Clinical Trials: Key Concepts
partitioned regression
Is it more likely that he comes from Texas or from Massachusetts?
POSC 202A: Lecture Lecture: Substantive Significance, Relationship between Variables 1.
Introduction to Econometrics
Empirical Tools of Public Finance
Positive analysis in public finance
An Introduction to Evaluating Federal Title Funding
Presentation transcript:

Nature gives us correlations… Evaluation Research (8521) Prof. Jesse Lecy Lecture 0 1

2 Policy / Program Black Box Input Outcome The Program Evaluation Mindset (something happens here)

3 Policy / Program Input Outcome The Program Evaluation Mindset (something happens here) The slopes tells us how much impact we expect a program to have when we spend one additional unit of input. The outcome is some function of the program and the amount of inputs into the process. It can sometimes be represented by this simple input-output equation.

Effect

5 Heart Rate Treatment (Caffeine) Control (No caffeine)

7 4:15-

How do we know when the interpretation is causal? Effect?

NATURE GIVES US CORRELATIONS 9

xy z Example #1

xy z Example #2

xy z Example #3

xy xy

Examples of Poor Causal Inference

1.Ice cream consumption causes polio 2.Investments in public buildings creates economic growth 3.Early retirement and health decline 4.Hormone replacement therapy and heart disease: In a widely-studied example, numerous epidemiological studies showed that women who were taking combined hormone replacement therapy (HRT) also had a lower-than-average incidence of coronary heart disease (CHD), leading doctors to propose that HRT was protective against CHD. But randomized controlled trials showed that HRT caused a small but statistically significant increase in risk of CHD. Re-analysis of the data from the epidemiological studies showed that women undertaking HRT were more likely to be from higher socio-economic groups (ABC1), with better than average diet and exercise regimes. The use of HRT and decreased incidence of coronary heart disease were coincident effects of a common cause (i.e. the benefits associated with a higher socioeconomic status), rather than cause and effect as had been supposed. (Wikipedia)

Examples of Complex Causal Inference

MODERN PROGRAM EVAL 17

To Experiment or Not Experiment 18

CASE STUDY – EDUCATION REFORM 19

Classroom Size and Performance State Laws Limiting Class Size Notwithstanding the ongoing debate over the pros and cons of reducing class sizes, a number of states have embraced the policy of class size reduction. States have approached class size reduction in a variety of ways. Some have started with pilot programs rather than state-wide mandates. Some states have specified optimum class sizes while other states have enacted mandatory maximums. Some states have limited class size reduction initiatives to certain grades or certain subjects. Here are three examples of the diversity of state law provisions respecting class size reduction. California – The state of California became a leader in promoting class size reduction in 1996, when it commenced a large-scale class size reduction program with the goal of reducing class size in all kindergarten through third grade classes from 30 to 20 students or less. The cost of the program was $1 billion annually.

Classroom Size and Performance Florida – Florida residents in 2002 voted to amend the Florida Constitution to set the maximum number of students in a classroom. The maximum number varies according to the grade level. For prekindergarten through third grade, fourth grade through eighth grade, and ninth grade through 12th grade, the constitutional maximums are 18, 22, and 25 students, respectively. Schools that are not already in compliance with the maximum levels are required to make progress in reducing class size so that the maximum is not exceeded by The Florida legislature enacted corresponding legislation, with additional rules and guidelines for schools to achieve the goals by Georgia -- Maximum class sizes depend on the grade level and the class subject. For kindergarten, the maximum class size is 18 or, if there is a full-time paraprofessional in the classroom, 20. Funding is available to reduce kindergarten class sizes to 15 students. For grades one through three, the maximum is 21 students; funding is available to reduce the class size to 17 students. For grades four through eight, 28 is the maximum for English, math, science, and social studies. For fine arts and foreign languages in grades K through eight, however, the maximum is 33 students. Maximums of 32 and 35 students are set for grades nine through 12, depending on the subject matter of the course. Local school boards that do not comply with the requirements are subject to lose funding for the entire class or program that is out of compliance.

Class Size Case Study - The Theory: Class Size Test Scores Class Size Test Scores SES ? Scenario 1 Scenario 2

Example: Classroom Size 23 ∆Y ∆X The regression coefficient represents a slope. In policy we think of the slope as an input-output formula. If I decrease class size (input) standardized tests scores increase (output). Note changes in slopes and standard errors when you add variables.

The Naïve Model:

With Teacher Skill as a Control:

Add SES to the Model:

Example: Classroom Size 27 Why are slopes and standard errors changing when we add “control” variables?

How do we interpret results causally? Class Size Test Scores SES Teacher Skill X

COURSE OUTLINE 29

The Origins of Modern Program Evaluation 30 The “Great Society” introduced unprecedented levels of spending on social services – marks the dawn of the modern welfare state. Econometrics also comes of age, creating tools the provide opportunity for rigorous analysis of social programs.

31

32

33 We need effective programs, not expensive programs

Modern Program Evaluation 34 Course Objectives: 1.Understanding why regressions are biased Seven Deadly Sins of Regression: 1.Multicollinearity 2.Omitted variable bias 3.Measurement error 4.Selection / attrition 5.Misspecification 6.Population heterogeneity 7.Simultaneity

Modern Program Evaluation 35 Course Objectives: 2.Understand tools of program evaluation – Fixed effect models – Instrumental variables – Matching – Regression discontinuity – Time series – Survival analysis

Modern Program Evaluation 36 Course Objectives: 3.How to talk to economists (and other bullies)

Modern Program Evaluation 37 Course Objectives: 4.Correctly apply and critique evaluation designs Experiments Pretest-posttest control group Posttest only control group Quasi-Experiments Pretest-posttest comparison group Posttest only comparison group Interrupted time series with comparison group Reflexive Design Pretest-posttest design Simple time series

Course Structure 38 First half: Understanding bias – No text, course notes online – Weekly homework Second half: Evaluation design – Text is required – Campbell Scores policy-research.net/programevaluation

Evaluating Internal Validity: The Campbell Scores 39 A Competing Hypothesis Framework Omitted Variable Bias Selection Nonrandom Attrition Trends in the Data Maturation Secular Trends Study Calibration Testing Regression to the Mean Seasonality Study Time-Frame Contamination Factors Intervening Events Measurement Error

Homework Policy 40 Homework problems each week 1 st half of semester – Graded pass/fail – Submit via D2L please – Work in groups is strongly encouraged Campbell Scores are due each class for the second half of the semester Midterm Exam (30%) – Confidence intervals, Standard error of regression, Bias Final Exam (20%) – Covers evaluation design and Internal validity No late homework accepted! 50% of final grade. See syllabus for policy on turning in by .

41 1st Qu.75.5 Median89 Mean rd Qu.96 1st Qu.82 Median86 Mean rd Qu.97.5

42