Non Experimental Design in Education Ummul Ruthbah.

Slides:



Advertisements
Similar presentations
PhD Research Seminar Series: Valid Research Designs
Advertisements

BOARD POLICY GRADES 3, 6, 8. Passing grade on final report card in READING AND MATH Summer school not required, promote to next grade, participate in.
REGRESSION, IV, MATCHING Treatment effect Boualem RABTA Center for World Food Studies (SOW-VU) Vrije Universiteit - Amsterdam.
Regression Discontinuity. Basic Idea Sometimes whether something happens to you or not depends on your ‘score’ on a particular variable e.g –You get a.
Mywish K. Maredia Michigan State University
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
Baela Raza Jamil Zara Khan Zaheer Abbas PRESENTED AT: Comparative and International Education Society (CIES) – 10 th to 16 th March 2014 Effects of Parental.
A Guide to Education Research in the Era of NCLB Brian Jacob University of Michigan December 5, 2007.
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
D O E XTRA H OURS OF T UITION P AY O FF ? Atonu Rabbani (Department of Economics, University of Dhaka) Ummul Hasanath Ruthbah (Department of Economics,
Designing Influential Evaluations Session 5 Quality of Evidence Uganda Evaluation Week - Pre-Conference Workshop 19 th and 20 th May 2014.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
Nasih Jaber Ali Scientific and disciplined inquiry is an orderly process, involving: problem Recognition and identification of a topic to.
Methods and Approaches to investigate the UK Education System Sandra McNally, University of Surrey and Centre for Economic Performance, London School of.
GS/PPAL Section N Research Methods and Information Systems A QUANTITATIVE RESEARCH PROJECT - (1)DATA COLLECTION (2)DATA DESCRIPTION (3)DATA ANALYSIS.
Experiments and Observational Studies.  A study at a high school in California compared academic performance of music students with that of non-music.
The Impact of Court Decentralization on Domestic Violence Against Women Raúl Andrade Jimena Montenegro March 2009.
Selecting a Research Design. Research Design Refers to the outline, plan, or strategy specifying the procedure to be used in answering research questions.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
Experiments and Observational Studies. Observational Studies In an observational study, researchers don’t assign choices; they simply observe them. look.
Chapter 13 Observational Studies & Experimental Design.
TRANSLATING RESEARCH INTO ACTION What is Randomized Evaluation? Why Randomize? J-PAL South Asia, April 29, 2011.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Case Studies Harry Anthony Patrinos World Bank November 2009.
Overview of Evaluation Designs. Learning objectives By the end of this presentation, you will be able to: Explain evaluation design Describe the differences.
1 The Need for Control: Learning what ESF achieves Robert Walker.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.
Chapter 3.1.  Observational Study: involves passive data collection (observe, record or measure but don’t interfere)  Experiment: ~Involves active data.
Public Policy Analysis ECON 3386 Anant Nyshadham.
MSRP Year 1 (Preliminary) Impact Research for Better Schools RMC Corporation.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
Measuring Impact 1 Non-experimental methods 2 Experiments
Statistics for the Social Sciences Psychology 340 Spring 2010 Introductions & Review of some basic research methods.
Framework of Preferred Evaluation Methodologies for TAACCCT Impact/Outcomes Analysis Random Assignment (Experimental Design) preferred – High proportion.
Randomized Assignment Difference-in-Differences
Building an evidence-base from randomised control trials Presentation of the findings of the impact evaluation of the Reading Catch-Up Programme 18 August.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
4.1 Statistics Notes Should We Experiment or Should We Merely Observe?
Measuring causal impact 2.1. What is impact? The impact of a program is the difference in outcomes caused by the program It is the difference between.
Do European Social Fund labour market interventions work? Counterfactual evidence from the Czech Republic. Vladimir Kváča, Czech Ministry of Labour and.
Revamping the Teaching Profession by Attracting Non-Teachers to It: Evidence from Enseña Chile Mariana Alfonso Education Division, Inter-American Development.
School of Public Administration & Policy Dr. Kaifeng Yang 研究设计 : 实验研究的基本问题.
Impact Evaluation Methods Regression Discontinuity Design and Difference in Differences Slides by Paul J. Gertler & Sebastian Martinez.
Issues in Evaluating Educational Research
EXPERIMENTAL RESEARCH
Bo Yan Blue Valley School District
Quasi Experimental Methods I
Impact evaluations at IFAD-IOE
Quasi-Experimental Methods
March 2017 Susan Edwards, RTI International
Quasi-Experimental Methods
Impact evaluation: The quantitative methods with applications
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Impact Evaluation Methods
Matching Methods & Propensity Scores
Implementation Challenges
Brahm Fleisch Research supported by the Zenex Foundation October 2017
1/18/2019 1:17:10 AM1/18/2019 1:17:10 AM Discussion of “Strategies for Studying Educational Effectiveness” Mark Dynarski Society for Research on Educational.
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
Sampling for Impact Evaluation -theory and application-
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
Reminder for next week CUELT Conference.
Presentation transcript:

Non Experimental Design in Education Ummul Ruthbah

A way around RCT Experiments are difficult to run. Is there a way around? There are several techniques Difference in Difference Regression discontinuity Propensity score matching

Methodology 1: Difference in Difference (DID) Suppose we want to evaluate the impact of supplementary tutoring on primary school students (Ruthbah, Rabbani, Hossain & Sarwar 2012). One way to do it is to assign students randomly to the program. What if the program is already in place and it did not follow the RCT protocol? How do we create a control group now?

Evaluating the Education Support Program of the CDIP The Center for Development Innovation and Practices provide 2 hours of supplementary tuition to nursery, grade 1 and 2 students in many districts of Bangladesh. Operating learning centres adjacent to primary schools since Supplementary tuition (about 10 hours per week) to primary school students in nursery, grade 1 and grade 2. 1,750 learning centres adjacent to the primary schools. We want to estimate the effect of the program on the participants test score and dropout rate.

Treatment and Control We could compare students who participated in the program with those who did not. But it could be that only the weak students participated into the program and therefore the treatment and control students are not similar

Methodology Treatment group Control group Pre-treatment observation 2007 (Grade I) Pre-treatment observation Students attending primary schools Post- treatment observations 2008 (Grade (II) CDIP intervention Students attending primary schools and CDIP LCs Students attending primary schools only 2009 ( Grade (III) Students attending primary schools 2010 (Grade IV)-do (Grade V)-do- 6

Methodology Test Scores in FinalDifference in test scores between 2007 and (grade 1)2008 (grade 2) Students who participated in ESP (Treatment) X T 2007 X T 2008 X T X T 2007 (a) Students who did not participate in ESP (Control) X C 2007 X C 2008 X C X C 2007 (b) Difference between treatment and control groups X T X C 2007 X T X C 2008 (a) – (b) = DID estimate (c) 7

Sampling Strategy 304 learning centres in 2008 in 33 unions of 8 upazilas in Bangladesh. Only 262 centres had students from grade 2. A sample of 1900 students (950 in each of treatment and control groups) from 159 learning centres and the associated primary schools. 8

Sampling Strategy Multistage sampling  Select the learning centres  Select students who were in grade 2 in 2008 and participated in the program  Select control students (6 on average) from the schools who were in grade 2 in 2008 but did not participate in the program. 9

The Surveys Three sets of questionnaire on: Performance of the treatment and control students in the final exams. Background of students School information 10

The Field Experience Could not get the complete list of students who were in grade two in 2008 and attended the ESP. We collected data on 2147 students, of whom 1078 students attended 144 different CDIP learning centers in The schools could provide the marks/test scores for 2007 for only 1215 students. 11

Figure 1: total marks obtained 12

Figure 2: difference in marks between pre-post treatment years 13

Results: class performance (did estimates) TotalBengaliEnglishMath (1)(2)(3)(4) Grade (.06) 0.12** (.06) (0.06) 0.05 (0.07) Grade (.06).01 (.05).040 (.07) (.07) Grade 4.06 (.07).06 (.06).09 (.07) -.02 (.08) Grade (.06) 0.03 (.06) 0.01 (.06) (.06) 14

Methodology 2: Regression Discontinuity Design Jacob and Lefgren (2002) examine the effect of summer school and grade retention on students’ achievement. The ideal situation would be to randomly assign students with poor grades to summer school or retain in the same grade. And compare them with those who did not go to summer school or repeated the grade (the control group). But it is not possible to ethical and or other reasons. How do we find the counterfactual (the control group)?

Measuring the Impact of Remedial Education Chicago Public Schools introduced an accountability policy in Schools should decide who goes to summer school and who should repeat the grade depending on the student’s performance in a standardized test on Math and Reading.

Figure 3: The Design

Treatment – Control Groups Students just below the cut-off in June test constitute the treatment group and those at the cut-off belong to the control group for assessing the impact of summer school on future achievement. Students just below the cut-off in August test constitute the treatment group and those at the cut-off belong to the control group for assessing the impact of grade retention on future achievement.

Figure 4: the Relationship between June Reading Scores and the Probability of Attending Summer School or being Retained

Figure 5: Relationship between June reading and next year reading and math performance for third grade students

Figure 6: Relationship between June reading and next year reading and math performance for sixth grade students

Figure 7: Relationship between August reading and next year reading and math performance for third grade students

Figure 8: Relationship between August reading and next year reading and math performance for sixth grade students

The DID Estimate can use the following DID estimator to find the impact of summer school: Where, = mean achievement c = student at the cut-off c-1 = student just below the cut-off T = the probability of attending the summer school t = time period

RDD: Main Idea

Methodology 3: the Propensity Score Matching If the program affects the treatment group in a different way then it would have affected the control group. The DID estimates are of no use. It happens is selection into the program depends on factors that also affect the outcome of interest. Example: the decision to attend the leaning centers may depend on the parents years of education and parents education is believed to have influence on students test scores. How to create a treatment – control group is this case?

Matching For the same level of parental education there are some students who attend the LCs and some who do not. For each level of parental education those who attend the LCs belong to the treatment group and those who do not belong to the control group. We take the average difference in test scores of treatment and control students for each level of parental education. The average of the differenced test scores over all parental education level is out treatment effect.

More than One Determinants What if there are more than one variable (factor) that affect both selection and outcome variable? For example: parental education and income. Use propensity scores. Propensity score = probability of getting treatment = f(parental education, income). Students with same parental education and income will have the same probability of getting treatment (propensity score)

Treatment an Control Groups For the same propensity score students who went to the LCs belong to the treatment group and those who did not belong to the control group.