Comments on Schleicher and Stanat AERA Session 72.011 Impact of the Programme for International Student Assessment (PISA) on Educational Quality Daniel.

Slides:



Advertisements
Similar presentations
The Reasons behind the Results Markku Linna
Advertisements

Cross Cultural Research
School autonomy and student achievement. An international study with a focus on Italy Angelo Paletta Maria Magdalena Isac Daniele Vidoni.
Why this Research? 1.High School graduates are facing increased need for high degree of literacy, including the capacity to comprehend texts, but comprehension.
Unit 3 Siobhan Carey Department for International Development Making cross-national comparisons using micro data.
C R E S S T / Harvard Daniel Koretz Harvard Graduate School of Education National Center for Research on Evaluation, Standards, and Student Testing “Believe.
Standards for Qualitative Research in Education
Communicating through Data Displays October 10, 2006 © 2006 Public Consulting Group, Inc.
Threats to the Validity of Measures of Achievement Gains Laura Hamilton and Daniel McCaffrey, RAND Corporation Daniel Koretz, Harvard University November.
Experimental Design, Statistical Analysis CSCI 4800/6800 University of Georgia Spring 2007 Eileen Kraemer.
Multiple Regression Models Advantages of multiple regression Important preliminary analyses Parts of a multiple regression model & interpretation Differences.
Epistemology and Methods Data Selection, Operationalization and Measurement May
Objectives and Indicators for MCH Programs MCH in Developing Countries January 25, 2011.
Chapter 2 Finding Ideas to Research. Generating Topics Translate ideas into valid and reliable ways of measuring them Collect evidence Unique Topics Innovative.
Institut zur Qualitätsentwicklung im Bildungswesen Dirk Richter Eurasian Educational Dialogue, Jaroslawl Monitoring student achievement in Germany:
Assessment in the early years © McLachlan, Edwards, Margrain & McLean 2013.
OECD/INFE tools for cross- country surveys of financial literacy
What We Know About Effective Professional Development: Implications for State MSPs Part 2 Iris R. Weiss June 11, 2008.
Inferences about School Quality using opportunity to learn data: The effect of ignoring classrooms. Felipe Martinez CRESST/UCLA CCSSO Large Scale Assessment.
The Systematic Vocabulary Instruction Project Based on Marzano’s Classroom Instruction That Works.
Validation of the Assessment and Comparability to the PISA Framework Hao Ren and Joanna Tomkowicz McGraw-Hill Education CTB.
Results H1 was supported. Paired samples t-tests revealed statistically significant (p
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
Chapter 1: The What and the Why of Statistics
Validity in an Era of Accountability Daniel Koretz CRESST/ Harvard Graduate School of Education.
Conditions de Travail et Formation – Comment - Gerhard Bosch, IAQ.
Professional Development Activity Log: A New Approach to Design, Measurement, Data Collection, and Analysis AERA Annual Meeting San Diego April 13, 2004.
Slide 1 Estimating Performance Below the National Level Applying Simulation Methods to TIMSS Fourth Annual IES Research Conference Dan Sherman, Ph.D. American.
How are inequality of opportunity and mean student performance related? A quantile regression approach using PISA data Zoltán Hermann – Dániel Horn Institute.
Key Findings of the OECD Policy Review of Migrant Education and Implications for the Irish Intercultural Strategy Miho Taguma Project Manager of Policy.
Assessing assessment: the role of student effort in comparative studies Ray Adams Jayne Butler.
Are Numbers Enough? Classroom Information and Feedback.
The What and the Why of Statistics The Research Process Asking a Research Question The Role of Theory Formulating the Hypotheses –Independent & Dependent.
Approaches to educational research Research topics Research problems Research purpose Research questions.
Chapter 1: The What and the Why of Statistics  The Research Process  Asking a Research Question  The Role of Theory  Formulating the Hypotheses  Independent.
LEVEL 3 I can identify differences and similarities or changes in different scientific ideas. I can suggest solutions to problems and build models to.
ראמ " ה The National Authority for Measurement and Evaluation in Education Correlation between Pre-primary Education and Achievements in PISA 2009 Joel.
Quantitative SOTL Research Methods Krista Trinder, College of Medicine Brad Wuetherick, GMCTE October 28, 2010.
Cautions About Inferences From International Assessments Kadriye Ercikan University of British Columbia This discussion is based on a paper co-authored.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
1 Copyright © 2011 by Saunders, an imprint of Elsevier Inc. Chapter 8 Clarifying Quantitative Research Designs.
Enhancing the Technical Quality of the North Carolina Testing Program: An Overview of Current Research Studies Nadine McBride, NCDPI Melinda Taylor, NCDPI.
The Choice Between Fixed and Random Effects Models: Some Considerations For Educational Research Clarke, Crawford, Steele and Vignoles and funding from.
“Value added” measures of teacher quality: use and policy validity Sean P. Corcoran New York University NYU Abu Dhabi Conference January 22, 2009.
A discussion of Comparing register and survey wealth data ( F. Johansson and A. Klevmarken) & The Impact of Methodological Decisions around Imputation.
Non-Experimental Design Where are the beakers??. What kind of research is considered the “gold standard” by the Institute of Education Sciences? A.Descriptive.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
D/RS 1013 Data Screening/Cleaning/ Preparation for Analyses.
Paper III Qualitative research methodology.  Qualitative research is designed to reveal a specific target audience’s range of behavior and the perceptions.
Two sides of optimism: The positive and negative consequences of dispositional optimism and optimistic attributional style Evgeny Osin (Higher School of.
Impact of Perceived Discrimination on Use of Preventive Health Services Amal Trivedi, M.D., M.P.H. John Z. Ayanian, M.D., M.P.P. Harvard Medical School/Brigham.
UAE National Agenda – A World Class Education
Unit 11: Use observation, assessment and planning
Report on the NCSEAM Part C Family Survey Batya Elbaum, Ph.D. National Center for Special Education Accountability Monitoring February 2005.
How Psychologists Do Research Chapter 2. How Psychologists Do Research What makes psychological research scientific? Research Methods Descriptive studies.
Developing Effective Statewide Kindergarten Assessments with the Work Sampling System Developing Effective Statewide Kindergarten Assessments with the.
PISA – an option to learn from other countries‘ educational systems On PISA and German educational reforms within the past decade Seminar in Tallinn, 19.
The influence of forgetting rate on complex span and academic performance Debbora Hall, Chris Jarrold, John Towse and Amy Zarandi.
Understanding Standards: Advanced Higher Statistics
Question 1: What is the baseline of high power?
Seminar on the importance of Education Research and Innovation
Esteban Villalobos, Diego Portales University
Instructional Practices in the Early Grades that Foster Language & Comprehension Development Timothy Shanahan University of Illinois at Chicago
Looking for Relationships…
Reporting the evidence:
Describe two features of…
Resource 1. Evaluation Planning Template
Performance Measurement Workgroup
Organisation for Economic Co-Operation and Development Indicators on the Quality of Educational Performance Quality of Education Teachers’ Professional.
Presentation transcript:

Comments on Schleicher and Stanat AERA Session Impact of the Programme for International Student Assessment (PISA) on Educational Quality Daniel Koretz Harvard Graduate School of Education Annual Meeting of the American Educational Research Association Chicago April 13, 2007

2 The impact of PISA  Provides a rich database for exploring patterns of performance and its correlates  Has sparked a wide range of careful and informative analyses  Has contributed to an international wave of education reform

3 Three cautions  PISA is one test  Beware of unwarranted causal inference  Omitted variables: the special case of ILSA  Monitor impact; don’t trust scores alone

4 Inconsistencies across assessments Source: Excerpted from Grønmo & Olsen (2006)

5 Decomposition of variance, TIMSS 1995 Middle-school mathematics Source: Koretz, D., McCaffrey, D., and Sullivan, T. (2001)

6 Explaining differences between the assessments  Differences in results can stem from:  Intended differences in content and format  Unintended differences in attributes measured  Sampling: age of students, age vs. grade sampling, school vs. classroom sampling; differences in exclusions and participation rates  Empirical reasons for specific disparities are not always clear—no empirical linking study  Unless there is a strong reason to prefer one test, place greatest weight on results that replicate

7 Avoiding unwarranted causal inferences  Inadequacy of intra-national survey databases for causal inference widely acknowledged  More severe for explanation of international differences; e.g.:  Cultural and other unmeasured social differences among countries  Variation in the meaning of proxy measures across countries

8 Examples of appropriate language on causality: Stanat and Christensen (2003) The report attempts to identify factors that might contribute to between country differences… Although it is not possible to estimate the effects of these factors on immigrant students’ school success using the PISA data, the analyses presented in the report provide a description…

9 Responding to omitted variables in PISA  Avoid strong causal inference  The residual is just the residual  View analyses as generating tentative conclusions and hypotheses for further testing  Unfortunately, explanations of international differences are riskier than international replications of within- country relationships

10 Omitted variables and the problem of immigration  Why are PISA background variables adequate controls for ethnic German immigrants but not for Turkish?  Thought experiments from the U.S. experience:  1990s immigrants from HK and El Salvador  Italian and Jewish immigrants,  One possible explanation is culture, which is very hard to measure in a comparative survey  An illuminating example: Caplan, Nathan, et al. (1991). Children of the Boat People: A Study of Educational Success. Ann Arbor, MI: The University of Michigan Press.

11 Monitoring impact  PISA is sparking a remarkable range of reform efforts throughout OECD  More focus on monitoring of outcomes and accountability  Widespread interest in additional ‘PISA-like’ measures  How will we ascertain their effects  U.S. experience: impact is likely to be mixed, and scores may become seriously inflated

12 Advice from this side of the pond  Put monitoring of PISA-related reforms in place at the outset  Monitor effects on educational practice  Validate gains: do not trust increases in scores on tests used for accountability  Put auditing measures in place  Guard against corruption of PISA itself by the proliferation of PISA-like assessments with consequences