Gary W. Phillips Vice President and Chief Scientist American Institutes for Research August 3, 2015.

Slides:



Advertisements
Similar presentations
Pennsylvania’s Continuous Improvement Process. Understanding AYP How much do you know about AYP?
Advertisements

Southern Regional Education Board ALPHABET SOUP Using HSTW to Meet the Goals of NCLB Scott Warren Director of Contracted Services High Schools That Work.
Alaska’s English Language Arts and Mathematics Standards Need for Change 1.
Chapter Fifteen Understanding and Using Standardized Tests.
BIE Accountability: Adequate Yearly Progress
What Parents Need to Know  TABS (Texas Assessment of Basic Skills)  TEAMS (Texas Educational Assessment of Minimum Skills)  TAAS (Texas Assessment.
5/16/ © 2011 California County Superintendents Educational Services Association.
Highlights from TIMSS 2011 Jack Buckley National Center for Education Statistics Washington, DC December 11, 2012.
Highlights from PIRLS and TIMSS 2011 Jack Buckley National Center for Education Statistics Washington, DC December 11, 2012.
Robert L. Linn CRESST, University of Colorado at Boulder Paper presented at a symposium sponsored by the National Association of Test Directors entitled.
Assessment & Accountability TEP 128A March 7, 2006.
The Assessment System in the Ministry of Education of the United Arab Emirates Monday, April 17, 2017.
Catherine Cross Maple, Ph.D. Deputy Secretary Learning and Accountability
Richfield Elementary School Improvement Plan Amber Lawrence, Chairperson Brian K. Barrett, Principal.
Common Questions What tests are students asked to take? What are students learning? How’s my school doing? Who makes decisions about Wyoming Education?
Challenges in International Large-Scale Assessments Higher School of Economics, Moscow, Russia, May 16, 2013.
Challenges in Developing a University Admissions Test & a National Assessment A Presentation at the Conference On University & Test Development in Central.
Development of educational assessment in Ukraine Liashenko O. I., doctor of educational science, professor, chief scientific secretary of Academy of Educational.
English Language Development Assessment (ELDA) Background to ELDA for Test Coordinator and Administrator Training Mike Fast, AIR CCSSO/LEP-SCASS March.
This Report is available at: Condensed essay available at:
What ACCESS, the New Virginia Test for LEP Students, Means for School Districts LEP Caucus Presentation July 2008.
Understanding and Using Standardized Tests
High School Mathematics: Where Are We Headed? W. Gary Martin Auburn University.
Mathematics The Bigger Picture. Mathematics is not arithmetic… Arithmetic is about –math facts –computation –algorithms –word problems Mathematics is.
NAEP and International Assessments August 28, 2013 DAC Annual Meeting.
1 Paul Tuss, Ph.D., Program Manager Sacramento Co. Office of Education August 17, 2009 California’s Integrated Accountability System.
Teaching and Learning Mathematics How did we get to where we are?
Highlights from PIRLS and TIMSS 2011 Jack Buckley National Center for Education Statistics Washington, DC December 11, 2012.
Benchmarking with National and International Assessments Larry V. Hedges Northwestern University This paper is intended to promote the exchange of ideas.
Assessment in Early Childhood Legislation. Legislation for Young Children The need for measurement strategies and tests to evaluate federal programs led.
International Assessments September 5, 2012 DAC Annual Meeting
Dubai’s 15 year old students Results from PISA 2009 Fatma Al Janahi Head of International Assessments Knowledge & Human Development Authority GCES Symposium,
Teaching and Learning Mathematics How did we get to where we are?
Understanding International Assessments Tom Loveless UNLV Las Vegas, Nevada January 17, 2012.
Pontotoc City School District. Pontotoc City School District believes LEARNING is a priority, a need, and a desire. To be successful, we must nurture.
Understanding the TerraNova Test Testing Dates: May Kindergarten to Grade 2.
Metairie Academy for Advanced Studies WELCOME LEAP Information February 18, 2014.
Student Data at Claremont Immersion What Does That Mean?? NCLB - No Child Left Behind –Federal regulations guiding test requirements and passing.
1 Connecticut Mastery Test (CMT) Steve Martin, CMT Program Manager Bureau of Research, Evaluation, and Student Assessment Connecticut State Department.
1 Comparing International Assessments to NAEP May 30, 2008 Eugene Owen Senior Advisor, NCES International Activities Program This paper is intended to.
NJ ASSESSMENTS CYCLE II REPORT GRADES 3-8 and 11 October 30, 2008 Haddonfield Public Schools.
No Child Left Behind Education Week
State Assessment Parent Presentation. 05/14/03 Why is this in place? Federal legislation passed in 2001  No Child Left Behind Reauthorization of the.
RECYCLE Jenkins Independent District Data STATE NCLB DATA 103 of 175 school districts (58.9%) met 100% of their No Child Left Behind (NCLB) AYP goals.
Assessment: Results & Implications for Instruction Parent meeting – October 13, 2011.
ADEQUATE YEARLY PROGRESS. Adequate Yearly Progress Adequate Yearly Progress (AYP), – Is part of the federal No Child Left Behind Act (NCLB) – makes schools.
I NTERNATIONAL B ENCHMARKING S TUDY — C ONTENT A LIGNMENT Mary J. Pitoniak, Nancy Glazer, Luis Saldivia Educational Testing Service June 22, 2015 National.
Obtaining International Benchmarks for States Through Statistical Linking: Presentation at the Institute of Education Sciences (IES) National Center for.
University of Colorado at Boulder National Center for Research on Evaluation, Standards, and Student Testing Challenges for States and Schools in the No.
GEORGIA’S CRITERION-REFERENCED COMPETENCY TESTS (CRCT) Questions and Answers for Parents of Georgia Students February 11, 2009 Presented by: MCES.
Gary W. Phillips Vice President & Institute Fellow American Institutes for Research Next Generation Achievement Standard Setting Symposium CCSSO NCSA New.
KHS PARCC/SCIENCE RESULTS Using the results to improve achievement Families can use the results to engage their child in conversations about.
6/14/2016 “A Horse of a Different Color” No Child Left Behind and Accountability The State Testing Program Louisiana.
Granby Public Schools Annual Continuous Progress Review Presented by Diane Dugas Director of Curriculum September CMT Review.
1 Main achievement outcomes continued.... Performance on mathematics and reading (minor domains) in PISA 2006, including performance by gender Performance.
Adequate Yearly Progress (AYP). What is Adequate Yearly Progress (AYP)? As a condition of receiving federal funds under No Child Left Behind (NCLB), all.
A Close Look at Don’t Fail Idaho’s Student Achievement Message June 25, 2013 Bert Stoneberg, Ph.D. K-12 Research Idaho
Standards-based Grading and Reporting Information for Parents.
Common Core State Standards: Myths vs. Facts
INTERNATIONAL ASSESSMENTS IN QATAR TIMSS
PISA 2015 Excellence and Equity in Education Peter Adams
Metropolitan Nashville Public Schools
Student Achievement Data Displays Mathematics & Reading Grade 3
PISA • PIRLS • TIMSS Program for International Student Assessment
NECAP PRESENTATION.
Understanding and Using Standardized Tests
NAEP and International Assessments
2009 California Standards Test (CST) Results
EDN Fall 2002.
Presentation transcript:

Gary W. Phillips Vice President and Chief Scientist American Institutes for Research August 3, 2015

 States are busy developing new content standards and new criterion-referenced tests that measure success on those content standards.  This activity is related to the federal No Child Left Behind (NCLB) legislation that require states to report annually on whether they are making adequate yearly progress (AYP) toward meeting state standards.  However, when standard setting panelists set performance standards, they generally have little knowledge of how their state performance standards compare with other states.  They also typically have no knowledge of how their state standards compare to national standards such as those used on the National Assessment of Educational Progress (NAEP).

 They also have no understanding of how their state performance standards compare with international standards, such as those used in TIMSS, PIRLS, and PISA.  Consequently, the states do not know if their performance standards are internationally competitive and are often surprised to find that they are setting low standards (compared to the highest achieving countries in the world) when they thought they were setting a high standard.  My presentation will discuss why states should use international benchmarking as an integral part of their standard setting process.

 Before I talk about the importance of international benchmarks let me give you a brief tutorial on the three most commonly discussed international assessments that can be used for international benchmarking.

 TIMSS-Trends in International Mathematics and Science  PIRLS-Progress in International Reading  PISA-Programme for International Student Assessment

 TIMSS 4-year cycle, Grades 4 & 8, Math & Science ◦ 1995 Math, Science ◦ 1999 Math, Science ◦ 2003 Math, Science ◦ 2007 Math, Science ◦ 2011 Math, Science  PIRLS 5-year cycle, Grade 4, Reading ◦ 2001 Reading ◦ 2006 Reading ◦ 2011 Reading  PISA 3-year cycle, Age 15, Reading, Math, Science ◦ 2000 Reading(Math, Science) ◦ 2003 Math(Reading, Science) ◦ 2006 Science (Reading, Math) ◦ 2009 Reading (Math, Science) ◦ 2012 Math (Reading, Science) ◦ 2015 Science (Reading, Math)

 Mathematics ◦ Number ◦ Algebra ◦ Geometry ◦ Measurement ◦ Data  Science ◦ Life Science ◦ Chemistry ◦ Physics ◦ Earth Science ◦ Environmental Science

 Reading Comprehension ◦ Literary Experience ◦ Acquiring and Using Information

 Reading ◦ Interpreting Texts ◦ Reflection and Evaluation ◦ Retrieving Information  Mathematics ◦ Quantity ◦ Space and Shape ◦ Change and Relationships ◦ Uncertainty  Science ◦ Identifying Scientific Issues ◦ Explaining Phenomena Scientifically ◦ Using Scientific Evidence

 TIMSS & PIRLS ◦ 2-stage sampling (schools, classrooms) ◦ 150 schools (minimum) ◦ 1 (or 2) classroom per school ◦ 4,500 students (minimum) ◦ Participation rates (85% schools, 80% students) ◦ Population coverage 95%  PISA ◦ 2-stage sampling (schools, students) ◦ 150 schools (minimum) ◦ 30 students per school ◦ 4,500 students (minimum) ◦ Participation rates (85% schools, 80% students) ◦ Population coverage 95%

 TIMSS ◦ 28 blocks (14 Math, 14 Science) ◦ 300 items (G4), 400 items (G8) ◦ 6 blocks per booklet ◦ Each student takes both subjects ◦ 70 minutes (G4), 90 minutes (G8) per student testing time  PIRLS ◦ 10 passages (5 literary, 5 informational), 120 items ◦ 2 passages per student ◦ 80 minutes per student testing time  PISA ◦ 13 clusters (7 Science, 2 Reading, 4 Math) ◦ 108 (Science), 31 items (Reading), 48 items (Math) ◦ 4 clusters per booklet ◦ Each student takes all three subjects ◦ 120 minutes per student testing time

 Each state develops its own academic content standards  Each state develops its own performance standards  Each state develops its own test  Each state reports adequate yearly progress to the federal government based on ◦ different content standards, ◦ different performance standards, ◦ different tests. 14 Gary W. Phillips American Institutes for Research

 Data from the 50 states are not comparable. ◦ No inference about national progress is possible. ◦ We cannot tell if progress in one state is better than progress in another state.  From a scientific point of view the system lacks transparency. ◦ Transparency in measurement is the first and most fundamental requirement for progress in science. ◦ Common metrics are needed for transparency. ◦ The ability to derive comparable measures from different measuring devices is the very definition of transparency. 15 Gary W. Phillips American Institutes for Research

 We compare each state achievement standard to a high common international standard.  This gives us a comparable measure against which state standards from different tests can be compared. 16 Gary W. Phillips American Institutes for Research

17 Gary W. Phillips American Institutes for Research

18 Gary W. Phillips American Institutes for Research

19 Gary W. Phillips American Institutes for Research

20 Gary W. Phillips American Institutes for Research

21 Gary W. Phillips American Institutes for Research

22 Gary W. Phillips American Institutes for Research

23 Gary W. Phillips American Institutes for Research

24 Gary W. Phillips American Institutes for Research

25 Gary W. Phillips American Institutes for Research

26 Gary W. Phillips American Institutes for Research

 In all three grades and subjects ◦ Correlation between the performance standard and percent proficient was about ◦ This means 60% of the variance in results reported to NCLB is due to how high or low the states set their performance standards. ◦ The Expectation Gap (the difference between the highest and lowest performance standard) was about two standard deviations (twice the size of the Achievement Gap). Gary W. Phillips American Institutes for Research 27

 The fundamental problem with the existing state testing paradigm is lack of transparency. ◦ Low expectations in most states. ◦ Misleads the public because high levels of proficiency are often obtained by using low standards. ◦ Students in states with low standards do not have the opportunity to learn challenging content.  International benchmarking provides a mechanism for calibrating the difficulty and an external referent for gauging the global competitiveness of each state performance standard. 28 Gary W. Phillips American Institutes for Research

 Envisioned by the CCSSO/NGA Common Standards.  Promoted by the Race to the Top Assessment Program competition. Gary W. Phillips American Institutes for Research 29

Gary W. Phillips American Institutes for Research 30

Gary W. Phillips American Institutes for Research 31

Gary W. Phillips American Institutes for Research 32

 When the data were reanalyzed based on a level playing field there was a dramatic drop among the states reporting the highest levels of proficiency. Using Grade 4mathematics as an example we find: ◦ Tennessee and Colorado dropped from 90% and 91% to 29% and 40%, respectively. ◦ ◦ Massachusetts goes from being one of the states with the lowest percent proficient (49%) based on NCLB results to the highest achieving state with 63% percent proficient. ◦ This is consistent with NAEP results that indicate Massachusetts is the highest achieving state in the nation. Gary W. Phillips American Institutes for Research 33