THE MAP TEST -or- “You can attach a meaningless number to anything. Even a child.”

Slides:



Advertisements
Similar presentations
Ed-D 420 Inclusion of Exceptional Learners. CAT time Learner-Centered - Learner-centered techniques focus on strategies and approaches to improve learning.
Advertisements

Using the problem-finding, problem-solving model to present this issue.
Assessing Student Performance
Pennsylvania Value-Added Assessment System (PVAAS) High Growth, High Achieving Schools: Is It Possible? Fall, 2011 PVAAS Webinar.
Measures of Academic Progress
A Model Description By Benjamin Ditkowsky, Ph.D. Student Growth Models for Principal and Student Evaluation.
Changes to Assessing and Reporting Shawn Whyte Assessment Coach.
Standardized Tests: What Are They? Why Use Them?
Summative Assessment Kansas State Department of Education ASSESSMENT LITERACY PROJECT1.
MAP Information Martha - introduction October 23, 2013.
STUDENT GROWTH PART 3 10/16/14 ASSESSMENTS. REVIEW 1.Who are we going to target? 9 th grade 2.How much do they need to grow? Based on MAPS testing 1.5.
Item Response Theory in the Secondary Classroom: What Rasch Modeling Can Reveal About Teachers, Students, and Tests. T. Jared Robinson tjaredrobinson.com.
Target Testing 101 Parent Informational Power Point
Measures of Academic Progress. Make informed instructional decisions  Identify gaps/needs  Support specific skill development across content areas 
ASSESSMENT LITERACY PROJECT4 Student Growth Measures - SLOs.
Chapter Fifteen Understanding and Using Standardized Tests.
Using Growth Models for Accountability Pete Goldschmidt, Ph.D. Assistant Professor California State University Northridge Senior Researcher National Center.
7-2 Estimating a Population Proportion
MAP Measures of Academic Performance
About the tests PATMaths Fourth Edition:
Classroom Assessment A Practical Guide for Educators by Craig A
WHAT IS MAP? AMES Lunch and Learn.
Understanding Your Child’s Assessment: What Do Those Tests Really Tell You? Victoria B. Damiani, Ed.D., NCSP Center for Rural Gifted Education Indiana.
Section 9.1 Introduction to Statistical Tests 9.1 / 1 Hypothesis testing is used to make decisions concerning the value of a parameter.
Classroom Assessment and Grading
Understanding and Using Standardized Tests
Building Effective Assessments. Agenda  Brief overview of Assess2Know content development  Assessment building pre-planning  Cognitive factors  Building.
Austin Elementary School August 20 and 27, 2015
ASSESSMENT OF STUDENT LEARNING Manal bait Gharim.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Evaluation of software engineering. Software engineering research : Research in SE aims to achieve two main goals: 1) To increase the knowledge about.
PARAMETRIC STATISTICAL INFERENCE
Diagnostics Mathematics Assessments: Main Ideas  Now typically assess the knowledge and skill on the subsets of the 10 standards specified by the National.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
Grading and Analysis Report For Clinical Portfolio 1.
End-of –year Assessments MAP and DRA Testing Workshop for Parents
Lecture by: Chris Ross Chapter 7: Teacher-Designed Strategies.
Welcome to MMS MAP DATA INFO NIGHT 2015.
MAP: Measured Academic Progress© Parent Coffee February 10, 2010.
NWEA Measures of Academic Progress Universal American School Dubai.
TM Peoria Public Schools NWEA – Measure of Academic Progress
Tuesday, November 27, 2015 Board of Education Meeting District MAP.
The Normal Distribution and Norm-Referenced Testing Norm-referenced tests compare students with their age or grade peers. Scores on these tests are compared.
MAP TESTING MEASURES OF ACADEMIC PROGRESS DECEMBER 9, 2014.
Reliability EDUC 307. Reliability  How consistent is our measurement?  the reliability of assessments tells the consistency of observations.  Two or.
Assessment Assessment is the collection, recording and analysis of data about students as they work over a period of time. This should include, teacher,
Curriculum Night Elementary. What do I as a parent need to know to support student assessments at CCAS? Essential Question.
FSM NSTT Teaching Competency Test Evaluation. The NSTT Teaching Competency differs from the three other NSTT tests. It is accompanied by a Preparation.
What are your thoughts about Diverse Cultural Perspectives; particularly the section on Aboriginal Learners. Do you feel this is a true account of Mi'kmaq-Maliseet.
Curriculum Night Middle School. What do I as a parent need to know to support student assessments at CCAS? Essential Question.
Interpreting Test Results using the Normal Distribution Dr. Amanda Hilsmier.
Developmentally Appropriate Practices. Five Guidelines For Developmentally Appropriate Practices.
STAR Reading. Purpose Periodic progress monitoring assessment Quick and accurate estimates of reading comprehension Assessment of reading relative to.
M.A.P. Measures of Academic Progress
What is Value Added?.
M.A.P. Measures of Academic Progress
Formative and Summative Assessment
Update on Data Collection and Reporting
Concept of Test Validity
NWEA Measures of Academic Progress (MAP)
Prepared by: Toni Joy Thurs Atayoc, RMT
Assessing the Common Core Standards
Cognitive Abilities Test (CogAT)
Understanding and Using Standardized Tests
Developmentally Appropriate Practices
Validity and Reliability II: The Basics
Goals for Tonight Present the new Georgia Milestones Assessment System
Goals for Tonight Present the new Georgia Milestones Assessment System
Accelerated Reader® 101 for Parents
EDUC 2130 Quiz #10 W. Huitt.
Presentation transcript:

THE MAP TEST -or- “You can attach a meaningless number to anything. Even a child.”

This graph shows the most precise results obtained from MAP tests in the classroom of an expert teacher. Yet, 10 out of 38 of the growth scores (26%) fall outside a reasonable range. Further, scores did not match the student performance that the teacher was observing in the classroom. The graphs & statistical analysis for this presentation came from

MAP is based on the assumption that all students learn information in the same order, and all schools teach information in the same order. The RIT scores (RIT score = the final number assigned to each student by NWEA’s MAP test) are no more accurate than these assumptions.

What’s RIT ? That’s a loaded question… IRT is Item Response Theory. NWEA uses a specific IRT model conceived by Danish mathematician, Georg Rasch, ( ). They’ve apparently rebranded “IRT” as “RIT”. It’s proprietary…. NWEA claims that Characteristics of the RIT Scale include: It is an achievement scale. It is an accurate scale. It is an equal interval scale. It helps to measure growth over time. It has the same meaning regardless of grade or age of the student.

MAP levels are based on academic levels (curriculum sequencing), not cognitive levels (Bloom's Taxonomy). A high RIT score can represent low level cognition in advanced curriculum material. Thus, MAP is not a good measure for gifted performance, and should not be used to guide decisions for gifted programs. (as it has been repurposed in Seattle Public Schools)

MAP does not provide information as to why the student is having difficulty with specific material. MAP does not even provide specific information as to what material students need work on. As a result, MAP does not provide useful information that can guide teachers in meeting student's real needs. Teachers don’t see the test, and the test isn’t tied to their classes or curricula

All decisions based on test results should consider the 2-sigma range. (Greek letter sigma is a standard statistical result which describes the size of measurement error within a set of measurements. A low sigma represents more believable data. As astronomers say, ”The road to hell is paved with 3-sigma.”) NWEA reports sigma to be about 3 (except for high RIT scores where sigma is even larger.) Thus, all decisions should consider the possibility that each reported score could have been 6 point higher, or 6 points lower. Many schools that have retested with MAP have seen evidence that sigma is actually larger than 3.

RIT scores are actually determined by averaging together strand scores. Since a typical MAP test has 50 questions and 5 strands, only 10 questions were asked for each strand. Uncertainty for a 10 question multiple choice test is rather large. This is where the dream of the statistician meets the reality of the young student. Students can’t do tests for hours, and schools don’t have computer resources for students to do tests all the time. So you ask just a few questions, and you get results with a lot of, well, sigma.

One means of checking the reliability of a test is to compare its results to another test. This is particularly important if the goal is to increase the scores on the other test. Such would be the case for high stakes testing mandated by NCLB. Below is a comparison of the results of a MAP test to the results of an End of Grade Test (EOG) required by the state. Reliability was low. MAP ranked about 5 students a year higher in performance, and 4 students a year lower in performance, (23% total) than the EOG.

NWEA's technical manual uses r-values to estimate the reliability of the tests. The r- values they report range from 0.76 to 0.93 with most of the values being between 0.80 and But what does this mean for those wishing to use the tests to guide instruction? One can simulate the r-values to estimate what percentage of scores will lie within a reliable range for given r-values. An r-value of 0.92 could easily mean that over 25% of the data lies more than 6 points away from true. However, 6 points constitutes a year's normal growth for over half of the tests. As simulated in the graph above, even with an r-value of 0.92, 29% of the data is in error by more than a year's normal growth. This imprecision would lead to serious errors in identifying student needs, and then tracking the students incorrectly.

For high RIT scores (score that are normal for 8th grade and higher) the SEM (standard error measurement) is larger than a typical year's growth Thus, MAP is not only unable to identify what skills a student may need, MAP is unable to precisely determine the grade level at which student is performing.

For very high RIT scores, negative growth is more common that positive growth. This implies that either MAP is totally unreliable for high performers, or that MAP testing reinforces educational strategies that are counterproductive for high achievers.

If your school is considered adopting a standardized test, quiz the salesperson to demonstrate how the score reporting will help you. Give specific common examples from your experience. "I have students who can solve two-step equations with minimal difficulty, but frequently confuse the properties for negatives. How will the test report let me know what specific help these students need?" Be sure the testing company can give good answers to these questions. So, if the MAP test is not so useful, what should teachers look for ?

Ask the test salesperson to demonstrate how the test distinguishes between high level learning and accelerated knowledge.

Recommend formative assessments that identify specific strengths and weaknesses, as opposed to tests that rank for tracking purposes.

Parents & Teachers: Learn more about testing. Do not accept the “results” of any single type of assessment at face value. All tests have measurement errors, nowhere more so than in testing students, particularly when those students have no stake in the outcome. Test results are commonly used inappropriately by administration for invalid purposes like student placement and teacher evaluation. If you get a bad review based on test results alone, file a grievance. Insist the administrator demonstrate that the test was both precise and accurate enough to justify the results.