EDUC 5535 Spring 2013.  An artifact of the eugenics movement (in the 1920’s) - an attempt to sort people by their perceived intelligence or ability.

Slides:



Advertisements
Similar presentations
Interpreting Test Results For Grades K-8. What tests will my child take? Students are assessed through: DIBELS (Dynamic Indicators of Basic Early Literacy.
Advertisements

Assessment & Evaluation adapted from a presentation by Som Mony
Cal State Northridge Psy 427 Andrew Ainsworth PhD
Validity and Reliability
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 6 Validity.
Using statistics in small-scale language education research Jean Turner © Taylor & Francis 2014.
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
Chapter Fifteen Understanding and Using Standardized Tests.
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
Reliability, Validity, Trustworthiness If a research says it must be right, then it must be right,… right??
JHLA Junior High Literacy Assessment. The school year saw the first administration of the Junior High Literacy Assessment. The assessment was.
VALIDITY.
Challenge Question: Why is being organized on the first day of school important? Self-Test Questions: 1.How do I get my classroom ready? 2.How do I prepare.
BASIC PRINCIPLES OF ASSSESSMENT RELIABILITY & VALIDITY
Developing Literacy in English- language Learners: Key Issues and Promising Practices Diane August David Francis Claude Goldenberg Timothy Shanahan.
1 Chapter 6: Assessment of Educational Ability Survey Battery Diagnostic Readiness Cognitive Ability Tests.
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Understanding Validity for Teachers
Comprehensive Assessment System Webinar #6 December 14, 2011.
Questions to check whether or not the test is well designed: 1. How do you know if a test is effective? 2. Can it be given within appropriate administrative.
Chapter 4. Validity: Does the test cover what we are told (or believe)
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Test Validity S-005. Validity of measurement Reliability refers to consistency –Are we getting something stable over time? –Internally consistent? Validity.
Validity and Reliability
Creating Assessments with English Language Learners in Mind In this module we will examine: Who are English Language Learners (ELL) and how are they identified?
Problem Based Learning. Facts The school is on the state list to be taken over because of failing ELA scores. Parents can use a voucher if scores don’t.
Instrument Validity & Reliability. Why do we use instruments? Reliance upon our senses for empirical evidence Senses are unreliable Senses are imprecise.
Reliability and Validity what is measured and how well.
Instrumentation.
Technical Adequacy Session One Part Three.
Validity & Practicality
Chapter 15 - Testing Psychology McGonigle. Use of Tests Psychological Tests – can help people make decisions (Binet & Wechsler) Placement tests- Can indicate.
Chap. 2 Principles of Language Assessment
Validity Is the Test Appropriate, Useful, and Meaningful?
Reliability vs. Validity.  Reliability  the consistency of your measurement, or the degree to which an instrument measures the same way each time it.
Session 7 Standardized Assessment. Standardized Tests Assess students’ under uniform conditions: a) Structured directions for administration b) Procedures.
Measurement Validity.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Session 4 Reliability and Validity. Validity What does the instrument measure and How well does it measure what it is supposed to measure? Is there enough.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Amber Narczewski & Holly Kent ECED Purpose of the Assessment: Helps educators identify students who may need additional literacy instruction.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Ch 9 Internal and External Validity. Validity  The quality of the instruments used in the research study  Will the reader believe what they are readying.
Chapter 9 Correlation, Validity and Reliability. Nature of Correlation Association – an attempt to describe or understand Not causal –However, many people.
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
 A test is said to be valid if it measures accurately what it is supposed to measure and nothing else.  For Example; “Is photography an art or a science?
Chapter 6 - Standardized Measurement and Assessment
Chapter 3 Selection of Assessment Tools. Council of Exceptional Children’s Professional Standards All special educators should possess a common core of.
PRINCIPLES OF LANGUAGE ASSESSMENT Riko Arfiyantama Ratnawati Olivia.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
Language Assessment Lecture 7 Validity & Reliability Instructor: Dr. Tung-hsien He
Assessing Intelligence. Test Construction Standardization – defining the meaning of scores by comparing to a pretested “standard group”. Reliability –
Reliability and Validity
Principles of Language Assessment
Assessing Intelligence
VALIDITY by Barli Tambunan/
Chapel Hill ISD Reading First Initiative
Concept of Test Validity
Test Validity.
Validity and Reliability
Validity and Reliability
Classroom Assessment Validity And Bias in Assessment.
Validity.
Week 3 Class Discussion.
پرسشنامه کارگاه.
Gazİ unIVERSITY M.A. PROGRAM IN ELT TESTING AND ASSESSMENT IN ELT «ValIdIty» PREPARED BY FEVZI BALIDEDE 2013, ANKARA.
Jennifer Rodriguez TB610-60
Qualities of a good data gathering procedures
Presentation transcript:

EDUC 5535 Spring 2013

 An artifact of the eugenics movement (in the 1920’s) - an attempt to sort people by their perceived intelligence or ability.  Short buzz: How are tests used to sort?

 Test  SAT  ITBS (Iowa Test of Basic Skills)  CELA  Sorting Purpose  Who goes to what college  Who is on grade level, who isn’t  Who is proficient in English/Who isn’t

 Were created because it was important to know if the tests used for sorting were accurate sorting measures (e.g. Are we sure that the children we are putting in special education really need to be there? Are we sure the ELL children know enough English to achieve in English medium classrooms?)

 Validity - Does the test measure what it purports to measure ◦ Construct ◦ Content ◦ Concurrent - criterion ◦ Predictive ◦ Consequential  More importantly - threats to validity and reliability

 The extent to which an assessment procedure adequately represents the content of the curricular aim being measured.  How do we determine content validity? ◦ Expert review (domain experts) ◦ External reviewers (to check on what the experts created) ◦ Expert Panel

 1. Which of these countries is not in South America? ◦ Brazil ◦ Canada ◦ Argentina ◦ Venezuela  2. How many continents are there?  Threats to validity? ◦ Reading and writing skill (if you do not read well you may not do well on the test) ◦ Cultural differences

 Red is to firetruck as _______________is to lemons.  Tortillas are to ________________ as _____________ is to a toaster.  Threats to validity – no OTL analogy, cultural issues, reading level.

 Every test in English is first and foremost a test of English.  Every paper/pencil test no matter what they content is first and foremost a test of literacy.  All tests have inherent cultural, linguistic and economic biases

 The extent to which empirical evidence confirms that an inferred construct exists and that a given assessment procedure is measuring the inferred construct accurately.  Construct validity is also determined by: ◦ Content experts ◦ External reviewers ◦ Expert panels

 Content  Conventions  Spelling  Genre  Audience  A writing assessment that has construct validity has all of the above constructs  What might be constructs related to being a good reader?

 Which title should be underlined? ◦ America the Beautiful ◦ Gone with the Wind ◦ Damn Yankees  Threats to construct validity? ◦ Child knows the rule for underlining but does not know which of the above is a song, book or play

 Read the text provided to you  Answer the comprehension questions  Final comprehension question: Choose a title for this essay.  Scoring:  Final question weighted more heavily because it is an inferential question  Threats to validity?

 Do outcomes from one assessment correlate positively to another assessment that purports to measure the same constructs?  Concurrent Validity is Measured by comparing outcomes of one assessment to another (e.g. compare LAS to CELA; ITBS to CSAP; SAT to ACT) look for correlations above.50

 Does the test/assessment predict future performance or behavior?  Do SATs predict preparedness for college?  Do GREs predict preparedness for graduate school?  Do 3 rd grade reading test scores predict who will struggle in high school?  Do school readiness test predict who is ready for kindergarten?  Does DIBELS predict reading comprehension?

 What are the consequences of outcomes on the assessment? Are the outcomes used the way the test creators intended?  Threats to validity  Tests given to a population they were not intended to be given to (e.g. CSAP to ELLs)  Tests used for unintended purposes (e.g. to rank schools and deem some ‘good’ and others ‘bad’

 Tests measure language skills of students and not knowledge of content  Tests contain economically, linguistically and culturally biased items  Tests were created for one population and given to another (e.g. ITBS or CSAP with L2 students)  Modifications of tests for L2 are inconsistent  Improperly trained people are administering the tests (e.g. paraprofessionals administering ACCESS)

 Invalid tests are often highly reliable  The consistent use of invalid tests creates the façade of an achievement gap

 Shepard discusses uses and abuses of tests (these are validity issues)  In your group identify what you think are the 3 most salient to YOUR work  Do these match practice in your district/school?  What should we do?