Assessment: Reliability & Validity. Reliability Refers to the repeatability of a given testing instrument The extent to which a student would be expected.

Slides:



Advertisements
Similar presentations
Wortham: Chapter 2 Assessing young children Why are infants and Preschoolers measured differently than older children and adults? How does the demand for.
Advertisements

1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
Reliability and Validity
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
VALIDITY AND RELIABILITY
Evaluation of the Iowa Algebra Aptitude Test Terri Martin Doug Glasshoff Mini-project 1 June 17, 2002.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 6 Validity.
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
Reliability, Validity, Trustworthiness If a research says it must be right, then it must be right,… right??
RESEARCH METHODS Lecture 18
Chapter 4 Validity.
RELIABILITY & VALIDITY What is Reliability? What is Reliability?What is Reliability?What is Reliability? How Can We Measure Reliability? How Can We Measure.
RELIABILITY & VALIDITY
Lecture 7 Psyc 300A. Measurement Operational definitions should accurately reflect underlying variables and constructs When scores are influenced by other.
Reliability & Validity the Bada & Bing of YOUR tailored survey design.
Chapter 7 Correlational Research Gay, Mills, and Airasian
Using statistics in small-scale language education research Jean Turner © Taylor & Francis 2014.
INTELLIGENCE AND PSYCHOLOGICAL TESTING. KEY CONCEPTS IN PSYCHOLOGICAL TESTING Psychological test: a standardized measure of a sample of a person’s behavior.
Classroom Assessment A Practical Guide for Educators by Craig A
Test Validity S-005. Validity of measurement Reliability refers to consistency –Are we getting something stable over time? –Internally consistent? Validity.
Classroom Assessment Reliability. Classroom Assessment Reliability Reliability = Assessment Consistency. –Consistency within teachers across students.
Curriculum-Based Measurements The What. Curriculum-Based Measurements  Curriculum-Based Measurements (CBM) Assessment tools derived from the curriculum,
NCLB AND VALUE-ADDED APPROACHES ECS State Leader Forum on Educational Accountability June 4, 2004 Stanley Rabinowitz, Ph.D. WestEd
VALIDITY, RELIABILITY, and TRIANGULATED STRATEGIES
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
Foundations of Recruitment and Selection I: Reliability and Validity
Principles of Test Construction
CCSSO Criteria for High-Quality Assessments Technical Issues and Practical Application of Assessment Quality Criteria.
Validity. Face Validity  The extent to which items on a test appear to be meaningful and relevant to the construct being measured.
Reliability & Validity
IES Evaluations and Data Collection Instruments Lauren Angelo National Center for Education Evaluation and Sally Atkins-Burnett Mathematica Policy Research.
Assessing Learners with Special Needs: An Applied Approach, 6e © 2009 Pearson Education, Inc. All rights reserved. Chapter 4:Reliability and Validity.
Reliability vs. Validity.  Reliability  the consistency of your measurement, or the degree to which an instrument measures the same way each time it.
Illustration of a Validity Argument for Two Alternate Assessment Approaches Presentation at the OSEP Project Directors’ Conference Steve Ferrara American.
Measurement Validity.
Selecting a Sample. Sampling Select participants for study Select participants for study Must represent a larger group Must represent a larger group Picked.
Oxford Preparatory Academy Scholar Academy Parent Social Topic: Changes in State Testing May 4, 5, and 6, 2015.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Designs and Reliability Assessing Student Learning Section 4.2.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
McGraw-Hill/Irwin © 2012 The McGraw-Hill Companies, Inc. All rights reserved. Obtaining Valid and Reliable Classroom Evidence Chapter 4:
Ch 9 Internal and External Validity. Validity  The quality of the instruments used in the research study  Will the reader believe what they are readying.
Minnesota's Approach to Comprehensive Assessment Megan E. Cox, Ph.D. Principal Leadership Academy January 11, 2016 Minnesota’s Approach to Comprehensive.
No Child Left Behind Impact on Gwinnett County Public Schools’ Students and Schools.
Chapter 3 Selection of Assessment Tools. Council of Exceptional Children’s Professional Standards All special educators should possess a common core of.
Testing. Psychological Tests  Tests abilities, interests, creativity, personality, behavior  Must be standardized, reliable, and valid  Timing, instructions,
Reliability EDUC 307. Reliability  How consistent is our measurement?  the reliability of assessments tells the consistency of observations.  Two or.
Career Counseling: A Holistic Approach
WHS AP Psychology Unit 7: Intelligence (Cognition) Essential Task 7-3:Explain how psychologists design tests, including standardization strategies and.
Measurement and Scaling Concepts
Terra Nova By Tammy Stegman, Robyn Ourada, Sandy Perry, & Kim Cotton.
Chapter 9 Intelligence. Objectives 9.1 The Nature of Intelligence Define intelligence from an adaptation perspective. Compare and contrast theories of.
Reduced STAAR test blueprints
Principles of Language Assessment
Reliability and Validity
Concept of Test Validity
Test Validity.
Tests and Measurements: Reliability
Reliability & Validity
پرسشنامه کارگاه.
Reliability and Validity of Measurement
Evaluating and Institutionalizing OD Interventions
Reliability & Validity.
Misc Internal Validity Scenarios External Validity Construct Validity
Qualities of a good data gathering procedures
Presentation transcript:

Assessment: Reliability & Validity

Reliability Refers to the repeatability of a given testing instrument The extent to which a student would be expected to perform similarly across multiple administrations of the test under similar conditions A test given today should produce, approximately, the same score if given tomorrow Assuming that no intervention has occurred

Types of Reliability Test-retest reliability. A method for assessing the reliability of an instrument by administering the same test at different points in time. The correlation between the 2 should be high. Alternate forms reliability. Refers to the relationship between scores on alternative forms of the same test, given closely in time. In the case of pre-post testing, alternate forms reliability is more important than test-retest.

Validity Refers to the extent to which a given test is measuring what it is expected to measure Ex. If a question is wrong on a math test because student couldn’t read the word problem, but s/he could do the math in the question, then the question is also measuring reading ability, not just math ability. The question, therefore, may not be a valid indicator of the student’s math ability.

Types of Validity Construct validity. Refers to evidence that a test is measuring the content & skills (or construct) it claims to measure. 2 common types: convergent validity: the extent to which the scores on a given test are related to scores on other tests designed to measure the same, or very similar, constructs discriminant validity: the extent to which the scores on a given test are related to scores on tests designed to measure something different.

Types of Validity Content validity. The extent to which the items/tasks adequately cover the area of interest. For ex.: in a test on reading comprehension, all of the items/tasks should: align w/the skills associated w/reading comprehension, and represent the full range of skills associated w/reading comprehension. Skills not associated w/this area, such as knowledge of science or social studies, should not influence performance External validity. The extent to which the results of a test given to one population of students are relevant to other groups of students. For ex.: test norms based on data collected from students enrolled in traditional classrooms may not be as useful for gauging the progress of students in alternative types of educational programs, especially w/respect to factors such as socioeconomic status & age

Works Cited National Education & Technical Assistance Center for the Education of Children & Youth who are Neglected, Delinquent, & At-Risk. A brief guide to selecting & using pre-post assessments. April, Found online at delinquent.org/nd/resources/spotlight/spotlight a.asphttp:// delinquent.org/nd/resources/spotlight/spotlight a.asp