In Today’s Society Education = Testing Scores = Accountability Obviously, Students are held accountable, But also!  Teachers  School districts  States.

Slides:



Advertisements
Similar presentations
Ed-D 420 Inclusion of Exceptional Learners. CAT time Learner-Centered - Learner-centered techniques focus on strategies and approaches to improve learning.
Advertisements

Chapter 6 Process and Procedures of Testing
Standardized Tests: What Are They? Why Use Them?
Wortham: Chapter 2 Assessing young children Why are infants and Preschoolers measured differently than older children and adults? How does the demand for.
Evaluation of the Iowa Algebra Aptitude Test Terri Martin Doug Glasshoff Mini-project 1 June 17, 2002.
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Issues Related to Assessment with Diverse Populations
Assessment: Reliability, Validity, and Absence of bias
VALIDITY.
Teaching and Testing Pertemuan 13
Chapter 5 Instrument Selection, Administration, Scoring, and Communicating Results.
Chapter 7 Correlational Research Gay, Mills, and Airasian
Classroom Assessment A Practical Guide for Educators by Craig A
Assessing and Evaluating Learning
Questions to check whether or not the test is well designed: 1. How do you know if a test is effective? 2. Can it be given within appropriate administrative.
By: Allan & Nadeen Kaufman Published by: American Guidance Service.
Mini Project I-- Evaluation of a Standardized Test By Marcia Luebbe.
Foundations of Recruitment and Selection I: Reliability and Validity
Evaluation of Nelson-Denny Reading Test Mini Project I By Becky Brandl.
Research Seminars in IT in Education (MIT6003) Research Methodology I Dr Jacky Pow.
Process Skill identify methods used by archaeologists, anthropologists, historians, and geographers to analyze evidence.[WHS.29A] October 2014WORLD HISTORY.
Week 5 Lecture 4. Lecture’s objectives  Understand the principles of language assessment.  Use language assessment principles to evaluate existing tests.
Reliability vs. Validity.  Reliability  the consistency of your measurement, or the degree to which an instrument measures the same way each time it.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 2 Fundamentals of Assessment.
Session 7 Standardized Assessment. Standardized Tests Assess students’ under uniform conditions: a) Structured directions for administration b) Procedures.
The KOPPITZ-2 A revision of Dr. Elizabeth Koppitz’
Child Assessment Cognitive & Language Development The options are in 2 major categories 1)Parent report (e.g. PARCA) 2)Researcher administered (e.g. BAS,
C R E S S T / U C L A UCLA Graduate School of Education & Information Studies Center for the Study of Evaluation National Center for Research on Evaluation,
Research in Communicative Disorders1 Research Design & Measurement Considerations (chap 3) Group Research Design Single Subject Design External Validity.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Issues in Validity and Reliability Conducting Educational Research Chapter 4 Presented by: Vanessa Colón.
Unit 5: Improving and Assessing the Quality of Behavioral Measurement
1 Comprehensive Accountability Systems: A Framework for Evaluation Kerry Englert, Ph.D. Paper Presented at the Canadian Evaluation Society June 2, 2003.
Assessment Specifications Gronlund, Chapter 4 Gronlund, Chapter 5.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
8th Grade Criterion-Referenced Math Test By Becky Brandl, Shirley Mills and Deb Romanek.
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
Ch 9 Internal and External Validity. Validity  The quality of the instruments used in the research study  Will the reader believe what they are readying.
Alternative Assessment Chapter 8 David Goh. Factors Increasing Awareness and Development of Alternative Assessment Educational reform movement Goals 2000,
Mini-Project #2 Quality Criteria Review of an Assessment Rhonda Martin.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. English Language Learners Assessing.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Chapter 14: Affective Assessment
Chapter 6 - Standardized Measurement and Assessment
Test of Nonverbal Intelligence.  Used for screening  Nonverbal intelligence test  Measures intelligence, aptitude, abstract reasoning, and problem.
VALIDITY, RELIABILITY & PRACTICALITY Prof. Rosynella Cardozo Prof. Jonathan Magdalena.
The population in a statistical study is the entire group of individuals about which we want information The population is the group we want to study.
Chapter 3 Selection of Assessment Tools. Council of Exceptional Children’s Professional Standards All special educators should possess a common core of.
Foundations of American Education: Perspectives on Education in a Changing World, 15e © 2011 Pearson Education, Inc. All rights reserved. Chapter 11 Standards,
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
LISA A. KELLER UNIVERSITY OF MASSACHUSETTS AMHERST Statistical Issues in Growth Modeling.
Building an Interim Assessment System: A Workbook for School Districts CCSSO National Conference on Student Assessment Detroit, MI June 22, 2010.
UNIT Standardization and Technical Properties n Standardization Sample n Reliability Studies Internal Consistency Reliabilities at Decision-Making Points.
Critiquing Quantitative Research.  A critical appraisal is careful evaluation of all aspects of a research study in order to assess the merits, limitations,
San Luis Valley Gifted Education Network Meeting October 17, 2013.
Educational Research Chapter 5 Selecting Measuring Instruments Gay and Airasian.
Lesson Thirteen Standardized Test. Contents Components of a Standardized test Reasons for the Name “Standardized” Reasons for Using a Standardized Test.
Terra Nova By Tammy Stegman, Robyn Ourada, Sandy Perry, & Kim Cotton.
Quality Assurance processes
Principles of Language Assessment
CCMH 535 RANK Career Begins/cchm535rank.com
Standardized Test Overview
Reliability and Validity in Research
Weschler Individual Achievement Test
A process for geographical inquiry
Jennifer Rodriguez TB610-60
EDUC 2130 Quiz #10 W. Huitt.
Presentation transcript:

In Today’s Society Education = Testing Scores = Accountability Obviously, Students are held accountable, But also!  Teachers  School districts  States

You should gather the information you need to evaluate a test. 1. Have an idea of what you want the test to measure. 2. Review a sample test from publisher. 3. Analyze test reviews.  Buros and Pro-Ed Test Locators  MMY & ERIC databases 4. Determine for yourself whether the publisher has made a compelling case that the test is valid and appropriate for your intended use.

WE WILL DISCUSS THE FOLLOWING ISSUES CONCERNING TEST EVALUATION A. Test coverage and use B. Appropriate samples for test validation and norming C. Reliability D. Predictive validity E. Test administration F. Test reporting

A. Test coverage and use The principal question to be asked in evaluating a test is whether it is appropriate for your intended purposes and your students.

B. Appropriate samples for test validation and norming The individuals in the norming and validation samples should be representative of the group for which the test is intended in terms of age, experience, and background.

C. Reliability Reliability refers to the consistency of assessment results. Fundamental to the evaluation of any instrument is the degree to which test scores are free from various sources of measurement error and are consistent from one occasion to another.

D. Predictive validity Validity is the appropriateness of the interpretations made from test scores and other evaluation results. In terms of an achievement test, predictive validity refers to the extent to which a test can be appropriately used to draw inferences regarding achievement. Example: high or low validity.

E. Test Administration To insure your test is valid: Directions should be clear. Questions should be easy to read. (use basic vocabulary & sentence structure) Ample time allowed.

E. Test Administration, Cont. The test is not biased or offensive with regard to race, sex, native language, ethnic origin, geographic region or other factors. Tests are not expected to yield equivalent mean scores across population groups.

F. Test Reporting The methods used to report test results, including scaled scores, subtests results and combined test results, are described fully along with the rationale for each method. Test results should be presented in a manner that will help schools, teachers and students to make decisions that are consistent with appropriate uses of the test. Help should be available for interpreting and using the test results.

IN CONCLUSION: Since testing is such an important part of education, it must be done correctly! Test Selection & Evaluation is critical. Be sure your tests are: Reliable, Valid, Administered fairly!