MEASUREMENT AND EVALUATION

Slides:



Advertisements
Similar presentations
Ed-D 420 Inclusion of Exceptional Learners. CAT time Learner-Centered - Learner-centered techniques focus on strategies and approaches to improve learning.
Advertisements

Assessing Student Performance
Assessment types and activities
© McGraw-Hill Higher Education. All rights reserved. Chapter 3 Reliability and Objectivity.
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Standardized Tests What They Measure How They Measure.
Types of Tests. Why do we need tests? Why do we need tests?
Chapter Fifteen Understanding and Using Standardized Tests.
Grading. Why do we grade? To communicate To tell students how they are doing To tell parents how students are doing To make students uneasy To wield power.
Introduction to Measurement and Evaluation
Classroom Assessment (1)
Chapter 4 Validity.
STANDARDIZED TESTING MEASUREMENT AND EVALUATION  All teaching involves evaluation. We compare information to criteria and ten make judgments.  Measurement.
Chapter 6 Developing Eva. Procedures Key Concepts Purposes of Evaluation Evaluation Procedure Design General Techniques of Eva. Procedures.
Characteristics of Sound Tests
Uses of Language Tests.
Lesson Three Kinds of Test and Testing. Yun-Pi Yuan 2 Contents Kinds of Tests: Based on Purposes  Classroom use Classroom use  External examination.
Introduction to Measurement and Evaluation
Standardized Test Scores Common Representations for Parents and Students.
Computer Integrated assessment. Computer integrated assessment Measurement, testing, assessment and evaluation What is the difference?
WEEK 1 – TOPIC 1 OVERVIEW OF ASSESSMENT: CONTEXT, ISSUES AND TRENDS 1.
Now that you know what assessment is, you know that it begins with a test. Ch 4.
Formative and Summative Assessment
Chapter 1 Assessment in Elementary and Secondary Classrooms
Chapter 14 Understanding and Using Standardized Tests Viewing recommendations for Windows: Use the Arial TrueType font and set your screen area to at least.
1 Development of Valid and Reliable Case Studies for Teaching, Diagnostic Reasoning, and Other Purposes Margaret Lunney, RN, PhD Professor College of.
Educational Psychology, 11 th Edition ISBN © 2010 Pearson Education, Inc. All rights reserved. Classroom Assessment, Grading, and Standardized.
Standardization the properties of objective tests.
Becoming a Teacher Ninth Edition
Classroom Assessment and Grading
Measuring Performance in Students Why this really important!
Instrument Validity & Reliability. Why do we use instruments? Reliance upon our senses for empirical evidence Senses are unreliable Senses are imprecise.
SHOWTIME!! EVALUATING ACHIEVEMENT INTRODUCTION BOTH CHILDREN AND ADULTS WANT TO KNOW HOW THEY COMPARE TO OTHERS OR A STANDARD PRIMARY ROLE OF TEACHER.
Chapter 14 Assessment of Classroom Learning. Copyright © Cengage Learning. All rights reserved. 14 | 2 Overview The Role of Assessment in Teaching Ways.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
How Can Teacher Evaluation Be Connected to Student Achievement?
CSD 5100 Introduction to Research Methods in CSD Observation and Data Collection in CSD Research Strategies Measurement Issues.
An Introduction to Measurement and Evaluation Emily H. Wughalter, Ed.D. Summer 2010 Department of Kinesiology.
EDU 385 Education Assessment in the Classroom
Lesson Three Kinds of Test and Testing. Contents Kinds of Tests: Based on Purposes  Classroom use Classroom use  External examination Kinds of Testing:
Chapter 7 Instrumentation. Empirical Data We need DATA We can’t rely solely upon our senses We develop INSTRUMENTS to compensate for the limitations of.
Session 7 Standardized Assessment. Standardized Tests Assess students’ under uniform conditions: a) Structured directions for administration b) Procedures.
Assessment What is it? Collection of relevant information for the purpose of making reliable curricular decisions and discriminations among students (Gallahue,
Performance-Based Assessment HPHE 3150 Dr. Ayers.
The Teaching Process. Problem/condition Analyze Design Develop Implement Evaluate.
TEST,MEASUREMENT AND EVALUATION
EDM 311 – Educational Testing, Measurement and Program Evaluation
A ssessment & E valuation. Assessment Answers questions related to individuals, “What did the student learn?” Uses tests and other activities to determine.
Classroom Assessment (1) EDU 330: Educational Psychology Daniel Moos.
Assessment and Testing
Assessment Information from multiple sources that describes a student’s level of achievement Used to make educational decisions about students Gives feedback.
Assessment. Workshop Outline Testing and assessment Why assess? Types of tests Types of assessment Some assessment task types Backwash Qualities of a.
Criterion-Referenced Testing and Curriculum-Based Assessment EDPI 344.
REVIEW I Reliability scraps Index of Reliability Theoretical correlation between observed & true scores Standard Error of Measurement Reliability measure.
ASSESSMENT CRITERIA Jessie Johncock Mod. 2 SPE 536 October 7, 2012.
Chapter 3 Selection of Assessment Tools. Council of Exceptional Children’s Professional Standards All special educators should possess a common core of.
Chapter 7 Criterion-Referenced Measurement PoorSufficientBetter.
Assessing Student Performance Characteristics of Good Assessment Instruments (c) 2007 McGraw-Hill Higher Education. All rights reserved.
GRADING n n Grading that reflects the actual ability of the student n n Grading as a personal communication.
Dr. Jeffrey Oescher 27 January 2014 Technical Issues  Two technical issues  Validity  Reliability.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 7 Assessing and Grading the Students.
Ch4-1. Testing.
ASSESSMENT METHODS – Chapter 10 –.
Assessment of Learning 1
Process of measurement, evaluation, identification, and prescription.
Bursting the assessment mythology: A discussion of key concepts
Concepts in Tests and Measurements
Understanding and Using Standardized Tests
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
Presentation transcript:

MEASUREMENT AND EVALUATION

IMPORTANCE AND PURPOSE OF MEASUREMENT AND EVALUATION IN HUMAN PERFORMANCE

DEFINITIONS MEASUREMENT - COLLECTION OF INFORMATION ON WHICH A DECISION IS BASED EVALUATION - THE USE OF MEASUREMENT IN MAKING DECISIONS

• INTERDEPENDENT CONCEPTS AS EVALUATION IS A PROCESS THAT USES MEASUREMENTS AND THE PURPOSE OF MEASUREMENT IS TO ACCURATELY COLLECT INFORMATION USING TESTS FOR EVALUATION • IMPROVED MEASUREMENT LEADS TO ACCURATE EVALUATION “GARBAGE IN, GARBAGE OUT”

OBJECTIVE VERSUS SUBJECTIVE TEST CONTINUUM OBJECTIVE TEST - 2 OR MORE PEOPLE SCORE THE SAME TEST AND ASSIGN A SIMILAR GRADE DEFINED SCORING SYSTEM AND TRAINED TESTERS INCREASES OBJECTIVITY HIGHLY SUBJECTIVE TEST LACKS A STANDARDIZED SCORING SYSTEM

EVALUATION COLLECT SUITABLE DATA (MEASUREMENT) JUDGE THE VALUE OF THE DATA ACCORDING TO SOME STANDARD (I.E., CRITERION-REFERENCED STANDARD OR NORM-REFERENCED STANDARD) MAKE DECISIONS BASED ON THE DATA

FUNCTIONS OF MEASUREMENT AND EVALUATION

PLACEMENT in classes/programs or grouping based on ability DIAGNOSIS of weaknesses EVALUATION OF ACHIEVEMENT to determine if individuals have reached important objectives

PREDICTION of an individual’s level of achievement in future activities or predict one measure from another measure PROGRAM EVALUATION MOTIVATION

FORMATIVE AND SUMMATIVE EVALUATION

FORMATIVE EVALUATION JUDGMENT OF ACHIEVEMENT DURING THE PROCESS OF LEARNING OR TRAINING PROVIDES FEEDBACK DURING THE PROCESS TO BOTH THE LEARNER/ATHLETE AND TEACHER/COACH “WHAT IS SUCCESSFUL AND WHAT NEEDS IMPROVEMENT”

SUMMATIVE EVALUATION JUDGMENT OF ACHIEVEMENT AT THE END OF AN INSTRUCTIONAL UNIT OR PROGRAM TYPICALLY INVOLVES TEST ADMINISTRATION AT THE END OF AN INSTRUCTIONAL UNIT OR TRAINING PERIOD USED TO DECIDE IF BROAD OBJECTIVES HAVE BEEN ACHIEVED

STANDARDS FOR EVALUATION

“EVALUATION IS THE PROCESS OF GIVING MEANING TO A MEASUREMENT BY JUDGING IT AGAINST SOME STANDARD”

CRITERION-REFERENCED (C-R) STANDARD IS USED TO DETERMINE IF SOMEONE HAS ATTAINED A SPECIFIED STANDARD NORM-REFERENCE (N-R) STANDARD IS USED TO JUDGE AN INDIVIDUAL’S PERFORMANCE IN RELATION TO THE PERFORMANCES OF OTHER MEMBERS OF A WELL-DEFINED GROUP

CRITERION-REFERENCED (C-R) STANDARDS ARE USEFUL FOR SETTING PERFORMANCE STANDARDS FOR ALL NORM-REFERENCED (N-R) STANDARDS ARE VALUABLE FOR COMPARISONS AMONG INDIVIDUALS WHEN THE SITUATION REQUIRES A DEGREE OF SENSITIVITY OR DISCRIMINATION IN ABILITY

• NORM-REFERENCED STANDARDS - DEVELOPED BY TESTING A LARGE GROUP OF PEOPLE - USING DESCRIPTIVE STATISTICS TO DEVELOP STANDARDS - PERCENTILE RANKS ARE A COMMON NORMING METHOD • MAJOR CONCERN - GROUP CHARACTERISTICS USED TO DEVELOP NORMS MAY NOT RESULT IN DESIRABLE NORMS; EXAMPLES WITH BODY COMPOSTION AND BLOOD CHOLESTEROL LEVELS WERE AVERAGE MAY NOT BE DESIRABLE

CRITERION-REFERENCED STANDARDS - PREDETERMINED STANDARD OF PERFORMANCE SHOWS THE INDIVIDUAL HAS ACHIEVED A DESIRED LEVEL OF PERFORMANCE - PERFORMANCE OF INDIVIDUAL IS NOT COMPARED WITH THAT OF OTHER INDIVIDUALS “COMMON PRACTICE TO APPLY A CRITERION-REFERENCED STANDARD TO A NORM-REFERENCED TEST”

DETERMINING ACCURACY OF CRITERION-REFERENCED (C-R) STANDARDS ACCURACY EXAMINED BY USING A 2 X 2 CONTIGENCY TABLE C-R TEST RELIABILITY EXAMINES THE CONSISTENCY OF CLASSIFICATION

LIMITATIONS OF CRITERION-REFERENCED (C-R) STANDARDS NOT ALWAYS POSSIBLE TO FIND A CRITERION THAT EXPLICITLY DEFINES MASTERY, PARTICULARLY IN SOME SKILLS

LIMITATIONS OF CRITERION-REFERENCED (C-R) STANDARDS ACCURACY OF C-R TEST VARIES WITH THE POPULATION BEING TESTED

EXAMPLE: ACCURACY OF EXERCISE STRESS TEST VARIES WITH THE DISEASE PREVALENCE IN THE GROUP STUDIED (I.E., PERCENTAGE OF PATIENTS WHO TRULY HAVE CORNOARY ARTERY DISEASE

MODELS OF EVALUATION

EDUCATIONAL MODEL

ADULT FITNESS MODEL

QUESTIONS OR COMMENTS??