Assessment Fall Quarter 2007.

Slides:



Advertisements
Similar presentations
Assessing Student Performance
Advertisements

Performance Assessment
Goals for Tonight Present the new Georgia Milestones Assessment System Discuss Our Instructional Strategies Discuss Your Role in supporting your child.
L2 program design Content, structure, evaluation.
Goals for Tonight Present the new Georgia Milestones Assessment System Discuss Our Instructional Strategies Discuss Your Role in supporting your child.
Testing What You Teach: Eliminating the “Will this be on the final
The Network of Dynamic Learning Communities C 107 F N Increasing Rigor February 5, 2011.
Alternative Assesment There is no single definition of ‘alternative assessment’ in the relevant literature. For some educators, alternative assessment.
Chapter Fifteen Understanding and Using Standardized Tests.
Chapter 8 Criteria and Validity PERSIAN GROUP. ارزیابی امتحان آزمون ارزیابی امتحان آزمون ارزیابی امتحان آزمون ارزیابی امتحان آزمون ارزیابی امتحان آزمون.
Developing Rubrics Presented by Frank H. Osborne, Ph. D. © 2015 EMSE 3123 Math and Science in Education 1.
C R E S S T / U C L A Improving the Validity of Measures by Focusing on Learning Eva L. Baker CRESST National Conference: Research Goes to School Los Angeles,
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Alignment Inclusive Assessment Seminar Brian Gong Claudia.
Dr. Robert Mayes University of Wyoming Science and Mathematics Teaching Center
Classroom Assessment A Practical Guide for Educators by Craig A
Becoming a Teacher Ninth Edition
Pre-Conference Workshop – June 2007 BUILDING A NATIONAL TEAM: Theatre Education Assessment Models Robert A. Southworth, Jr., Ed.D. TCG Assessment Models.
ASSESSMENT 19 September 2015Laila N. Boisselle 1.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Classroom Assessment A Practical Guide for Educators by Craig A
Classroom Assessment A Practical Guide for Educators by Craig A. Mertler Chapter 7 Portfolio Assessments.
Measuring Complex Achievement
Assessment in Education Patricia O’Sullivan Office of Educational Development UAMS.
Teaching Today: An Introduction to Education 8th edition
General Information Iowa Writing Assessment The Riverside Publishing Company, 1994 $39.00: 25 test booklets, 25 response sheets 40 minutes to plan, write.
Validity Is the Test Appropriate, Useful, and Meaningful?
Developing Assessments for and of Deeper Learning [Day 2b-afternoon session] Santa Clara County Office of Education June 25, 2014 Karin K. Hess, Ed.D.
Factors that determine the Cognitive Complexity of PARCC Mathematics Items Cognitive Complexity Mathematical Content Mathematical Practices Stimulus Material.
Week 5 Lecture 4. Lecture’s objectives  Understand the principles of language assessment.  Use language assessment principles to evaluate existing tests.
Investigating Assessment, Evaluation, and Reporting Hamilton Wentworth DSB Fri Apr 8, 2011 MaryLou Kestell
Classroom Evaluation & Grading Chapter 15. Intelligence and Achievement Intelligence and achievement are not the same Intelligence and achievement are.
Performance-Based Assessment HPHE 3150 Dr. Ayers.
Assessment Information from multiple sources that describes a student’s level of achievement Used to make educational decisions about students Gives feedback.
Alternative Assessment Chapter 8 David Goh. Factors Increasing Awareness and Development of Alternative Assessment Educational reform movement Goals 2000,
Mini-Project #2 Quality Criteria Review of an Assessment Rhonda Martin.
 A test is said to be valid if it measures accurately what it is supposed to measure and nothing else.  For Example; “Is photography an art or a science?
Criterion-Referenced Testing and Curriculum-Based Assessment EDPI 344.
Performance Based Assessment. What is Performance Based Assessment? PBA is a form of assessment that requires students to perform a task rather than an.
Foundations of American Education: Perspectives on Education in a Changing World, 15e © 2011 Pearson Education, Inc. All rights reserved. Chapter 11 Standards,
Alternative Assessment Larry D. Hensley University of Northern Iowa Chapter 8.
Language Assessment.
EVALUATING EPP-CREATED ASSESSMENTS
Classroom Assessments Checklists, Rating Scales, and Rubrics
Chapter 8: Performance-Based Strategies
Take-Home Message: Principles Unique to Alternate Assessments
Designing Rubrics with the Three Categories of Knowledge
Assessment Fall Quarter 2007.
Classroom Assessment A Practical Guide for Educators by Craig A
Assessment and Evaluation
Formative and Summative Assessment
Chapter 5: Assessment and Accountability
ASSESSMENT OF STUDENT LEARNING
Classroom Assessments Checklists, Rating Scales, and Rubrics
Classroom Assessment Validity And Bias in Assessment.
Research on Using Observation Systems with Special Educators
Learning About Language Assessment. Albany: Heinle & Heinle
Designing Assessment Things to be considered:
What Are Rubrics? Rubrics are components of:
February 24, 2009 Joint Session of Congress
Understanding and Using Standardized Tests
Assessment Methods.
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
Alternative Assessment
HS Physical Science Spring 2017
Assessment Elementary Mathematics
Characteristics of good assessments
Alternative Assessment
Why do we assess?.
Looking at Student Work
Presentation transcript:

Assessment Fall Quarter 2007

High-Stakes Testing All Tests Have Shortcomings Tests Tend to Concentrate on Discrete Thinking Knowledge/skill in bits and pieces Emphasis on recognition rather than production of knowledge

When Testing Has A Lesser Role Discrete Test Items Sample From A Larger Valued Set of Proficiencies But These Test Items Do Not Embody All that We Mean by A High-Quality Education. Still, Inferences to the Domain are Valid

When Testing has a Larger Role If the Stakes are Raised, Then the Content of Testing Begins to Define Educational Goals Instruction is Shaped Accordingly Test Scores Are No Longer Predictive of a Broader and More Significant Educational Achievement Test scores lose their meanings They no longer “sample” from the larger domain”

How to Respond? Possibly, Reduce the Extent of Testing But there is growing accountability Possibly, Make Tests More Worthy of Their New Role So that teaching to the test does good rather than harm

Features of New Assessments Alternative, Authentic, and Performance Assessment All Involve Tasks That Are Complex, Extended in Time, and Intrinsically Worthwhile The Emphasize Student Production It’s Implicit That Students Are Thinking at a High Level Critical thinking and problem solving rather than recognition and recall

Assessment in the Service of Instruction Blending of Assessment and Instruction Classroom or formal assessment Teaching (Like Testing) Should Involve Diagnosis and Response Assessment (Like Teaching) Should Always Have Value for Learning Instruction Assessment

Where Do Assessments Come From? Goals and Objectives for Education Are the Prerogative of Multiple Stakeholders Standards Frameworks National and state frameworks Illustrative tasks General Rubric E.g., characteristics of a good essay

Creating the Assessment Creating the Assessment Framework (Specifications) Knowledge objectives by process dimensions Knowledge: Planets, molecules, vertebrates Process: Reasoning, representational forms, types of items (e.g., constructed response, multiple-choice). Fairness Entails Opportunity to Learn

Scoring Rubrics Scoring Must Be Reliable and Fair Scoring Rubrics Can Focus Efforts of Students and Teachers This is what constitutes good work Clear, Unambiguous, Meaningful Separated Into Performance Levels Descriptions of each level (anchoring)

Types of Scores Holistic or Analytic? Holistic is overall impression Analytic involves aggregation of several dimensions Norm-Referenced or Criterion-Referenced? Norm-referenced is comparative (e.g., standard scores) Criterion-referenced cites performance against a standard Can be both, and the two can cross-map

Reliability Reliability is the Replicability of the Test Score What happens when a tape measure is made of rubber? One Positive Feature of Traditional Testing Large number of questions (items) tends to produce high reliability Performance testing often has lower reliability Whenever Possible, Increase the Number of Observations On Performance Grading, Ensure that Scorers are Well-Trained Reliable Measurement Does Not Promise Validity

Ensuring Reliable Scoring Clear, Well-Defined, and Defensible Criteria Well-Trained Scorers Evidence of Inter-Rater Agreement Document this Procedure of Resolving Score Divergence A third scorer? Beware of Rater Fatigue; Rater Differences Is Training of Scorers Professional Development?

Validity Validity is Concerned with the Meaning of What is Measured Validity is an Argument That a Test Really Measures What it Purports to Measure Construct Validity is the Most General Form A construct is the psychological dimension measured Other Forms of Validity Content: Established partly by test specifications Predictive: Does the test relate to future success? Reliability Places an Upper Bound on Validity If you can’t measure it accurately, it can’t be meaningful