Assessment Framework and Test Blueprint

Slides:



Advertisements
Similar presentations
An Introduction to Test Construction
Advertisements

Performance Assessment
Wynne Harlen. What do you mean by assessment? Is there assessment when: 1. A teacher asks pupils questions to find out what ideas they have about a topic.
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Benchmark Assessment Item Bank Test Chairpersons Orientation Meeting October 8, 2007 Miami-Dade County Public Schools Best Practices When Constructing.
Evaluating tests and examinations What questions to ask to make sure your assessment is the best that can be produced within your context. Dianne Wall.
Score Reporting for Interim and Summative Assessments: Claims & Targets Nancy Thomas Price, Comprehensive Assessment System Coordinator.
Consistency of Assessment
Principles of High Quality Assessment
Implementing Mathematics K-6 Using the syllabus for consistency of teacher judgement © 2006 Curriculum K-12 Directorate, NSW Department of Education and.
Classroom Assessment A Practical Guide for Educators by Craig A
NCCSAD Advisory Board1 Research Objective Two Alignment Methodologies Diane M. Browder, PhD Claudia Flowers, PhD University of North Carolina at Charlotte.
Building Effective Assessments. Agenda  Brief overview of Assess2Know content development  Assessment building pre-planning  Cognitive factors  Building.
ASSESSMENT IN EDUCATION ASSESSMENT IN EDUCATION. Copyright Keith Morrison, 2004 PERFORMANCE ASSESSMENT... Concerns direct reality rather than disconnected.
Validation of the Assessment and Comparability to the PISA Framework Hao Ren and Joanna Tomkowicz McGraw-Hill Education CTB.
Looking At Your Assessment System: A Graphical Perspective Michigan Educational Research Association Fall Conference 2012 Monday, November 19, 2012.
Quality in language assessment – guidelines and standards Waldek Martyniuk ECML Graz, Austria.
Workshop 3 Early career teacher induction: Literacy middle years Workshop 3 Literacy teaching and NSW syllabus 1.
Validity & Practicality
Principles in language testing What is a good test?
1 Issues in Assessment in Higher Education: Science Higher Education Forum on Scientific Competencies Medellin-Colombia Nov 2-4, 2005 Dr Hans Wagemaker.
Teaching Today: An Introduction to Education 8th edition
Chap. 2 Principles of Language Assessment
CURRICULUM AND ASSESSMENT POLICY STATEMENT
Sue Drew Learning and Teaching Institute Sheffield Hallam University
USEFULNESS IN ASSESSMENT Prepared by Vera Novikova and Tatyana Shkuratova.
Programme Objectives Analyze the main components of a competency-based qualification system (e.g., Singapore Workforce Skills) Analyze the process and.
Understanding Alaska Measures of Progress Results: Reports 1 ASA Fall Meeting 9/25/2015 Alaska Department of Education & Early Development Margaret MacKinnon,
Assessment Specifications Gronlund, Chapter 4 Gronlund, Chapter 5.
Lecture by: Chris Ross Chapter 7: Teacher-Designed Strategies.
Developing NAEP Frameworks: A Look Inside the Process Mary Crovo Deputy Executive Director National Assessment Governing Board November 17, 2007.
Assessment Design. Four Professional Learning Modules 1.Unpacking the AC achievement standards 2.Validity and reliability of assessments 3. Confirming.
Assessment. Workshop Outline Testing and assessment Why assess? Types of tests Types of assessment Some assessment task types Backwash Qualities of a.
1 Scoring Provincial Large-Scale Assessments María Elena Oliveri, University of British Columbia Britta Gundersen-Bryden, British Columbia Ministry of.
Parent Workshop Year 2 Assessment without levels January 2016.
Moderation and Validation of Teacher Judgements in School.
LISA A. KELLER UNIVERSITY OF MASSACHUSETTS AMHERST Statistical Issues in Growth Modeling.
COURSE AND SYLLABUS DESIGN
The definition of table of specification. Table of specification is a chart that provides graphic representations of a related to the content of a course.
GET Intermediate Phase CAPS Training 1pas/2012. ASSESSMENT OUTCOMES : At the end of this Activity: participants will be able to differentiate between.
Required Skills for Assessment Balance and Quality: 10 Competencies for Educational Leaders Assessment for Learning: An Action Guide for School Leaders.
Training Week, August 2016 Assessment for
Curriculum Forum Secondary Tuesday 6 June 2017
Classroom Assessments Checklists, Rating Scales, and Rubrics
What is moderation and why should we moderate?
NEEDS ANALYSIS.
Classroom Assessment A Practical Guide for Educators by Craig A
ECML Colloquium2016 The experience of the ECML RELANG team
Assessments for Monitoring and Improving the Quality of Education
Test Blueprints for Adaptive Assessments
EDU 704-Assessment and Evaluation
Smarter Balanced Assessment Results
Classroom Assessment A Practical Guide for Educators by Craig A
國立臺灣師範大學英語系陳秋蘭 PISA 與英語閱讀素養 國立臺灣師範大學英語系陳秋蘭
Assessment Literacy Series
Validity and Reliability
Classroom Assessments Checklists, Rating Scales, and Rubrics
Clinical Assessment Dr. H
Managing Assessment in the Classroom
Booklet Design and Equating
پرسشنامه کارگاه.
TOPIC 4 STAGES OF TEST CONSTRUCTION
Assessment of student achievement of outcomes from the 7-10 Science syllabus Use this as a screen saver.
Brian Gong Center for Assessment
Critically Evaluating an Assessment Task
Assessment for Learning
COMPONENTS OF CURRICULUM
AACC Mini Conference June 8-9, 2011
VCE Religion and Society School-based Assessment 2019
Presentation transcript:

Assessment Framework and Test Blueprint Margaret Wu

Development of a Framework and Test Blueprint Draft items Item Development Development of a Framework and Test Blueprint Draft items Item panelling (shredding!) Iterative process: Draft items to illustrate, and clarify and sharpen up framework. Framework to guide item development.

Framework and Test Blueprint-1 Clearly identify ‘Why’ you are assessing (Purpose) ‘Whom’ to assess (Population) ‘What’ to assess (Construct domain) What need to be reported Overall score Subscores? For individual students or for cohorts (by district, gender, school)? Define parameters for the test,e.g.: Duration of the test and test administration procedures Scoring/marking constraints; item formats Other issues: security, feedback.

Specifying the Purpose How will the results be used? Determine pass/fail, satisfactory/unsatisfactory Award prizes Provide diagnostic information Compare students Set standards Provide information to policy makers Who will use the information? Teachers, parents, students, managers, politicians

Specifying the Population Grade, age level. In an industry. Profession. Ethnicity/Culture/Language issues. Gender. Notion of population and sample. Sampling method: random, convenience Size of sample Validity of test results could depend on the population/sample you are assessing

Specifying the Construct Domain - Examples Familiarity with sport: What is meant by ‘sport’? Include “Gym”? “Taichi?” “Gymnastics”? In Australian contexts? Define Problem Solving: As viewed in the workforce Workforce competencies As viewed in the cognitive sciences Cognitive processes: decoding, reasoning, domain specific knowledge Interpersonal skills Negotiation/conflict resolution skills Leadership skills Work with people from diverse backgrounds

Specifying the Construct Domain – Examples Achievement domains: Content oriented: Number, measurement, data, algebra Competency oriented: Conceptual understanding, procedural knowledge, problem solving Taxonomy of educational objectives (Bloom’s taxonomy of learning outcomes): cognitive and affective.

Blooms’ Taxonomy - cognitive Knowledge Comprehension Application Analysis Synthesis Evaluation

Considerations in defining the Construct of a test Validity Consideration Does the construct cover what the test is claimed to be assessing? E.g., language proficiency: speaking, listening, reading, writing Measurement Consideration How well the specifications for a construct “hang together” to provide meaningful scores? The idea of “unidimensionality” On-balance judgement Boundaries are never clear

Test Blueprint Sufficiently detailed so that test developers can work from these specifications. Range of difficulty Target reliability Item format. Weights of sub-domains Test administration procedures Timing, equipment, resources Marking requirements

Test Blueprint – example (PISA Reading) Aspect % of test % constructed % MC Retrieving information 20 7 13 Broad understanding Interpretation 30 11 19 Reflecting on content 15 10 5 Reflecting on form Total 100

Report by individual student What will be reported Any subscore By content domains. If so, then there needs to be sufficient number of items for each content domain Report by individual student Need to work out measurement error: the accuracy with which each student is measured. Typically, we need a very long test to measure individuals accurately. This has an impact on the test length. Report by subgroups Ensure that sample size is sufficient to report by subgroups Report possible factors related to achievement Need to collect student background information

Define parameters for the test Duration of the test Has an impact on measurement error If the purpose of the test is for cohort reporting, the tests can be shorter. Total number of items needed How many items are needed to cover the content domain? As many as possible. The use of rotated test booklet design Allow sufficient number of items to cover the content domain Prevent cheating Moderate fatigue effect How many open-ended/constructed items The need to hire markers The cost of marking

Uses of Frameworks & Blueprints To guide item development Don’t ignore specifications. Cross-check with specs constantly. To ensure that there is a clear and well-defined construct that can be stable from one testing occasion to another. Different item writing team Parallel tests To establish test validity Design test -> Administer test -> Process test results -> Interpret test results The last stage of interpreting test results relies on the test specifications (test blueprints) to establish what is being tested.

Adapting/borrowing test blueprints from other assessment projects Assessment framework may not suit – different purposes of assessment. e.g., differences between PISA and TIMSS (curriculum based or competency based) Identify learning goals, and not just what is taught Look beyond the textbook May avoid rote learning Avoid the possibility of students memorising test items Talk to stakeholders Policy makers Employers Educators

Using rotated booklet design to cover content areas and counteract fatigue effect Block 1 Block 2 Block 3 1 C1 C2 C4 2 C3 C5 3 C6 4 C7 5 6 7

An example of rotated test booklet design Block 1 Block 2 Block 3 1 C1 C2 C4 2 C3 C5 3 C6 4 C7 5 6 7