Design framework Item development Cognitive complexity Sample items

Slides:



Advertisements
Similar presentations
Level 1 Recall Recall of a fact, information, or procedure. Level 2 Skill/Concept Use information or conceptual knowledge, two or more steps, etc. Level.
Advertisements

Learning Outcomes Participants will be able to analyze assessments
Understanding Depth 0f knowledge
The Network of Dynamic Learning Communities C 107 F N Increasing Rigor February 5, 2011.
Leadership for the Common Core in Mathematics, University of Wisconsin-Milwaukee Reviewing the Cognitive Rigor Matrix and DOK Tuesday September.
CCSS Verbs of Math: Habits of a Mathematical Thinker July , 2013.
Elementary Principals CCSS Update Robyn Seifert & Rita Reimbold April 10, 2013.
Benchmark Assessment Item Bank Test Chairpersons Orientation Meeting October 8, 2007 Miami-Dade County Public Schools Best Practices When Constructing.
Understanding the Smarter BalanceD Math Summative Assessment
Leadership for the Common Core in Mathematics, University of Wisconsin-Milwaukee Linking Assessment Targets to Instructional Tasks and DOK This.
THE VISION OF THE COMMON CORE: EMBRACING THE CHALLENGE UCDMP SATURDAY SERIES SECONDARY SESSION 5 MAY 3, 2014.
Smarter Balanced Assessment Consortium. Structure of the Common Core State Standards for Mathematics Research-based learning progressions Internationally.
An Overview of Webb’s Depth of Knowledge
Webb’s Depth of Knowledge USD 457. Objectives CO – Analyze and apply the four levels of Webb’s Depth of Knowledge. LO – Read for the purpose of understanding.
M-Step Math Prep Grades 3-5 February 25, 2015 KRISTI HINEMAN, TRISH DUNN, MARIANNE SROCK,
ABC Curriculum Showcase Coventry Public Schools February 10, 2009 An introduction for parents and community members Left click your mouse to advance to.
Math Learning Progression
Introduction to Depth of Knowledge
Moving to the Common Core Janet Rummel Assessment Specialist Indiana Department of Education.
Design framework Item development Cognitive complexity Sample items ELA ITEM DESIGN.
UNDERSTANDING CCSS AND K- 12 INSTRUCTIONAL SHIFTS IN MATH Jamie Sirois, Cooperative Middle School, Statham, NH Adapted From: Maxine Mosely CCRS 101 NH-CCRS.
1 North Dakota Common Core State Standards (CCSS) Grades K-12 Adopted June 2011 Effective July 1, 2013 “After July 1, 2013, all public school districts.
The Common Core State Standards for Mathematics Transitioning to the Common Core.
Three Shifts of the Alaska Mathematics Standards.
© 2013 Boise State University1 What the Shifts in the Standards Mean for Learning and Instruction Michele Carney, PhD Spring 2014.
Philomath School District Board of Directors Work Session May 10, 2012.
Mathematics Shift 1: Focus What the Student Does…What the Teacher Does…What the Principal Does… Spend more time thinking and working on.
Evaluating Student Growth Looking at student works samples to evaluate for both CCSS- Math Content and Standards for Mathematical Practice.
RUBRIC DEVELOPMENT Why use rubrics? Characteristics of good rubrics Definition & types of rubric Strategies for rubric development SOLO taxonomy Phenomeno-
Adapted from Engageny.org. What the Student Does… What the Teacher Does… What the Principal Does… Spend more time thinking and working on fewer concepts.
Assessment Literacy Module
DOK Depth of Knowledge An Introduction.
Math rigor facilitating student understanding through process goals
Introduction ELA Selected Themes Assessment targets Progress Map ENGLISH LANGUAGE ARTS LEARNING PROGRESSION.
The Depth of Knowledge (DOK) Matrix
A CLOSER LOOK AT THE CCSS FOR MATHEMATICS COMMON CORE STATE STANDARDS PRESENTED BY: BEATRIZ ALDAY.
Writing Learning Outcomes David Steer & Stephane Booth Co-Chairs Learning Outcomes Committee.
Lunch and Learn: A Collaborative Time for North Carolina Teachers Your Presenter Information.
The Revised Bloom’s Taxonomy (RBT): Improving Curriculum, Instruction, and Assessment in an Accountability-Driven, Standards-Based World Developed and.
NEW REALITY STUDENTS MUST HAVE HIGHER-ORDER THINKING SKILLS 1.
Quality Instruction The ♥ of the matter is the depth of knowledge.
Review of BAS Learning progression Item design Outcomes/Sco ring Assessment quality Outline OVERVIEW OF BEAR ASSESSMENT SYSTEM.
Developing Assessments for and of Deeper Learning [Day 2b-afternoon session] Santa Clara County Office of Education June 25, 2014 Karin K. Hess, Ed.D.
Depth of Knowledge Assessments (D.O.K.) Roseville City School District Leadership Team.
1 Alignment of Alternate Assessments to Grade-level Content Standards Brian Gong National Center for the Improvement of Educational Assessment Claudia.
Overview Criteria Item Examples.  Create a sense of “measured urgency” ◦ Timeline is not immediate, but close ◦ Urgency in you and your teachers ◦ Difference.
CHAPTER 3 SCORING & MODERATION Strategies for scoring Reliability Scoring moderation.
Lecture by: Chris Ross Chapter 7: Teacher-Designed Strategies.
An Analysis of Three States Alignment Between Language Arts and Math Standards and Alternate Assessments Claudia Flowers Diane Browder* Lynn Ahlgrim-Delzell.
Depth of Knowledge and the Cognitive Rigor Matrix 1.
Bridge Year (Interim Adoption) Instructional Materials Criteria Facilitator:
By Benjamin Newman.  Define “Cognitive Rigor” or “Cognitive Demand”  Understand the role (DOK) Depth of Knowledge plays with regards to teaching with.
Welcome to Curriculum Night I am so glad you are here. Feel free to ask questions. Remember our time is short. Welcome to Curriculum Night for Ms. Figurelli’s.
Welcome Principals Please sit in groups of 3 2 All students graduate college and career ready Standards set expectations on path to college and career.
Mathematics in the Common Core 6 instructional shifts cassandra sponseller director of teaching and learning.
Evaluating individuals Evaluating classes Noting unusual patterns or trends Identifying gaps in learning USE OF SCORES.
What is RIGOR in the Math Classroom?. Presenter Introduction: Stephanie Darley East Valley RttT Math Coach Curriculum Coach at DHHS Been with TLG for.
Write your personal definition of “cognitive rigor” What do rigorous academic environments look and sound like?
Balancing on Three Legs: The Tension Between Aligning to Standards, Predicting High-Stakes Outcomes, and Being Sensitive to Growth Julie Alonzo, Joe Nese,
SBAC-Mathematics November 26, Outcomes Further understand DOK in the area of Mathematics Understand how the new SBAC assessments will measure student.
New Hope-Solebury School District. Develop a shared understanding of the concept of cognitive rigor Begin the conversation about Webbs’ Depth of Knowledge.
Performance Assessment: The Core of Competency Based Learning Rose ColbyIowa ASCDJune, 2014.
Linking Curriculum, Instruction and Assessment in Mathematics March 10, 2016 CONNECTICUT STATE DEPARTMENT OF EDUCATION.
Day Two: February 25, :30-3:00. Series Goals Participants will have the opportunity to:  Work collaboratively to:  Deepen their knowledge of the.
Training of Trainers Workshop
Claudia Flowers, Diane Browder, & Shawnee Wakeman UNC Charlotte
Claudia Flowers, Diane Browder, & Shawnee Wakeman UNC Charlotte
Presentation transcript:

Design framework Item development Cognitive complexity Sample items Math Item Design Once we’ve selected the types of items we want to develop, we then need to think about design frameworks, the alignment of items to the standards, and other issues of cognitive complexity. We’ll cover some of those ideas in this chapter.

4 building blocks Learning Progression Item Design Assessment Quality We’re still working on the second building block of the BEAR Assessment System – item design. Assessment Quality Outcomes/Scoring

Item Design Framework BEAR Assessment System Step 2 A match between what is taught and what is assessed Constructive alignment  aligning teaching and assessment to the learning outcomes/standards (Biggs, 2003) Proposed items are located along the LP map Learning Outcomes/Standards Assessment Task Teaching & Instruction One Framework (Wilson & Sloane, 2000) This is based on the principle that there should be a match between what is taught and what is assessed. In fact, things work best when there is an alignment between teaching, assessment, and the desired learning outcomes (in our case the Common Core standards). We work to achieve this by locating our proposed items along our map of the learning progression.

Alignment framework Item design framework used by Smarter-Balanced Assessment under the evidence-centered design approach (Mislevy, Steinberg, & Almond, 2003) Defined as the degree to which expectations and assessments work together to improve and measure students learning The Smarter Balanced assessment also uses the idea of alignment, which is defined as the degree to which expectations and assessments can be used to both measure and improve student learning.

4 criteria to determine the degree of alignment Categorical concurrence Commonality between the content categories of the standards and those of the assessment items Range of knowledge correspondence Number of objectives within each standard covered by item(s) Balance of representation Relative coverage of content categories by items in a test Depth of Knowledge consistency Match between the cognitive demand of items and the level of cognitive demand communicated by the wording of the objectives The Smarter Balanced consortium uses 4 criteria to determine the degree of alignment of items to the standards. These are defined as follows: First, Categorical concurrence: This is the commonality between the content categories of the standards and those covered by the assessment items Second is range of knowledge correspondence: Which is defined as the number of objectives within each standard covered by item(s) in question Third is balance of representation: Relative coverage of content categories by items in a test – how many items represent each of the content categories that we intend to cover Finally, we consider depth of knowledge consistency: This is the degree of match between the cognitive demand of items and the level of cognitive demand communicated by the wording of the objectives

Item development Universal design Design item that accurately assess the targeted competency for all students Ensure item fairness – make sure that items are equally difficult for groups of equal ability (e.g. males and females; urban and rural) Vocabulary & language use content-specific language appropriate to the assessed grade For non-content-specific material, use vocabulary/language from previous grade levels Other issues to consider when developing items are the ideas of universal design, and the issues of vocabulary and language Universal design requires that we Design item that accurately assess the targeted competency for all students, including those with special needs We must also ensure item fairness – we need to make sure that items are equally difficult for groups of equal ability (e.g. males and females; urban and rural). For example, we should take care not to develop math items that might draw too heavily on specific sports knowledge that might be less familiar to girls on average, or on farming ideas that would be more familiar to students in a rural environment than an urban one. We should pay attention to vocabulary & language by using content-specific language that is appropriate to the assessed grade. For non-content-specific material, use vocabulary and language from previous grade levels. Math in particular should not use complex vocabulary not needed, for example, in a story problem

Item development Grade appropriateness Design items that assess a primary content domain/standard of the appropriate grade “For non-reading items, the reading level is approximately one grade level below the grade level of the test, except for specifically assessed content terms or concepts” (SBAC, 2012) Using items to link tests For pre-post test designs, include some items that appeared on previous test(s) to measure student progress If the time between tests is relatively long (i.e. 2-3 months), the same test can be used If the time is short (i.e. 2-3 weeks), pick a few items to reuse and include new ones More issues include Grade appropriateness We should design items that assess a primary content domain or standard of the appropriate grade, not focusing on either an earlier or later grade. For non-reading items, the Smarter Balanced consortium recommends that the reading level be approximately one grade level below the grade level of the test, except for specifically assessed math-related content terms or concepts (e.g. the distributive law, the hypotenuse of a triangle) Sometimes, we must use the same items on more than one test. This is called linking It is most common in pretest-posttest designs, where we often include some items that appeared on previous tests to measure student progress We must be careful, however, that progress is real progress, and not just students’ memory of specific items If the time between tests is relatively long (say, around 2 to 3 months), the same test can usually be reused, unless it included very long or unusual items If the time is short (say 2 to 3 weeks), it is best to pick a few items to reuse and include new ones

Cognitive complexity Use of Modified Bloom’s Taxonomy Definition An example of Cognitive Rigor Matrix (Hess, et al., 2009) Demonstration on how to align standards and proposed item(s) on the LP map In the previous chapter, we mentioned the issue of cognitive complexity. The Smarter Balanced consortium represents this using the Cognitive Rigor matrix, which was included with the Chapter 5 video. This is based on a modification of Bloom’s taxonomy, which we’ll be looking at next. Then we will examine some sample items and how they align to both the learning progression and the cognitive rigor matrix.

Modified Bloom’s Taxonomy Modified by Anderson & Krathwohl (2001) Knowledge Comprehension Application Analysis Synthesis Evaluation The original Bloom’s taxonomy is on the right, the modification is on the left. In the original taxonomy, demonstrating knowledge was considered the lowest level skill, while evaluation was seen as the highest. Although this was useful, it was sometimes not the case that the levels were ordered as expected. In the newer, “cognitive process” version, several changes were made. First, some of the cells are renamed. The lowest level is now simply remembering information. The highest levels include evaluating and analyzing, but also creating something new Second, the top three levels are not seen as hierarchical. Creating, evaluating, and analyzing are all seen as highest-level skills Finally, this new version includes in each cell a number of verbs that can be used in item creating to cue students as to which of the skills to use in constructing their responses. For example, remembering could be cued by “describe” or “identify”, while analyzing could be cued by “compare and contrast” or creating by “design” or “construct” Old Bloom’s Taxonomy (Bloom, 1956)

Cognitive rigor matrix Here is the cognitive rigor matrix outline. The new cognitive process levels are listed as the row headers, and levels of knowledge depth are listed as the column headers. In the version of the matrix you have, each cell contains a detailed description of the kinds of skills a student using those cognitive processes, at that depth of knowledge, would be required to use. This can be used to classify items based on their depth of knowledge.

Sample items focus on Application process Level 1: Recall & Reproduction Level 2: Skills & Concepts Level 3: Strategic Thinking/ Reasoning Level 4: Extended Thinking Apply: Carry out or use a procedure in a given situations; carry out (apply to a familiar task) or use (apply) to an unfamiliar task. Follow simple/routine procedures. Apply an algorithm or formula. Solve routine problems applying multiple concepts or decision points. Use or apply concepts to solve non-routine problems. Select or devise an approach among many alternatives to solve a novel problem. The sample items we will be examining focus on the process of application. Here are the descriptions of each level of application. At level 1, students follow simple or routine procedures, such as applying a formula or algorithm. At level 2, students solve routine problems that require multiple concepts or decision points. At level 3, students apply concepts to solve non-routine problems. And, at the highest level of extended thinking, students must either select from among many alternatives, or devise their own approach to solve a novel problem.

Sample item* Intended level: Grade 8 Domain: Expressions and Equations Content standard: Analyze and solve linear equations and pairs of simultaneous linear equations. CCSS: Analyze and solve pairs of simultaneous linear equations (8.EE.8). Intended claims: 1, 2, 3 and 4 Here is the setup for our sample problems. It was originally a Smarter Balanced assessment example item at the 8th grade level, in the domain of Expressions and Equations , designed to assess the standard: Analyze and solve linear equations and pairs of simultaneous linear equations, and addresses the problem solving claim. Max is organizing a trip to the airport for a party of 75 people. He can use two types of taxi. A small taxi costs $40 for the trip and holds up to 4 people. A large taxi costs $63 for the trip and holds up to 7 people. Max is organizing a trip to the airport for a party of 75 people. He can use two types of taxi. A small taxi costs $40 for the trip and holds up to 4 people. A large taxi costs $63 for the trip and holds up to 7 people. * Adapted from SBAC (2013, p. 134)

Sample item Grade 8 How many taxis of each type should Max order, to keep the total cost as low as possible? Explain. LP level Specific Standard Cognitive Rigor Analyze and solve linear equations and pairs of simultaneous linear equations. Analyze and solve pairs of simultaneous linear equations (8.EE.8). Apply, Level 3 (Strategic Thinking/reasoning) Here is a sample problem at grade 8. Here Max must figure out how many of each taxi to order, to minimize cost. This will require the use of linear equations to determine the cost of persons in each sort of taxi, as well as a number of calculations to find the optimal scenario. This represents the skills described at the learning progression Grade 8 level, as shown in the table It also represents the Strategic thinking and reasoning level of the Apply process (as this is a non-routine problem).

Sample item Grade 7 Let L be the number of large taxis needed and S be the number of small taxis needed. a. Write an expression to show the number of taxis needed. b. If Max orders 6 large taxis, how many small taxis will he need? b. How much will the cost be? LP level Specific Standard Cognitive Rigor Solve real-life and mathematical problems using numerical and algebraic expressions and equations. Use variables to represent quantities in a real-world or mathematical problem, and construct simple equations and inequalities to solve problems by reasoning about the quantities. (7.EE.4). Apply, Level 2 (Skills & Concepts Here is a similar problem, designed for Grade 7. In this case, the problem is simpler, as the student is asked to write an equation, is given the number of large taxis, and then asked to find the number of small taxis and the cost This is at the grade 7 level on the learning progression, in which students reason about one variable at a time In addition, it is at the Skills and Concepts level of the apply process. It is a more routine problem than was the Grade 8 problem.

Sample item Grade 6 a. Let x be the number of large taxis needed. If Max wants to order large taxis only, evaluate x. b. If each person must equally share the cost of taking large taxi defined previously, write an expression for the cost of each person. Let y be the amount of money each person must pay. Calculate y. LP level Specific Standard Cognitive Rigor Apply and extend previous understandings of arithmetic to algebraic expressions. Write, read, and evaluate expressions in which letters stand for numbers. (6.EE.2). Apply, Level 1 (Recall and Reproduction) Finally, here is modification of the problem for Grade 6, using only one variable – the number of large taxis. If one were to use only this problem, it would be possible to change the problem description to remove the reference to both large and small taxis, and refer only to taxis which cost $63 and hold 7 people. This would then be at the Grade 6 level of the learning progression, requiring students to use expressions in which letters stand for numbers. This is also at the Recall and Reproduction level of the Apply process, which involves solving one-step problems.

Bibliography Bloom, B. S. (1956). Taxonomy of educational objectives. Handbook I: The Cognitive Domain. New York, NY: David McKay Co. Anderson, L. W., & Krathwohl, D. (2001). A Taxonomy for Learning, Teaching, and Assessing: A Revision of Bloom's Taxonomy of Educational Objectives. New York, NY: Longman. Hess, K., Carloc, D., Jones, B., & Walkup, J., (2009). What exactly do “fewer, clearer, and higher standards” really look like in the classroom? Using a cognitive rigor matrix to analyze curriculum, plan lessons, and implement assessments. Paper presented at CCSSO, Detroit, Michigan. Nitko, A. J., & Brookhart, S. (2007). Educational assessment of students. Upper Saddle River, NJ: Pearson Education, Inc. McMillan, J. H. (2007). Classroom assessment. Principles and practice for effective standard-based instruction (4th ed.). Boston: Pearson - Allyn & Bacon. Oregon Department of Education. (2014, June). Assessment guidance. Webb, N. (2007). Aligning assessments and standards. Retrieved from http://www.wcer.wisc.edu/news/coverStories/aligning_assessments_and_standards.php Wilson, M. (2005). Constructing measures: An item response modeling approach. New York, NY: Psychology Press, Taylor & Francis Group. Wilson, M., & Sloane, K. (2000). From principles to practice: An embedded assessment system. Applied Measurement in Education, 13 (2), pp. 181-208. Smarter Balanced Assessment Consortium. (2012, April). General item specifications. Smarter Balanced Assessment Consortium. (2013, June). Content specifications for the summative assessment of the Common Core State Standards for Mathematics. Revised draft. That’s the end of our chapter on item design. If you would like any additional information about what has been covered in this chapter, here are some possible references for you.  

Creative Commons License Item Design (Math) PPT by the Oregon Department of Education and Berkeley Evaluation and Assessment Research Center is licensed under a CC BY 4.0. You are free to: Share — copy and redistribute the material in any medium or format Adapt — remix, transform, and build upon the material Under the following terms: Attribution — You must give appropriate credit, provide a link to the license, and indicate if changes were made. You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use. NonCommercial — You may not use the material for commercial purposes. ShareAlike — If you remix, transform, or build upon the material, you must distribute your contributions under the same license as the original. Oregon Department of Education welcomes editing of these resources and would greatly appreciate being able to learn from the changes made. To share an edited version of this resource, please contact Cristen McLean, cristen.mclean@state.or.us.