AERA April 2005 Models and Tools for Drawing Inferences from Student Work: The BEAR Scoring Engine Cathleen Kennedy & Mark Wilson University of California,

Slides:



Advertisements
Similar presentations
Math Action Research Team
Advertisements

Level 1 Recall Recall of a fact, information, or procedure. Level 2 Skill/Concept Use information or conceptual knowledge, two or more steps, etc. Level.
Computer-Based Performance Assessments from NAEP and ETS and their relationship to the NGSS Aaron Rogat Educational Testing Service.
Well written objectives will… Provide clear direction for instruction Convey what is to be learned to students Provide a clear guide for assessment.
Understanding the Smarter BalanceD Math Summative Assessment
Designing Content Targets for Alternate Assessments in Science: Reducing depth, breadth, and/or complexity Brian Gong Center for Assessment Web seminar.
Show Me an Evidential Approach to Assessment Design Michael Rosenfeld F. Jay Breyer David M. Williamson Barbara Showers.
Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI International Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI.
University of Maryland Slide 1 July 6, 2005 Presented at Invited Symposium K3, “Assessment Engineering: An Emerging Discipline” at the annual meeting of.
Information from NCTM (1989, 1991, 1995, 2000) and Lelon Capps (personal communique 1993) Jamar Pickreign, Ph.D. Mathematics Assessment What is it? What’s.
Overview of the WVDE Interim Assessment Resources and the Interim Assessments Stacey Murrell, Ed.D. Interim Assessments Coordinator
Collaborating for Student Success Using Collaborative Inquiry with Student Teachers to Support Teacher Professional Development Sponsored by Teachers for.
Career and College Readiness (CCR) NGSS, CCSS ELA/Literacy, CCSS Mathematics, MMC K-12 Science Framework and NGSS Review in Terms of CCR 1.
Item Response Theory for Survey Data Analysis EPSY 5245 Michael C. Rodriguez.
Assessment Cadre #3: “Assess How? Designing Assessments to Do What You Want”
Collaborating for Student Success Using Collaborative Inquiry with Student Teachers to Support Teacher Professional Development Sponsored by Teachers for.
Middle School Liaison Meeting
P.E.R.T. Diagnostic Learning Pathways Math, Reading, Writing.
Terry Vendlinski Geneva Haertel SRI International
March16, To keep up with the current and future standards of high school graduates. To align with college and career readiness standards. To ensure.
ECD in the Scenario-Based GED ® Science Test Kevin McCarthy Dennis Fulkerson Science Content Specialists CCSSO NCSA June 29, 2012 Minneapolis This material.
Yosemite Mathematics Unit 1 Overview. Yosemite Not just a great valley, but a shrine to human foresight, the strength of granite, the power of glaciers,
Educator Effectiveness Academy STEM Follow-Up Webinar December 2011.
September ERD 5 th Grade. Welcome to Harmony! Norms House keeping.
WELCOME TO PARK VIEW MIDDLE SCHOOL NECAP REPORT NIGHT.
 Connecting to the Common Core State Standards for Mathematics- Assessment Tools for Student Success Anton Jackson, OSPI Sandy Christie, Puget Sound ESD.
Protocols for Mathematics Performance Tasks PD Protocol: Preparing for the Performance Task Classroom Protocol: Scaffolding Performance Tasks PD Protocol:
Assessment Practices That Lead to Student Learning Core Academy, Summer 2012.
1 Issues in Assessment in Higher Education: Science Higher Education Forum on Scientific Competencies Medellin-Colombia Nov 2-4, 2005 Dr Hans Wagemaker.
What is the TPA? Teacher candidates must show through a work sample that they have the knowledge, skills, and abilities required of a beginning teacher.
The present publication was developed under grant X from the U.S. Department of Education, Office of Special Education Programs. The views.
Committee on the Assessment of K-12 Science Proficiency Board on Testing and Assessment and Board on Science Education National Academy of Sciences.
What is it we expect students to learn? Curriculum learning tasks need to be clearly stated 1.What policies and practices, regarding curriculum development,
Overview Criteria Item Examples.  Create a sense of “measured urgency” ◦ Timeline is not immediate, but close ◦ Urgency in you and your teachers ◦ Difference.
Computer Control Lou Loftin FETC Conference Orlando, FL January 28 – 31, 2014.
The PLC Team Learning Process Review Step One: Identify essential (key) learning standards that all students must learn in each content area during each.
Assessing Local Item Dependence in Building Explanation Tasks Han Bao & Dr. Robert J. Mislevy University of Maryland, College Park Han Bao & Dr. Robert.
© 2015 The College Board The Redesigned SAT/PSAT Key Changes.
Welcome Principals Please sit in groups of 3 2 All students graduate college and career ready Standards set expectations on path to college and career.
INQUIRY AND CRITICAL THINKING What is inquiry??? Simply put…. CRITICAL THINKING….
Alaska Measures of Progress (AMP) Summative Assessment Framework 1.
What’s Needed for this Session Handouts: 1.Performance Task s 2.Task writing template.
Dorthea Litson April 30, Purposes of Assessment Purposes of Assessment Making instructional decisions Monitoring student progress Evaluating programs.
Castle / Kahuku Complex Area Support Team (CK CAST) Phase I Professional Development
ACTion for Mathematics-ASPIRE. Background The math assessment was developed to reflect students’ knowledge and skill accumulation over time; capturing.
Performance Task and the common core. Analysis sheet Phases of problem sheet Performance task sheet.
Using PADI Templates as an Alternative Structure for Specifying GLOBE Investigation Strategies AERA April 2005 Angela Haydel DeBarger, SRI International.
LEAP TH GRADE. DATES: APRIL 25-29, 2016 Test Administration Schedule:  Day 1 April 25- ELA Session 1: Research Simulation Task (90mins) Mathematics.
Assessing Math Looking closer at PARCC Task Types 2.
Performance Assessment: The Core of Competency Based Learning Rose ColbyIowa ASCDJune, 2014.
Next Generation Iowa Assessments.  Overview of the Iowa Assessments ◦ Purpose of Forms E/F and Alignment Considerations ◦ Next Generation Iowa Assessments.
Linking Curriculum, Instruction and Assessment in Mathematics March 10, 2016 CONNECTICUT STATE DEPARTMENT OF EDUCATION.
High School Biology EOC Planning and Updates December 4, 2012  Welcome- Networking and Coffee  Door Prizes  Contest  Entry Task.
Next Generation Iowa Assessments
Inquiry learning and SimQuest
Using Cognitive Science To Inform Instructional Design
STATISTICAL TOOLS FOR AUDITING
8th Grade Mathematics Curriculum
Assessing Math Looking closer at PARCC Task Types.
Lindsay Ruhter, Lori Andersen, & Allison Lawrence
Increasing Rigor to Develop Critical Thinking Skills
Mike Timms and Cathleen Kennedy University of California, Berkeley
Common Ground: Addressing Known Challenges with Alignment for CATs
Principled Assessment Designs for Inquiry (PADI)
Innovative Approaches for Examining Alignment
Assessment In Mathematics
Assessing Understanding
2008 Multiple Choice.
Presentation transcript:

AERA April 2005 Models and Tools for Drawing Inferences from Student Work: The BEAR Scoring Engine Cathleen Kennedy & Mark Wilson University of California, Berkeley

2 © BEAR Center, 2005 Overview  Features of “complex” tasks.  How PADI addresses complex task features.  The “big” assessment picture and where inferences are drawn (in measurement models implemented in Scoring Engine).  An example of the PADI and Scoring Engine views.  Next steps: Wizard to guide designers in developing solid chain of reasoning.

3 © BEAR Center, 2005 Example from FOSS

4 © BEAR Center, 2005 Example from FOSS Two measures: Physics (speed) Mathematics 5 responses: Equation choice Fill-in Numbers Fill-in Units Calculate Units in Answer Are the responses dependent?

5 © BEAR Center, 2005 “Complex” Task Features  Multiple measures of interest  Content & inquiry  Multiple aspects of inquiry  Response dependencies  Common stimulus  Sequence of steps

6 © BEAR Center, 2005 Complex Measurement Requires Clear Chain of Reasoning 1.Inferences one wishes to draw (cognition vertex) 2.Evidence required to draw the inferences (interpretation vertex) 3.Observations required to generate evidence (observations vertex) 4.Inferences are then interpretable in the context of the purpose of the assessment

7 © BEAR Center, 2005 PADI Addresses “Complex” Task Features  Multiple measures of interest  Content & inquiry  Multiple aspects of inquiry  Response dependencies  Common stimulus  Sequence of steps (within task) PADI Approach: Multidimensional IRT measurement model. Well-defined evaluation phases model response dependencies (rather than ignoring them).

8 © BEAR Center, 2005 Assessment System Architecture Task Specifications Student Database Scoring Engine Delivery System Design System Students Design Team DesignImplementation

9 © BEAR Center, 2005 Chain of Inferential Reasoning Task Specifications Student Database Scoring Engine Delivery System Design System Students Design Team Assessment Purpose Assessment Evidence Inferences about what students know and can do

10 © BEAR Center, 2005 FOSS Example: PADI View Two student model variables: Physics (speed) Mathematics 6 observable variables: Equation choice (Physics) Fill-in Numbers (Physics) Fill-in Units (Physics) Calculate (Math) Units in Answer (Physics) Bundled physics items

11 © BEAR Center, 2005 FOSS Example : Bundling Rules  Defined in PADI Design System  Template (task specification) − Activity − Evaluation Procedure − Evaluation Phase  Implemented in Delivery System

12 © BEAR Center, 2005 FOSS Example : Bundling Rules EquationEq. Fill InEq. UnitsAns. UnitsBundle

13 © BEAR Center, 2005 FOSS Example : Scoring Engine View Two student model variables: Physics (speed) Mathematics 2 observable variables: Math score Bundled Physics score “Between Item” MD: Each observable variable provides evidence of one SMV. Scoring Engine returns two proficiency estimates per student to the Assessment Delivery System.

14 © BEAR Center, 2005 Next Steps Develop Measurement Model Design Wizard  Evaluate design needs of users (how do they do it now, what would work better?)  Guide thinking from the “assessment purpose” standpoint  Align inferences, evidence and observations