Terry Vendlinski Geneva Haertel SRI International

Slides:



Advertisements
Similar presentations
Victorian Curriculum and Assessment Authority
Advertisements

Level 1 Recall Recall of a fact, information, or procedure. Level 2 Skill/Concept Use information or conceptual knowledge, two or more steps, etc. Level.
Computer-Based Performance Assessments from NAEP and ETS and their relationship to the NGSS Aaron Rogat Educational Testing Service.
TAILS: COBWEB 1 [1] Online Digital Learning Environment for Conceptual Clustering This material is based upon work supported by the National Science Foundation.
Software Modeling SWE5441 Lecture 3 Eng. Mohammed Timraz
Aligning to standards from the "get go:" Designing alternate assessments based on states’ standards, expanded benchmarks, and universal design Sue Bechard,
ARCH-01: Introduction to the OpenEdge™ Reference Architecture Don Sorcinelli Applied Technology Group.
Marzano Art and Science Teaching Framework Learning Map
Assessing Computational Thinking
© 2013 SRI International - Company Confidential and Proprietary Information Center for Technology in Learning SRI International NSF Showcase 2014 SIGCSE.
Show Me an Evidential Approach to Assessment Design Michael Rosenfeld F. Jay Breyer David M. Williamson Barbara Showers.
Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI International Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI.
SRI Technology Evaluation WorkshopSlide 1RJM 2/23/00 Leverage Points for Improving Educational Assessment Robert J. Mislevy, Linda S. Steinberg, and Russell.
NCTM’s Focus in High School Mathematics: Reasoning and Sense Making.
Knowledge Acquisitioning. Definition The transfer and transformation of potential problem solving expertise from some knowledge source to a program.
MCAS-Alt: Alternate Assessment in Massachusetts Technical Challenges and Approaches to Validity Daniel J. Wiener, Administrator of Inclusive Assessment.
University of Maryland Slide 1 May 2, 2001 ECD as KR * Robert J. Mislevy, University of Maryland Roy Levy, University of Maryland Eric G. Hansen, Educational.
University of Maryland Slide 1 July 6, 2005 Presented at Invited Symposium K3, “Assessment Engineering: An Emerging Discipline” at the annual meeting of.
ECOLT 2006 Slide 1 October 13, 2006 Prospectus for the PADI design framework in language testing ECOLT 2006, October 13, 2006, Washington, D.C. PADI is.
U Iowa Slide 1 Sept 19, 2007 Some Terminology and Concepts for Simulation-Based Assessment Robert J. Mislevy University of Maryland In collaboration with.
Teaching with Depth An Understanding of Webb’s Depth of Knowledge
FERA 2001 Slide 1 November 6, 2001 Making Sense of Data from Complex Assessments Robert J. Mislevy University of Maryland Linda S. Steinberg & Russell.
The Software Product Life Cycle. Views of the Software Product Life Cycle  Management  Software engineering  Engineering design  Architectural design.
Foundations This chapter lays down the fundamental ideas and choices on which our approach is based. First, it identifies the needs of architects in the.
1 © 2003 Cisco Systems, Inc. All rights reserved. PTC 6/18/04 What Cisco’s ADI Group is Doing in Performance Testing PTC: June 18, 2004.
1 New York State Mathematics Core Curriculum 2005.
ADL Slide 1 December 15, 2009 Evidence-Centered Design and Cisco’s Packet Tracer Simulation-Based Assessment Robert J. Mislevy Professor, Measurement &
Principled Assessments of 21st Century Skills across Disciplines in a Community College Curriculum Louise Yarnall SRI International Jane Ostrander Foothill-DeAnza.
Can we make a test fun? Eric Church – BreakAway Games/ University of Baltimore
INTRODUCTION TO THE MINISTRY OF EDUCATION’s ACHIEVEMENT CHART Bedford Park PS September 2013.
COPYRIGHT WESTED, 2010 Calipers II: Using Simulations to Assess Complex Science Learning Diagnostic Assessments Panel DRK-12 PI Meeting - Dec 1–3, 2010.
The Use of Student Work as a Context for Promoting Student Understanding and Reasoning Yvonne Grant Portland MI Public Schools Michigan State University.
ECD in the Scenario-Based GED ® Science Test Kevin McCarthy Dennis Fulkerson Science Content Specialists CCSSO NCSA June 29, 2012 Minneapolis This material.
Nicole Paulson CCSSO Webinar March 21, 2012 Transition to the Common Core State Standards in Elementary Math.
The Design Phase: Using Evidence-Centered Assessment Design Monty Python argument.
Some Implications of Expertise Research for Educational Assessment Robert J. Mislevy University of Maryland National Center for Research on Evaluation,
Learning Progressions: Some Thoughts About What we do With and About Them Jim Pellegrino University of Illinois at Chicago.
1 Issues in Assessment in Higher Education: Science Higher Education Forum on Scientific Competencies Medellin-Colombia Nov 2-4, 2005 Dr Hans Wagemaker.
The present publication was developed under grant X from the U.S. Department of Education, Office of Special Education Programs. The views.
UCLA Graduate School of Education & Information Studies National Center for Research on Evaluation, Standards, and Student Testing Online Assessment within.
Metadata Models in Survey Computing Some Results of MetaNet – WG 2 METIS 2004, Geneva W. Grossmann University of Vienna.
ATTRIBUTEDESCRIPTION Focal Knowledge, Skills, Abilities The primary knowledge / skills / abilities (KSAs) targeted by this design pattern. RationaleHow/why.
Robert J. Mislevy University of Maryland Geneva Haertel & Britte Haugan Cheng SRI International Robert J. Mislevy University of Maryland Geneva Haertel.
Illustration of a Validity Argument for Two Alternate Assessment Approaches Presentation at the OSEP Project Directors’ Conference Steve Ferrara American.
On Layers and Objects in Assessment Design Robert Mislevy, University of Maryland Michelle Riconscente, University of Maryland Robert Mislevy, University.
Student assessment Assessment tools AH Mehrparvar,MD Occupational Medicine department Yazd University of Medical Sciences.
Object-Oriented Software Engineering using Java, Patterns &UML. Presented by: E.S. Mbokane Department of System Development Faculty of ICT Tshwane University.
Algebraic Reasoning Institute Math & Science Collaborative at the Allegheny Intermediate Unit.
National Council of Teachers of Mathematics Principles and Standards for grades pre-K-2.
Based on the work of Dr. M.S. Smith, University of Pgh. Key Ingredients to Developing Mathematical Understanding: Anticipating, Monitoring, Selecting,
Resources and Reflections: Using Data in Undergraduate Geosciences Cathy Manduca SERC Carleton College DLESE Annual Meeting 2003.
Backwards Design STEP 1: Agree on the overarching science concept of the 5E sequence STEP 2: Design the Evaluate and decide what evidence students will.
1 Assessing Student Understanding David Niemi UCLA Graduate School of Education & Information Studies National Center for Research on Evaluation, Standards,
Welcome Principals Please sit in groups of 3 2 All students graduate college and career ready Standards set expectations on path to college and career.
Robert J. Mislevy University of Maryland National Center for Research on Evaluation, Standards, and Student Testing (CRESST) NCME San Diego, CA April 15,
SOFTWARE ENGINEERING. Objectives Have a basic understanding of the origins of Software development, in particular the problems faced in the Software Crisis.
AERA April 2005 Models and Tools for Drawing Inferences from Student Work: The BEAR Scoring Engine Cathleen Kennedy & Mark Wilson University of California,
Using the PADI Design System to Examine the Features of a NAEP Performance Assessment Kathleen C. Haynie Kathleen Haynie Consulting Andrea A. Lash, Geneva.
CTB CADDS Sally Valenzuela Director, Publishing Strategic Initiatives CTB/McGraw-Hill.
Using Evidence-Centered Design to develop Scenario Based Interactive Computer Tasks Daisy Rutstein.
Using PADI Templates as an Alternative Structure for Specifying GLOBE Investigation Strategies AERA April 2005 Angela Haydel DeBarger, SRI International.
Teaching with CHRONOS Data and Tools A Framework for Design Cathy Manduca Science Education Resource Center Carleton College June 13, 2006.
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
Knowing What Students Know Ganesh Padmanabhan 2/19/2004.
Center for Assessment and Improvement of Learning
Principles of Evidence-Centered Design: Practical Considerations for State Assessment Development Efforts Carole Gallagher Director of Research.
Mike Timms and Cathleen Kennedy University of California, Berkeley
Principled Assessment Designs for Inquiry (PADI)
Principles of Evidence-Centered Design: Practical Considerations for State Assessment Development Efforts Carole Gallagher Director of Research.
Presentation transcript:

Terry Vendlinski Geneva Haertel SRI International The Use of a Co-Design Process in ECD to Support the Development of Large-Scale Assessments Terry Vendlinski Geneva Haertel SRI International CCSSO’s National Conference on Student Assessment Minneapolis, MN June 29, 2012

PADI Special Ed AERA 2009 Acknowledgements 4/21/2017 Research findings and assessment tasks described in this presentation were supported by the following projects: Principled Assessment Design in Inquiry [National Science Foundation, REC-0089122 and REC-0129331]; An Application of Evidence-Centered Design to a State’s Large Scale Science Assessment [National Science Foundation, DRK-12 initiative, DRL-0733172];Principled Assessment Science Assessment Designs for Students with Disabilities [Institute of Education Sciences, US Department of Education, R324A070035]; Applying Evidence-Centered Design to Alternate Assessments in Mathematics for Students with Significant Cognitive Disabilities [US Department of Education, Contract to State of Utah, 09679];Alternate Assessment Design Reading (AAD-R): Evidence-Centered Design for Alternate Assessment [US Department of Education, Contract to State of Idaho, S368A090032]. In addition, SRI International provided Strategic Business Thrust (SBT) funds. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the funding agencies.

Geneva Haertel – SRI International Robert Mislevy – ETS PADI Special Ed AERA 2009 Co-Design Process 4/21/2017 Geneva Haertel – SRI International Robert Mislevy – ETS Britte Cheng – SRI International Angela DeBarger – SRI International Daisy Rutstein – SRI International Terry Vendlinski – SRI International

Evidence-Centered Assessment Design Mislevy, Steinberg, & Almond at ETS in late 1990s Cisco / ETS / University of Maryland Principled Assessment Design in Inquiry (PADI) project SRI, University of Maryland, UC Berkeley, FOSS, BioKIDS National Science Foundation ECD for Large-Scale State Assessments SRI, Pearson, University of Maryland, Haney Research & Evaluation, GED Assessment Developers

Evidence-Centered Assessment Design Formal, multiple-layered framework from Messick’s (1994) guiding questions: What complex of knowledge, skills, or other attributes should be assessed? What behaviors or performances should reveal those constructs? What tasks or situations should elicit those behaviors?

What is an ECD approach? A process by which evidence is gathered. PADI Special Ed AERA 2009 What is an ECD approach? 4/21/2017 A process by which evidence is gathered. Uses the framework to document information that supports the validity argument Documents what decisions have been made with regards to the assessment and the justification for those decisions.

Co-Design in the Context of ECD PADI Special Ed AERA 2009 Co-Design in the Context of ECD 4/21/2017 What is Co-Design? What sorts of expertise are required? What are the processes that might occur?

What is important about this domain? Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? Domain Modeling How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Conceptual Assessment Framework Design structures: Student, evidence, and task models. Generativity. Assessment Implementation Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability. Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four-process delivery architecture. From Mislevy & Riconscente, 2006

PADI Special Ed AERA 2009 Domain Analysis 4/21/2017 Gather substantive information about the domain of interest that has implications for assessment; how knowledge is constructed, acquired, used, communicated. Domain concepts, terminology, tools, knowledge representations, research findings, situations of use (heads up display), patterns of interaction. Representational forms and symbol systems used in domain (e.g., algebraic notation, Punnett squares, maps, computer program interfaces, content standards, concept maps). Could take days or weeks (two-hour blocks)

PADI Special Ed AERA 2009 Domain Modeling 4/21/2017 Express assessment argument in narrative form based on information from Domain Analysis. Specifications of knowledge, skills, or other attributes to be assessed; features of situations that can evoke evidence; kinds of performances that convey evidence. Design patterns; “big ideas”, Toulmin and Wigmore diagrams for assessment arguments; assessment blueprints, ontologies, generic rubrics. Could take from an hour to a day (one to two hour blocks)

PADI Special Ed AERA 2009 Design Pattern 4/21/2017

Design Pattern Attributes Focal Knowledge, Skills & Attributes (KSAs) The primary KSAs targeted by the design pattern. What we want to make inferences about. Additional KSAs Other KSAs that may be required for successful performance on the assessment tasks. Potential Observations Features of the things students say, do, or make. Potential Work Products Some possible things one could see students doing that would give evidence about the KSAs. 12

Design Pattern Attributes Characteristic Features Aspects of assessment situations that are likely to evoke the desired evidence. Variable Features Aspects of assessment situations that can be varied in order to shift difficulty or emphasis. 13

Conceptual Assessment Framework PADI Special Ed AERA 2009 Conceptual Assessment Framework 4/21/2017 Express assessment argument in structures and specifications for tasks and tests, evaluation procedures, measurement models. Student, evidence, and task models; student, observable, and task variables; rubrics; measurement models; test assembly specifications; task templates and task specifications. Algebraic and graphical representations of measurement models; task templates and task specifications; item generation models; generic rubrics; algorithms for automated scoring. Can take from days to weeks

PADI Special Ed AERA 2009 4/21/2017 Visual CAF

PADI Special Ed AERA 2009 Task Model Template 4/21/2017

Assessment Implementation Implement assessment, including presentation-ready tasks and calibrated measurement models Item writing and task materials (including all materials, tools, affordances); pilot test data to hone evaluation procedures and fit measurement models. Coded algorithms for rendering tasks, interacting with examinees and evaluating work products; tasks as displayed; IMS/QTI/APIP representation of materials; ASCII files of item parameters. Time required varies according to number and complexity of items and tasks.

PADI Special Ed AERA 2009 Assessment Delivery 4/21/2017 Coordinate interactions of students and tasks: task-and test-level scoring; reporting. Tasks as presented; work products as created; scores as evaluated. Renderings of materials; numerical and graphical summaries for individual and groups; specifications for results files.

Why Co-design? Co-Design can improve at any / all the ECD layers. Not all layers are required. Co-design may be most powerful at top three layers. Can be complex … so requires structure May take more time … and produce better products.

More Information Visit us: padi.sri.com Email us: Geneva.Haertel@sri.com Terry.Vendlinski@sri.com