University of Maryland Slide 1 July 6, 2005 Presented at Invited Symposium K3, “Assessment Engineering: An Emerging Discipline” at the annual meeting of.

Slides:



Advertisements
Similar presentations
OSSE CSSS Educator Leader Institute Secondary English Language Arts July 31 st to Aug 3 rd, 2012 Day 1 Facilitated by Heidi Beeman.
Advertisements

Assessing Computational Thinking
© 2013 SRI International - Company Confidential and Proprietary Information Center for Technology in Learning SRI International NSF Showcase 2014 SIGCSE.
Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI International Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI.
Object-Oriented Analysis and Design
SRI Technology Evaluation WorkshopSlide 1RJM 2/23/00 Leverage Points for Improving Educational Assessment Robert J. Mislevy, Linda S. Steinberg, and Russell.
University of Maryland Slide 1 May 2, 2001 ECD as KR * Robert J. Mislevy, University of Maryland Roy Levy, University of Maryland Eric G. Hansen, Educational.
SLRF 2010 Slide 1 Oct 16, 2010 What is the construct in task-based language assessment? Robert J. Mislevy Professor, Measurement, Statistics and Evaluation.
Reading and Writing in Science: Multiple Text Types in Scientific Literacy Learning Elizabeth Birr Moje LeeAnn M. Sutherland Tanya Cleveland Mary Heitzman.
ECOLT 2006 Slide 1 October 13, 2006 Prospectus for the PADI design framework in language testing ECOLT 2006, October 13, 2006, Washington, D.C. PADI is.
U Iowa Slide 1 Sept 19, 2007 Some Terminology and Concepts for Simulation-Based Assessment Robert J. Mislevy University of Maryland In collaboration with.
FERA 2001 Slide 1 November 6, 2001 Making Sense of Data from Complex Assessments Robert J. Mislevy University of Maryland Linda S. Steinberg & Russell.
Next Generation Science Standards
Computational Thinking Related Efforts. CS Principles – Big Ideas  Computing is a creative human activity that engenders innovation and promotes exploration.
Systems Engineer An engineer who specializes in the implementation of production systems This material is based upon work supported by the National Science.
ADL Slide 1 December 15, 2009 Evidence-Centered Design and Cisco’s Packet Tracer Simulation-Based Assessment Robert J. Mislevy Professor, Measurement &
1 Welcome back!. Vision for Science Teaching and Learning 2 View free PDF from The National Academies Press at *Will also be posted.
Basic Concepts The Unified Modeling Language (UML) SYSC System Analysis and Design.
Principled Assessments of 21st Century Skills across Disciplines in a Community College Curriculum Louise Yarnall SRI International Jane Ostrander Foothill-DeAnza.
Seeing the Destination So We Can Direct Others to It
Can we make a test fun? Eric Church – BreakAway Games/ University of Baltimore
Crosscutting Concepts and Disciplinary Core Ideas February24, 2012 Heidi Schweingruber Deputy Director, Board on Science Education, NRC/NAS.
Supporting the CCSS in the Science Classroom through the Science and Engineering Practices of the Next Generation Science Standards (NGSS) John Spiegel.
The Use of Student Work as a Context for Promoting Student Understanding and Reasoning Yvonne Grant Portland MI Public Schools Michigan State University.
Terry Vendlinski Geneva Haertel SRI International
ECD in the Scenario-Based GED ® Science Test Kevin McCarthy Dennis Fulkerson Science Content Specialists CCSSO NCSA June 29, 2012 Minneapolis This material.
The Four Strands of Scientific Proficiency Students who understand science:  Know, use, and interpret scientific explanations of the natural world  Generate.
The Design Phase: Using Evidence-Centered Assessment Design Monty Python argument.
Engineering Design By Brian Nettleton This material is based upon work supported by the National Science Foundation under Grant No Any opinions,
Some Implications of Expertise Research for Educational Assessment Robert J. Mislevy University of Maryland National Center for Research on Evaluation,
ITEC224 Database Programming
January 29, 2010ART Beach Retreat ART Beach Retreat 2010 Assessment Rubric for Critical Thinking First Scoring Session Summary ART Beach Retreat.
9/14/2012ISC329 Isabelle Bichindaritz1 Database System Life Cycle.
LIVE INTERACTIVE YOUR DESKTOP 1 Start recording—title slide—1 of 3 Introducing the Next Generation Science Standards Originally presented by:
The present publication was developed under grant X from the U.S. Department of Education, Office of Special Education Programs. The views.
Robert J. Mislevy University of Maryland Geneva Haertel & Britte Haugan Cheng SRI International Robert J. Mislevy University of Maryland Geneva Haertel.
The Next Generation Science Standards From Awareness to Transition
Putting Research to Work in K-8 Science Classrooms Ready, Set, SCIENCE.
REVISIONS TO GENERAL EDUCATION STUDENT LEARNING OUTCOMES Auburn University Senate Information Item, August 2014.
Maria E. Fernandez, Ph.D. Associate Professor Health Promotion and Behavioral Sciences University of Texas, School of Public Health.
Sharing Design Knowledge through the IMS Learning Design Specification Dawn Howard-Rose Kevin Harrigan David Bean University of Waterloo McGraw-Hill Ryerson.
On Layers and Objects in Assessment Design Robert Mislevy, University of Maryland Michelle Riconscente, University of Maryland Robert Mislevy, University.
What does a Framework for k-12 Science Education have to do with PER?
Differentiation PLC.
Shaping a Health Statistics Vision for the 21 st Century 2002 NCHS Data Users Conference 16 July 2002 Daniel J. Friedman, PhD Massachusetts Department.
NGSS-Health Science August Connection to the Common Core.
Based on the work of Dr. M.S. Smith, University of Pgh. Key Ingredients to Developing Mathematical Understanding: Anticipating, Monitoring, Selecting,
Theme 2: Data & Models One of the central processes of science is the interplay between models and data Data informs model generation and selection Models.
1 Assessing Student Understanding David Niemi UCLA Graduate School of Education & Information Studies National Center for Research on Evaluation, Standards,
Evidence-Centered Game Design Kristen DiCerbo, Ph.D. Principal Research Scientist, Pearson Learning Games Scientist, GlassLab.
CSE 303 – Software Design and Architecture
Invention and Innovation Unit #4 Mrs. Mattson Carl Ben Eielson Middle School ETP 2006—Tanya Mattson This material is based upon work supported by the National.
S.P.S. - HMS May 19, New 2011 ELA & Literacy and Math Curriculum Frameworks  comprised primarily of the Common Core State Standards  written explicitly.
National Research Council Of the National Academies
Robert J. Mislevy University of Maryland National Center for Research on Evaluation, Standards, and Student Testing (CRESST) NCME San Diego, CA April 15,
CALIFORNIA DEPARTMENT OF EDUCATION Tom Torlakson, State Superintendent of Public Instruction 2014 ELA/ELD Framework A Focus on Making Meaning.
AERA April 2005 Models and Tools for Drawing Inferences from Student Work: The BEAR Scoring Engine Cathleen Kennedy & Mark Wilson University of California,
Using the PADI Design System to Examine the Features of a NAEP Performance Assessment Kathleen C. Haynie Kathleen Haynie Consulting Andrea A. Lash, Geneva.
Using PADI Templates as an Alternative Structure for Specifying GLOBE Investigation Strategies AERA April 2005 Angela Haydel DeBarger, SRI International.
Teaching with CHRONOS Data and Tools A Framework for Design Cathy Manduca Science Education Resource Center Carleton College June 13, 2006.
Introducing Critical and Creative Thinking. Agenda The importance of Critical and Creative Thinking What is in the curriculum? Questions Planning for.
Model-Facilitated Learning Overview Gordon Graber 2008.
Principles of Evidence-Centered Design: Practical Considerations for State Assessment Development Efforts Carole Gallagher Director of Research.
Assessment Directors WebEx March 29, :00-3:00 pm ET
Mike Timms and Cathleen Kennedy University of California, Berkeley
Principled Assessment Designs for Inquiry (PADI)
This material is based upon work supported by the National Science Foundation under Grant #XXXXXX. Any opinions, findings, and conclusions or recommendations.
Serena Alderson Manager of School Partnerships Carnegie Learning, Inc
Principles of Evidence-Centered Design: Practical Considerations for State Assessment Development Efforts Carole Gallagher Director of Research.
Presentation transcript:

University of Maryland Slide 1 July 6, 2005 Presented at Invited Symposium K3, “Assessment Engineering: An Emerging Discipline” at the annual meeting of the National Council on Measurement in Education, Chicago, IL, April 10-12, 2007 PADI is supported by the Interagency Educational Research Initiative (IERI) under grants REC (PADI Planning Grant)and REC (PADI Implementation Grant).Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation. Implications of Evidence-Centered Design for Educational Testing: Lessons from the PADI Project Robert J. Mislevy University of Maryland Geneva D. Haertel SRI International April 12, 2007

University of Maryland Slide 2 July 6, 2005 Outline l Remarks on assessment engineering l Evidence-centered assessment design  Assessment arguments  Layers l Implications  Explicitness, representations, reusability, generativity l PADI l Too many notes?

University of Maryland Slide 3 July 6, 2005 Assessment Engineering l Is assessment design art or science? l Assessment engineering submits there are recurring structures and relationships we can exploit – in terms of language, data structures, representational forms, and processes – to enhance efficiency and validity in assessment/task design.

University of Maryland Slide 4 July 6, 2005 Assessment Engineering l Idea not new or exclusive; some examples:  Guttman, Suppes, Hively et al., Obsbourne, Bormuth ~ 60s.  Roid & Haladyna (1982): Technology for Test-Item Writing  Embretson (1985): Test design: Developments in psychology and psychometrics.  Embretson, Wilson, Baker, Luecht, Gorin, Tatsuoka, Bejar, Shavelson, Irvine & Kyllonen… l Evidence-centered design (ECD)  Mislevy, Steinberg, & Almond (2003)  PADI project

University of Maryland Slide 5 July 6, 2005 Assessment Arguments What complex of knowledge, skills, or other attributes should be assessed? What behaviors or performances should reveal those constructs? What tasks or situations should elicit those behaviors? (Messick, 1994)

From Mislevy & Riconscente, in press Assessment Delivery How do students and tasks actually interact? How do we report examinee performance? How do students and tasks actually interact? How do we report examinee performance? Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Design structures: Student, evidence, and task models How do we choose and present tasks, and gather and analyze responses? Layers in the assessment enterprise

From Mislevy & Riconscente, in press Assessment Delivery How do students and tasks actually interact? How do we report examinee performance? How do students and tasks actually interact? How do we report examinee performance? Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Design structures: Student, evidence, and task models How do we choose and present tasks, and gather and analyze responses? Cog Psych, expertise studies, domain research Cog Psych, expertise studies, domain research

From Mislevy & Riconscente, in press Assessment Delivery How do students and tasks actually interact? How do we report examinee performance? How do students and tasks actually interact? How do we report examinee performance? Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Design structures: Student, evidence, and task models How do we choose and present tasks, and gather and analyze responses? Specific implementations. Surface elements. Specific implementations. Surface elements.

From Mislevy & Riconscente, in press Assessment Delivery How do students and tasks actually interact? How do we report examinee performance? How do students and tasks actually interact? How do we report examinee performance? Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Design structures: Student, evidence, and task models How do we choose and present tasks, and gather and analyze responses? Assessment Argument

From Mislevy & Riconscente, in press Assessment Delivery How do students and tasks actually interact? How do we report examinee performance? How do students and tasks actually interact? How do we report examinee performance? Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Design structures: Student, evidence, and task models How do we choose and present tasks, and gather and analyze responses? Explicit connection to domain & purpose Generative structures for recurring kinds of proficiencies (e.g., inquiry cycles, troubleshooting) across projects PADI Design patterns

From Mislevy & Riconscente, in press Assessment Delivery How do students and tasks actually interact? How do we report examinee performance? How do students and tasks actually interact? How do we report examinee performance? Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Design structures: Student, evidence, and task models How do we choose and present tasks, and gather and analyze responses? Generative Design Schemas

From Mislevy & Riconscente, in press Assessment Delivery How do students and tasks actually interact? How do we report examinee performance? How do students and tasks actually interact? How do we report examinee performance? Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Design structures: Student, evidence, and task models How do we choose and present tasks, and gather and analyze responses? Explicit connection of argument to machinery Generative structures for recurring task situations (e.g., item shells) Re-usable / interoperable data structures PADI templates

University of Maryland Slide 13 July 6, 2005 Supported by NSF, via IERI program Focus on scaling up ECD ideas specifically focusing on science inquiry tasks Partners: »SRI International (Geneva, co-PI) »University of Maryland (Mislevy, co-PI) »University of California at Berkeley (BEAR – Wilson – & FOSS –Long ) »University of Michigan (BioKIDS – Songer) PADI: Principled Assessment Design for Inquiry

University of Maryland Slide 14 July 6, 2005 PADI Conceptual frameworks & representations »Domain Modeling: Design patterns for inquiry »CAF: Task-design (templates) Object model for task design »Data structure Software tools to aid the design process »Design system »Wizards

University of Maryland Slide 15 July 6, 2005 PADI Libraries of exemplars … »Design patterns for assessing inquiry »Task templates and pointers to tasks “Scoring engine” (MRCMLM) Worked-through applications »FOSS, BioKids; GLOBE, FCI, Mystery Powders, Mystery Boxes Evaluation studies – FOSS & BioKids

University of Maryland Slide 16 July 6, 2005 PADI For more information: padi.sri.com/ Includes lots of tech reports.

University of Maryland Slide 17 July 6, 2005 Too many notes? As Emperor Joseph II said to Mozart. Power in general/abstract representations, »Reveals fundamental similaries under assessments that look different on the surface. »Right perspective for design of systems, but not necessarily for work within systems. Some people need to work from first principles, but… Scaling up for wide range of users requires exemplars, support tools, tuned interfaces.