U Iowa Slide 1 Sept 19, 2007 Some Terminology and Concepts for Simulation-Based Assessment Robert J. Mislevy University of Maryland In collaboration with.

Slides:



Advertisements
Similar presentations
Analyzing Student Work
Advertisements

MMAP Middle School Math Through Applications Project Dahwun Deepak Gazi Scott Sun-Young.
Advances in the PARCC Mathematics Assessment August
The Network of Dynamic Learning Communities C 107 F N Increasing Rigor February 5, 2011.
Yvan Rooseleer – BiASC – MAY 2013
© 2013 SRI International - Company Confidential and Proprietary Information Center for Technology in Learning SRI International NSF Showcase 2014 SIGCSE.
Show Me an Evidential Approach to Assessment Design Michael Rosenfeld F. Jay Breyer David M. Williamson Barbara Showers.
Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI International Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI.
SRI Technology Evaluation WorkshopSlide 1RJM 2/23/00 Leverage Points for Improving Educational Assessment Robert J. Mislevy, Linda S. Steinberg, and Russell.
NCTM’s Focus in High School Mathematics: Reasoning and Sense Making.
University of Maryland Slide 1 May 2, 2001 ECD as KR * Robert J. Mislevy, University of Maryland Roy Levy, University of Maryland Eric G. Hansen, Educational.
University of Maryland Slide 1 July 6, 2005 Presented at Invited Symposium K3, “Assessment Engineering: An Emerging Discipline” at the annual meeting of.
Overview: Competency-Based Education & Evaluation
1 TALK 5 Anita Pincas Course Design 2 – PART A The pedagogic media - continued.
ECOLT 2006 Slide 1 October 13, 2006 Prospectus for the PADI design framework in language testing ECOLT 2006, October 13, 2006, Washington, D.C. PADI is.
Developing Ideas for Research and Evaluating Theories of Behavior
FERA 2001 Slide 1 November 6, 2001 Making Sense of Data from Complex Assessments Robert J. Mislevy University of Maryland Linda S. Steinberg & Russell.
The Software Product Life Cycle. Views of the Software Product Life Cycle  Management  Software engineering  Engineering design  Architectural design.
ED 3501: Curriculum and Instruction Section GHI - Fall Understanding by Design Understanding and Creating Effective Instructional Design.
1 © 2003 Cisco Systems, Inc. All rights reserved. PTC 6/18/04 What Cisco’s ADI Group is Doing in Performance Testing PTC: June 18, 2004.
The Modeling Method of Physics Teaching Taken from the MM Web Site.
ADL Slide 1 December 15, 2009 Evidence-Centered Design and Cisco’s Packet Tracer Simulation-Based Assessment Robert J. Mislevy Professor, Measurement &
Can we make a test fun? Eric Church – BreakAway Games/ University of Baltimore
Supporting the CCSS in the Science Classroom through the Science and Engineering Practices of the Next Generation Science Standards (NGSS) John Spiegel.
COPYRIGHT WESTED, 2010 Calipers II: Using Simulations to Assess Complex Science Learning Diagnostic Assessments Panel DRK-12 PI Meeting - Dec 1–3, 2010.
Terry Vendlinski Geneva Haertel SRI International
ECD in the Scenario-Based GED ® Science Test Kevin McCarthy Dennis Fulkerson Science Content Specialists CCSSO NCSA June 29, 2012 Minneapolis This material.
Interstate New Teacher Assessment and Support Consortium (INTASC)
QUALITATIVE MODELING IN EDUCATION Bert Bredweg and Ken Forbus Yeşim İmamoğlu.
Central concepts:  Assessment can measure habits of mind or habits of recall.  Tests have their limits.  It is important to know the purpose the test.
A Framework for Inquiry-Based Instruction through
1 WEB Engineering Introduction to Electronic Commerce COMM1Q.
Advances in the PARCC Mathematics Assessment August
Some Implications of Expertise Research for Educational Assessment Robert J. Mislevy University of Maryland National Center for Research on Evaluation,
Delaware Professional Teaching Standards 3.0 Content Knowledge 4.0 Human Development and Learning 5.0 Diverse Learners 6.0 Communication 7.0 Learning Environment.
Math rigor facilitating student understanding through process goals
Learning Progressions: Some Thoughts About What we do With and About Them Jim Pellegrino University of Illinois at Chicago.
LIVE INTERACTIVE YOUR DESKTOP 1 Start recording—title slide—1 of 3 Introducing the Next Generation Science Standards Originally presented by:
1 Issues in Assessment in Higher Education: Science Higher Education Forum on Scientific Competencies Medellin-Colombia Nov 2-4, 2005 Dr Hans Wagemaker.
Robert J. Mislevy University of Maryland Geneva Haertel & Britte Haugan Cheng SRI International Robert J. Mislevy University of Maryland Geneva Haertel.
UCLA Graduate School of Education & Information Studies National Center for Research on Evaluation, Standards, and Student Testing V3, 4/9/07 SIG: Technology.
Putting Research to Work in K-8 Science Classrooms Ready, Set, SCIENCE.
Developing Assessments for and of Deeper Learning [Day 2b-afternoon session] Santa Clara County Office of Education June 25, 2014 Karin K. Hess, Ed.D.
Sharing Design Knowledge through the IMS Learning Design Specification Dawn Howard-Rose Kevin Harrigan David Bean University of Waterloo McGraw-Hill Ryerson.
On Layers and Objects in Assessment Design Robert Mislevy, University of Maryland Michelle Riconscente, University of Maryland Robert Mislevy, University.
Object-Oriented Software Engineering using Java, Patterns &UML. Presented by: E.S. Mbokane Department of System Development Faculty of ICT Tshwane University.
“When learners are at a distance, the careful design of assessments is particularly important, because society somewhat unfairly imposes higher expectations.
Effective Teaching Strategies Day 2
How People Learn – Brain, Mind, Experience, and School (Bransford, Brown, & Cocking, 1999) Three core principles 1: If their (students) initial understanding.
1 © 2005 Cisco Systems, Inc. All rights reserved. 9429_03_2004_c1 Using Cisco Network Simulator.
Resources and Reflections: Using Data in Undergraduate Geosciences Cathy Manduca SERC Carleton College DLESE Annual Meeting 2003.
1 Assessing Student Understanding David Niemi UCLA Graduate School of Education & Information Studies National Center for Research on Evaluation, Standards,
Evidence-Centered Game Design Kristen DiCerbo, Ph.D. Principal Research Scientist, Pearson Learning Games Scientist, GlassLab.
Standards-Based Science Assessment. Ohio’s Science Cognitive Demands Science is more than a body of knowledge. It must not be misperceived as lists of.
Robert J. Mislevy University of Maryland National Center for Research on Evaluation, Standards, and Student Testing (CRESST) NCME San Diego, CA April 15,
Click to edit Master title style Overview of the NGSS Framework.
CREATING AN ACTIVE LEARNING ENVIRONMENT Using Inquiry and Primary Sources.
“…the ability to think and act flexibly with what one knows.” David Perkins in Teaching for Understanding, Wiske et al (1998)
Nevada State Science Standards Revision: Why NGSS?
CDIO: Overview, Standards, and Processes (Part 2) Doris R. Brodeur, November 2005.
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
Designing Quality Assessment and Rubrics
Inquiry learning and SimQuest
Using Cognitive Science To Inform Instructional Design
Creating an Active Learning environment
Principled Assessment Designs for Inquiry (PADI)
What Do Scientists Really Do?
Educational Technology Lab, National Kapodistrian
Teacher Evaluator Student Growth Retraining Academy
Presentation transcript:

U Iowa Slide 1 Sept 19, 2007 Some Terminology and Concepts for Simulation-Based Assessment Robert J. Mislevy University of Maryland In collaboration with John T. Behrens & Dennis C. Frezzo Cisco Systems, Inc. Presented at the Lidquist Center at the University of Iowa, Iowa City, IA, September 19, 2007

U Iowa Slide 2 Sept 19, 2007 Examples Computer-based, interactive, simulation-based. Info-processing and sociocultural roots. l PacketTracer (Cisco Systems) Computer network design & troubleshooting l The Architectural Registration Examination Architectural design; CAD-like environment l Hydrive Troubleshooting the F15 hydraulics system l Nat’l Board of Medical Examiner’s Primum ® Computer Case Simulations (CCS) l DISC simulator Simulations for problem-solving in dental hygiene

U Iowa Slide 3 Sept 19, 2007 Two Not-Entirely-Satisfactory Ways of Assessing Competence in Computer Networking The Cisco Networking Academy supports 60,000+ academies in schools. CNA supplies curriculum & assessment resources online. Standardized multiple-choice exams »Okay on content knowledge, but... »nothing about interacting with functioning networks or troubleshooting in inquiry cycles. High fidelity local hands-on exams »Authentic, high face validity. »Huge local variation, unreliable, expensive

U Iowa Slide 4 Sept 19, 2007 PacketTracer Goal is improving instruction by fostering communication and enabling agents The patterns of activity and structure in our software embody both cognitive theory and theory of the domain structure. Tasks are tasks whether you are teaching with them or testing with them.

U Iowa Slide 5 Sept 19, 2007 PacketTracer Create and edit networks Multiple representations in the logical layer Inspection tool links to a deeper physical world Simulation mode »Detailed visualization and data presentation Standard support for world authoring »Library of elements »Simulation of relevant deep structure »Copy, paste, save, edit, annotate, lock Skip Screens

Instructors and students can author their own activities

Explanation

Experimentation

U Iowa Slide 9 Sept 19, 2007 The WRONG way Ask ‘how do you score it?’ after you’ve built the simulator and scripted a handful of idiosyncratic scenarios. Two ways to design simulation-based assessment system when you’re looking at thousands of people, with hundreds of tasks, at high cost and with high stakes...

U Iowa Slide 10 Sept 19, 2007 A RIGHT way Design the simulator and the scenarios around what you want to make inferences about, what you need to see to ground them, and the structure of the interrelationships. Two ways to design simulation-based assessment system when you’re looking at thousands of people, with hundreds of tasks, at high cost and with high stakes...

U Iowa Slide 11 Sept 19, 2007 Evidence-Centered Assessment Design Formal, multiple-layered framework from Messick’s (1994) guiding questions: What complex of knowledge, skills, or other attributes should be assessed? What behaviors or performances should reveal those constructs? What tasks or situations should elicit those behaviors?

U Iowa Slide 12 Sept 19, 2007 Evidence-Centered Assessment Design Principled framework for designing, producing, and delivering assessments Process model, object model, design tools Explicates the connections among assessment designs, inferences regarding students, and the processes needed to create and deliver these assessments. Particularly useful for new / complex assessments.

U Iowa Slide 13 Sept 19, 2007 Evidence-Centered Assessment Design PADI: Principled Assessment Design for Inquiry »NSF project, collaboration with SRI et al. » Bob Mislevy home page » »Links to papers on ECD »Cisco NetPASS

From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability. Layers in the assessment enterprise

From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

U Iowa Slide 19 Sept 19, 2007 From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

U Iowa Slide 20 Sept 19, 2007 From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

U Iowa Slide 21 Sept 19, 2007 Packet Tracer: Task Models & Design Wizard - Inline editing of feedback for each assessable item. - Activity and observable level feedback possible - Scoring based on network configuration, functionality, and/or solution process

U Iowa Slide 22 Sept 19, 2007 Packet Tracer: Variable Manager - Variable Assignment - When the activity starts, instead of using the initial network as the starting values, the activity will configure the network with the contents of the variables as entered by the student.

U Iowa Slide 23 Sept 19, 2007 Sample evidence rule from HYDRIVE IF an active path which includes the failure has not been created and the student creates an active path which does not include the failure and the edges removed from the problem area are of one power class, THEN the student strategy is splitting the power path ELSE the student strategy is not splitting the power path. Note… Sequential dependence Contextualized definition Multiple ways to succeed & fail Higher-level definition of what we’re looking for

U Iowa Slide 24 Sept 19, 2007 NetPass: An Evidence-Model Bayes-Net Fragment for Troubleshooting Tasks

U Iowa Slide 25 Sept 19, 2007 From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

U Iowa Slide 26 Sept 19, 2007 From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

U Iowa Slide 27 Sept 19, 2007 From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

U Iowa Slide 28 Sept 19, 2007 From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

U Iowa Slide 29 Sept 19, 2007 From Mislevy & Riconscente, in press Assessment Delivery Students interact with tasks, performances evaluated, feedback created. Four- process delivery architecture. Assessment Implementation Conceptual Assessment Framework Domain Modeling Domain Analysis What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? What is important about this domain? What work and situations are central in this domain? What KRs are central to this domain? How do we represent key aspects of the domain in terms of assessment argument. Conceptualization. Design structures: Student, evidence, and task models. Generativity. Manufacturing “nuts & bolts”: authoring tasks, automated scoring details, statistical models. Reusability.

U Iowa Slide 30 Sept 19, 2007 Conclusion Insights from cognitive research can improve the practice of assessment. Doing so requires … »Shift in perspective and »deeper understanding of assessment design. “Too many notes” ? (Emperor Joseph II on Mozart) Suitable conceptual frameworks, tools, and exemplars are now appearing. »Generativity, re-usability, and inter-operability are keys.