Some Implications of Expertise Research for Educational Assessment Robert J. Mislevy University of Maryland National Center for Research on Evaluation,

Slides:



Advertisements
Similar presentations
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California USA
Advertisements

Department of Mathematics and Science
Level 1 Recall Recall of a fact, information, or procedure. Level 2 Skill/Concept Use information or conceptual knowledge, two or more steps, etc. Level.
IAEA 2008 Slide 1 September 8, 2008 Some Implications of Expertise Research for Educational Assessment Robert J. Mislevy, PhD Professor of Measurement.
Improving System Safety through Agent-Supported User/System Interfaces: Effects of Operator Behavior Model Charles SANTONI & Jean-Marc MERCANTINI (LSIS)
Computer-Based Performance Assessments from NAEP and ETS and their relationship to the NGSS Aaron Rogat Educational Testing Service.
Please print the three Cognitive Rigor Matrices full page. Thanks!
© 2013 SRI International - Company Confidential and Proprietary Information Center for Technology in Learning SRI International NSF Showcase 2014 SIGCSE.
The Art and Science of Teaching (2007)
Show Me an Evidential Approach to Assessment Design Michael Rosenfeld F. Jay Breyer David M. Williamson Barbara Showers.
Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI International Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI.
Object-Oriented Analysis and Design
Automated Analysis and Code Generation for Domain-Specific Models George Edwards Center for Systems and Software Engineering University of Southern California.
SRI Technology Evaluation WorkshopSlide 1RJM 2/23/00 Leverage Points for Improving Educational Assessment Robert J. Mislevy, Linda S. Steinberg, and Russell.
University of Maryland Slide 1 May 2, 2001 ECD as KR * Robert J. Mislevy, University of Maryland Roy Levy, University of Maryland Eric G. Hansen, Educational.
University of Maryland Slide 1 July 6, 2005 Presented at Invited Symposium K3, “Assessment Engineering: An Emerging Discipline” at the annual meeting of.
ECOLT 2006 Slide 1 October 13, 2006 Prospectus for the PADI design framework in language testing ECOLT 2006, October 13, 2006, Washington, D.C. PADI is.
LTRC 2007 Messick Address Slide 1 June 9, 2007 Toward a Test Theory for the Interactionalist Era Robert J. Mislevy University of Maryland Samuel J. Messick.
U Iowa Slide 1 Sept 19, 2007 Some Terminology and Concepts for Simulation-Based Assessment Robert J. Mislevy University of Maryland In collaboration with.
C R E S S T / U C L A Improving the Validity of Measures by Focusing on Learning Eva L. Baker CRESST National Conference: Research Goes to School Los Angeles,
Common Core State Standards Professional Learning Module Series
FERA 2001 Slide 1 November 6, 2001 Making Sense of Data from Complex Assessments Robert J. Mislevy University of Maryland Linda S. Steinberg & Russell.
The Software Product Life Cycle. Views of the Software Product Life Cycle  Management  Software engineering  Engineering design  Architectural design.
Computational Thinking Related Efforts. CS Principles – Big Ideas  Computing is a creative human activity that engenders innovation and promotes exploration.
ADL Slide 1 December 15, 2009 Evidence-Centered Design and Cisco’s Packet Tracer Simulation-Based Assessment Robert J. Mislevy Professor, Measurement &
Basic Concepts The Unified Modeling Language (UML) SYSC System Analysis and Design.
Can we make a test fun? Eric Church – BreakAway Games/ University of Baltimore
Terry Vendlinski Geneva Haertel SRI International
CEE2141 CE114/214 Theory: Models functions, forms, and behaviors of organizations, processes, and products Practice Models for AEC construct and interrelate.
ECD in the Scenario-Based GED ® Science Test Kevin McCarthy Dennis Fulkerson Science Content Specialists CCSSO NCSA June 29, 2012 Minneapolis This material.
-Nikhil Bhatia 28 th October What is RUP? Central Elements of RUP Project Lifecycle Phases Six Engineering Disciplines Three Supporting Disciplines.
The Four Strands of Scientific Proficiency Students who understand science:  Know, use, and interpret scientific explanations of the natural world  Generate.
The Design Phase: Using Evidence-Centered Assessment Design Monty Python argument.
Welcome to the Data Warehouse HOME HELP COGNITIVE LEVELS Assessments COGNITIVE LEVELS.
The Revised Bloom’s Taxonomy (RBT): Improving Curriculum, Instruction, and Assessment in an Accountability-Driven, Standards-Based World Developed and.
Learning Science and Mathematics Concepts, Models, Representations and Talk Colleen Megowan.
Learning Progressions: Some Thoughts About What we do With and About Them Jim Pellegrino University of Illinois at Chicago.
1 Issues in Assessment in Higher Education: Science Higher Education Forum on Scientific Competencies Medellin-Colombia Nov 2-4, 2005 Dr Hans Wagemaker.
Robert J. Mislevy University of Maryland Geneva Haertel & Britte Haugan Cheng SRI International Robert J. Mislevy University of Maryland Geneva Haertel.
REVISIONS TO GENERAL EDUCATION STUDENT LEARNING OUTCOMES Auburn University Senate Information Item, August 2014.
A language to describe software texture in abstract design models and implementation.
Baker ONR/NETC July 03 v.4  2003 Regents of the University of California ONR/NETC Planning Meeting 18 July, 2003 UCLA/CRESST, Los Angeles, CA ONR Advanced.
Implementing Inquiry Based Science Teaching Through Teacher Networks The SINUS model for educational development in Germany Matthias Stadler, Kiel (Germany)
Sharing Design Knowledge through the IMS Learning Design Specification Dawn Howard-Rose Kevin Harrigan David Bean University of Waterloo McGraw-Hill Ryerson.
On Layers and Objects in Assessment Design Robert Mislevy, University of Maryland Michelle Riconscente, University of Maryland Robert Mislevy, University.
Computational Thinking Class Overview web site:
Knowledge Representation of Statistic Domain For CBR Application Supervisor : Dr. Aslina Saad Dr. Mashitoh Hashim PM Dr. Nor Hasbiah Ubaidullah.
1 Orientation Session at 2003 Assessment Conference A Richer and More Coherent Set of Assessment Practices Peggy L. Maki Senior Scholar Assessing for Learning.
Basics of Research and Development and Design STEM Education HON4013 ENGR1020 Learning and Action Cycles.
UCLA Graduate School of Education & Information Studies National Center for Research on Evaluation, Standards, and Student Testing Designing Instructional.
1 Assessing Student Understanding David Niemi UCLA Graduate School of Education & Information Studies National Center for Research on Evaluation, Standards,
Evidence-Centered Game Design Kristen DiCerbo, Ph.D. Principal Research Scientist, Pearson Learning Games Scientist, GlassLab.
Robert J. Mislevy University of Maryland National Center for Research on Evaluation, Standards, and Student Testing (CRESST) NCME San Diego, CA April 15,
Maryland College and Career Readiness Conference Summer 2015.
Lifelong Learning: Construction of competencies and knowledge.
Knowing What Students Know: The Science and Design of Educational Assessment Committee on the Foundations of Assessment Board on Testing and Assessment,
An Online Support System to Scaffold Real-World Problem Solving Presenter: Chih-Ming Chen Advisor: Min-Puu Chen Date: 11/10/2008 Ge, X., & Er, N. (2005).
1 Science, Learning, and Assessment: (Eats, Shoots, and Leaves) Choices for Comprehensive Assessment Design Eva L. Baker UCLA Graduate School of Education.
Using the PADI Design System to Examine the Features of a NAEP Performance Assessment Kathleen C. Haynie Kathleen Haynie Consulting Andrea A. Lash, Geneva.
Learning Analytics isn’t new Ways in which we might build on the long history of adaptive learning systems within contemporary online learning design Professor.
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
HTA as a framework for task analysis a review by Andrew Wright (2.8.16)
Mathematical Practice Standards
Using Cognitive Science To Inform Instructional Design
OSEP Leadership Conference July 28, 2015 Margaret Heritage, WestEd
Chapter 1 OBJECT-ORIENTED ANALYSIS AND DESIGN
Standard Setting for NGSS
Automated Analysis and Code Generation for Domain-Specific Models
TAKS, Inquiry, Standards and Assessment
Principled Assessment Designs for Inquiry (PADI)
Presentation transcript:

Some Implications of Expertise Research for Educational Assessment Robert J. Mislevy University of Maryland National Center for Research on Evaluation, Standards, and Student Testing (CRESST) 2005 CRESST Conference UCLA, Los Angeles, CA September 8-9, 2005

The Central Idea lExpertise research seeks to understand in detail the knowledge, representations, development, and contexts of expertise. lResults can ground assessment arguments & task design. lMust recast in terms of purpose, perspective, principles, and structures of assessment. lEspecially valuable with complex proficiencies, situations, or performances.

Outline lAn example: Architectural design lImplications of expertise research lAssessment arguments lGenerative schemas for task design lConclusion

The Architectural Registration Examination lArchitectural design; CAD-like environment. (ETS: Katz, Bejar, Braun, Hone, Brittingham, Bennett, et al.) lTo replace 10-hour hand-drawn design problem lReflects changing of profession to CAD lPremium on thinking, not drawing

Cognitive task analysis Planning the firestation site: A “block diagram” design problem

An Example of a Task Prompt for the ARE

An Illustrative Base Diagram for an ARE Task

A Sample Solution to an ARE Task

Some Results Differences between novices and experts Success rate: 98% vs. 88% Planning time & execution time Patterns of revision involving rework

Assessment Arguments What complex of knowledge, skills, or other attributes should be assessed? What behaviors or performances should reveal those constructs? What tasks or situations should elicit those behaviors? (Messick, 1994)

Implications for task design (1) Experts perceive/understand/act via fundamental principles rather than surface features Chi, Feltovich, & Glaser on physics In architecture, Constraints Number Variation in importance/difficulty Degree of conflict Implicit constraints

Implications for task design (2) Importance of interaction with situation l Intelligence built into situations, tools, processes l “Attunement to affordances” l Cycles of …  Approach to solution, as in ARE  Hypothesize, test, revise, as in inquiry & troubleshooting

Implications for task design (3) External knowledge representations Maps, insurance forms, symbol systems, diagrams, Punnett squares, wiring charts, blueprints Roles in practice l Gather/create/share/hold/transform knowledge l Embody generative principles of domain l Central to practice in the domain Roles in assessment l Environment / stimuli / work of tasks l Design KRs to create / implement assessment

CTA for Assessment CTA can help ground assessment, re validity argument and principled/explicated task construction. l Must aim to specify elements of Student-, Evidence-, and Task-models l in light of the intended assessment’s purpose, l and constraints of the assessment-building and assessment-delivery contexts.

Design patterns Look across domains to find recurring difficulties, blockages, overloads (Salthouse, 1991) Classes of expertise / observables / situations that arise in many domains. Can create “Design patterns” in, e.g., Design under constraints (engineering) Problem-solving in finite domains (troubleshooting) Model-based reasoning (science)

Advances in assessment design A cognitive design system approach to generating valid tests (Embretson,1998) Model-basel assessment (Baker, 1997) Constructing measures (Wilson, 2004) Understanding by design (Wiggins, 1998) On the structure of educational assessments (Mislevy, Steinberg, & Almond, 2003)

Conclusion Insights from expertise research can improve the practice of assessment. Suitable conceptual frameworks, tools, and exemplars are now appearing. Design & delivery frameworks are key to making technology-based / complex assessment practical.