Serious Play Conference Los Angeles, CA – July 21, 2012 Girlie C. Delacruz and Ayesha L. Madni Setting Up Learning Objectives and Measurement for Game.

Slides:



Advertisements
Similar presentations
Modelling with expert systems. Expert systems Modelling with expert systems Coaching modelling with expert systems Advantages and limitations of modelling.
Advertisements

Stephanie Burba, Noyce Graduate Tyler Ghee, Noyce Scholar Shelby Overstreet, Noyce Scholar Kathryn Crawford, Noyce Graduate Hope Marchionda, PhD Using.
Computer-Based Performance Assessments from NAEP and ETS and their relationship to the NGSS Aaron Rogat Educational Testing Service.
Using the Crosscutting Concepts As conceptual tools when meeting an unfamiliar problem or phenomenon.
American Educational Research Association Annual Meeting New York, NY - March 23, 2008 Eva L. Baker and Girlie C. Delacruz What Do We Know About Assessment.
Serious Play Conference Los Angeles, CA July 22, 2014 Girlie C. Delacruz Games as Assessment: Validating Patterns of Gameplay Via Cluster Analysis.
Developing Classroom Assessments for the NGSS What evidence of student thinking is needed to determine if a student has met a PE (performance expectation)?
Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI International Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI.
A Linguistics-Based Approach for Use Case Driven Analysis Using Goal and Scenario Authoring Vijayan Sugumaran Oakland University Rochester, Michigan, USA.
Planning Value of Planning What to consider when planning a lesson Learning Performance Structure of a Lesson Plan.
Chapter Two SCIENTIFIC METHODS IN BUSINESS
Specifying a Purpose, Research Questions or Hypothesis
Developing Ideas for Research and Evaluating Theories of Behavior
C R E S S T / U C L A Improving the Validity of Measures by Focusing on Learning Eva L. Baker CRESST National Conference: Research Goes to School Los Angeles,
Principles of High Quality Assessment
Meaningful Learning in an Information Age
Science and Engineering Practices
Science Inquiry Minds-on Hands-on.
The 5 E Instructional Model
Framework for K-12 Science Education
RSBM Business School Research in the real world: the users dilemma Dr Gill Green.
Crosscutting Concepts and Disciplinary Core Ideas February24, 2012 Heidi Schweingruber Deputy Director, Board on Science Education, NRC/NAS.
COPYRIGHT WESTED, 2010 Calipers II: Using Simulations to Assess Complex Science Learning Diagnostic Assessments Panel DRK-12 PI Meeting - Dec 1–3, 2010.
TECHNOLOGY INTEGRATION & INSTRUCTION FOR THE 21 ST CENTURY LEARNER JUNE 15-17, 2009 HOPE BROWN, HIGH SCHOOL SCIENCE, ST. EDMOND, FORT DODGE VALERIE JERGENS,
How do Scientists Think?
CCSSO-National Conference on Student Assessment 2013 National Harbor, MD– June 21, 2013 Girlie C. Delacruz, Eva L. Baker, Gregory K. W. K. Chung Solving.
8th Grade Science Introduction to Physical Science
Cognitive Task Analysis and its Application to Restoring System Security by Robin Podmore, IncSys Frank Greitzer, PNNL.
Learning Law Orientation: August 16, Synthesis Judgment 4. Problem Solving 3. Spotting Issues 2. Understanding 1. Knowledge 1. Recognition vs.
Design Science Method By Temtim Assefa.
SLB /04/07 Thinking and Communicating “The Spiritual Life is Thinking!” (R.B. Thieme, Jr.)
Classroom Assessment A Practical Guide for Educators by Craig A
1 Brief Review of Research Model / Hypothesis. 2 Research is Argument.
EDU 385 EDUCATION ASSESSMENT IN THE CLASSROOM
Progression in ICT Key Stage 1 - Children learn how to…... explore ICT; use it confidently and purposefully to achieve outcomes; use ICT to develop their.
Forum - 1 Assessments for Learning: A Briefing on Performance-Based Assessments Eva L. Baker Director National Center for Research on Evaluation, Standards,
Course on Data Analysis and Interpretation P Presented by B. Unmar Sponsored by GGSU PART 2 Date: 5 July
Big Idea 1: The Practice of Science Description A: Scientific inquiry is a multifaceted activity; the processes of science include the formulation of scientifically.
WELNS 670: Wellness Research Design Chapter 5: Planning Your Research Design.
© Regents of University of California 1 Functional Validity: Extending the Utility of State Assessments Eva L. Baker, Li Cai, Kilchan Choi, Ayesha Madni.
Learning Progressions: Some Thoughts About What we do With and About Them Jim Pellegrino University of Illinois at Chicago.
1 Issues in Assessment in Higher Education: Science Higher Education Forum on Scientific Competencies Medellin-Colombia Nov 2-4, 2005 Dr Hans Wagemaker.
Language and understanding in Physics Acknowledgements: Brian McKittrick, Kim Falloon Helen McDonald & Geoff Davies.
Copyright©2007 Education Service Center Region XIII Mineral Wells ISD 5E CSCOPE Overview.
SOFTWARE DESIGN.
What is a Business Analyst? A Business Analyst is someone who works as a liaison among stakeholders in order to elicit, analyze, communicate and validate.
InterActions Overview This Presentation will touch on the following topics.  Brief Overview  Major Content Themes  Pedagogical Principles  Scaffolding.
ATTRIBUTEDESCRIPTION Focal Knowledge, Skills, Abilities The primary knowledge / skills / abilities (KSAs) targeted by this design pattern. RationaleHow/why.
Putting Research to Work in K-8 Science Classrooms Ready, Set, SCIENCE.
THEORETICAL FRAMEWORK and Hypothesis Development
PRINCIPAL SESSION 2012 EEA Day 1. Agenda Session TimesEvents 1:00 – 4:00 (1- 45 min. Session or as often as needed) Elementary STEM Power Point Presentation.
LEVEL 3 I can identify differences and similarities or changes in different scientific ideas. I can suggest solutions to problems and build models to.
1 Introduction to Software Engineering Lecture 1.
Construct-Centered Design (CCD) What is CCD? Adaptation of aspects of learning-goals-driven design (Krajcik, McNeill, & Reiser, 2007) and evidence- centered.
Computer Control Lou Loftin FETC Conference Orlando, FL January 28 – 31, 2014.
The Next Generation Science Standards: 4. Science and Engineering Practices Professor Michael Wysession Department of Earth and Planetary Sciences Washington.
© 2009 All Rights Reserved Jody Underwood Chief Scientist
Introduction to Physical Science “What is Physical Science?”
Process & Inquiry Skills: Scientists in EC Classrooms CDAEYC Annual Meeting May 19, 2015 Albany, New York.
Unpacking the Elements of Scientific Reasoning Keisha Varma, Patricia Ross, Frances Lawrenz, Gill Roehrig, Douglas Huffman, Leah McGuire, Ying-Chih Chen,
1 The Theoretical Framework. A theoretical framework is similar to the frame of the house. Just as the foundation supports a house, a theoretical framework.
Accelerate increase/decrease in speed or change in direction.
Alternative Assessment Chapter 8 David Goh. Factors Increasing Awareness and Development of Alternative Assessment Educational reform movement Goals 2000,
Science and Engineering Practices K–2 Condensed Practices3–5 Condensed Practices6–8 Condensed Practices9–12 Condensed Practices Developing and Using Models.
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part One INTRODUCTION TO BUSINESS RESEARCH.
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
Classroom Assessment A Practical Guide for Educators by Craig A
Part One INTRODUCTION TO BUSINESS RESEARCH
SCIENCE AND ENGINEERING PRACTICES
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
Presentation transcript:

Serious Play Conference Los Angeles, CA – July 21, 2012 Girlie C. Delacruz and Ayesha L. Madni Setting Up Learning Objectives and Measurement for Game Design

Components of Assessment Architecture Create assessment architecture (Your Example) Assessment Validity Overview

What is so hard?

What are some of your challenges?

Passed the Game

Gameplay Log data Domain

Challenges We Have Translating objectives into assessment outcomes – Purpose of assessment information – Communication between designers and educators Game is developed—need to assess its effectiveness – Cannot change code, wraparounds

How can we meet the challenge?

Assessment requirements Technology requirements Instructional requirements Front-end Efforts Support Effectiveness

Model-Based Engineering Design Communication Collaboration

Model-Based Engineering Design z

Part One ASSESSMENT VALIDITY

Assessment (noun) = Test What Is Assessment?

= Assessment As A Verb Process of drawing reasonable inferences about what a person knows by evaluating what they say or do in a given situation. ASSESSMENT

Games As Formative Assessment Formative Assessment: Use and interpretation of task performance information with intent to adapt learning, such as provide feedback. (Baker, 1974; Scriven, 1967). Formative Assessment: Use and interpretation of task performance information with intent to adapt learning, such as provide feedback. (Baker, 1974; Scriven, 1967).

Games As Formative Assessment Games as Formative Assessment: Use and interpretation of game performance information with intent to adapt learning, such as provide feedback. Games as Formative Assessment: Use and interpretation of game performance information with intent to adapt learning, such as provide feedback.

What is Validity?

Assessment Validity as a Quality Judgment Critical Analysis Legal Judgment Scientific Process

= Assessment Validity Bringing evidence and analysis to evaluate the propositions of interpretive argument. (Linn, 2010) Bringing evidence and analysis to evaluate the propositions of interpretive argument. (Linn, 2010) ASSESSMENT VALIDITY

How Does This Relate to Design? ① Identification of the inferences to be made. What do you want to be able to say? ② Specificity about the expected uses and users of the learning system. Define boundaries of the training system Determine need for supplemental resources ③ Translate into game mechanics ④ Empirical analysis of judgment of performance within context of assumptions. ① Identification of the inferences to be made. What do you want to be able to say? ② Specificity about the expected uses and users of the learning system. Define boundaries of the training system Determine need for supplemental resources ③ Translate into game mechanics ④ Empirical analysis of judgment of performance within context of assumptions.

What do you want to be able to say about the gameplayer(s)? Player mastered the concepts. How do you know? Because they did x, y, z (player history) Because they can do a, b, c (future events)

Identify Key Outcomes: Defining Success Metrics Quantitative Criteria (Generalizable) – % of successful levels/quests/actions – Progress into the game – Changes in performance Errors Time spent on similar levels Correct moves Qualitative Criteria (Game-specific) – Patterns of gameplay – Specific actions

motion pre 1 speed direction duration o1o1 o2o2 o3o3 pre 2 pre 3 pre 4 pre 5 o4o4 o5o5 o6o6 o7o7 o8o8 BACKGROUND LAYER Prior knowledge Game experience Age, sex Language proficiency CONSTRUCT LAYER Construct, subordinate constructs, and inter- dependencies INDICATOR LAYER Behavioral evidence of construct EVENT LAYER Player behavior and game states FUNCTION LAYER Computes indicator value f n (e 1, e 2, e 3,...; s 1, s 2, s 3,...): Computes an indicator value given raw events and game states Game events and states (e 1, e 2, e 3,...; s 1, s 2, s 3,...)

General Approach Derive structure of measurement model from ontology structure Define “layers” – Background: Demographic and other variables that may moderate learning and game performance – Construct: Structure of knowledge dependencies – Indicator: Input data (evidence) of construct – Function: Set of functions that operate over raw event stream to compute indicator value – Event: Atomic in-game player behaviors and game states Assumptions – Chain of reasoning among the layers are accurate

Part Two ASSESSMENT ARCHITECTURE

Components of Assessment Architecture COGNITIVE DEMANDS defines targeted knowledge, skills, abilities, practices domain-independent descriptions of learning COGNITIVE DEMANDS defines targeted knowledge, skills, abilities, practices domain-independent descriptions of learning DOMAIN REPRESENTATION instantiating domain-specific related information and practices guides development allows for external review DOMAIN REPRESENTATION instantiating domain-specific related information and practices guides development allows for external review TASK SPECIFICATIONS defines what the students (tasks/scenarios, materials, actions) defines rules and constraints) defines scoring TASK SPECIFICATIONS defines what the students (tasks/scenarios, materials, actions) defines rules and constraints) defines scoring

Cognitive Demands What kind of thinking do you want capture? Adaptive, complex problem solving Conceptual, procedural, and systemic learning of content Transfer Situation awareness and risk assessment Decision making Self-regulation Teamwork Communication What kind of thinking do you want capture? Adaptive, complex problem solving Conceptual, procedural, and systemic learning of content Transfer Situation awareness and risk assessment Decision making Self-regulation Teamwork Communication

Domain Representation External representation(s) of domain- specific models Defines universe (or boundaries) of what is to be learned and tested External representation(s) of domain- specific models Defines universe (or boundaries) of what is to be learned and tested

Ontologies Item specifications Example: Math Knowledge specifications

Task Specifications ① Operational statement of content and behavior for task Content = stimulus/scenario (what will the users see?) ② Behavior = what student is expected to do/ response (what will the users do?) Content limits ③ Rules for generating the stimulus/scenario posed to the student Permits systematic generation of scenarios with similar attributes Response descriptions ④ Maps user interactions to cognitive requirements ① Operational statement of content and behavior for task Content = stimulus/scenario (what will the users see?) ② Behavior = what student is expected to do/ response (what will the users do?) Content limits ③ Rules for generating the stimulus/scenario posed to the student Permits systematic generation of scenarios with similar attributes Response descriptions ④ Maps user interactions to cognitive requirements

Force and Motion Pushes and pulls, can have different strengths and directions. Pushing and pulling on an object can change the speed or direction of its motion and can start or stop it. Each force acts on one particular object and has both strength and a direction. Energy The faster a given object is moving, the more energy it possesses NGSS performance expectation Plan and conduct an investigation to compare the effects of different strengths of pushes on the motion of an object (K-PS2-1). Analyze data to determine if a design solution works as intended to change the speed or direction of an object with a push (K-PS2-2). Content limits Effects: change in position; increased or decreased acceleration Strengths of pushes: Qualitative (small, medium, big), or quantitative Type of Motion: Rotational Constraints on planar objects: Must be something that can be pushed horizontally and attached to its fulcrum (e.g., the door to a house) Allowable variations on objects: Mass, height and width, location of object Constraints on fulcrum objects: Must be attached to the planar object; position of fulcrum object cannot be changed Data: distance, slope, time, speed Speed change: increase in acceleration Direction: Vertical movement Constraints on planar objects: Must be something flat (e.g., book, frame, ruler) that can be placed on another object and can be pushed in a downward movement Allowable variations on planar objects: Mass, height and width, location of object in the room, surface material Constraints on fulcrum objects: The structural properties of the fulcrum should support some, but not all of the set of planar objects; position of fulcrum object can be changed Targeted science and engineering practice(s) Ask questions that can be investigated based on patterns such as cause and effect relationships. Use observations to describe patterns and/or relationships in that natural and designed world(s) in order to answer scientific questions and solve problems. Response description Ask questions: Query the MARI about the properties of the objects (e.g., what is the distance between the hinge and where I pushed) based on observed outcomes (e.g., how hard it was to push the door, or how far the door moved). Use observations: use snapshot images of activity in the HRLA with overlaid measurement data generated by the MARI to sort situations based on the physical features, behaviors, or functional roles in the design. Task complexity Student only has 4 attempts to pass the ball to the girl and can only vary position and strength of push. Easy: Student can vary the position and strength of the push, but must apply force by placing additional objects on the planar object and pushing downward with both hands (to connect the kinesthetic experience of applying the force with hands on experience of the object). Harder: Student can vary both the position and strength of the push and how the planar object is placed on the fulcrum (e.g., load is moved closer or further away from fulcrum) Available resources Iconic and graphical representation of underlying physics laws will be on the screen, and will change based on student actions. Guided questions will ask students about distance, mass, force magnitude and direction, height, and slope based on observed outcomes.

Components of Computational Model

Components of Decision Model Do nothing: move on, end task Get more evidence or information: repeat same task, perform similar task, ask a question Intervene (instructional remediation): give elaborated feedback, worked example or add scaffolding, more supporting information Intervene (task modification): new task (reduced or increased difficulty), new task (qualitatively different) Courses of Action

Components of Decision Model Confidence of diagnosis : How certain are we about hypothesized causal relation? Consequence of misdiagnosis: What happens if we get it wrong? What are the implications of ignoring other possible states or causal relations? Effectiveness of intervention: How effective is the intervention we will give after diagnosis? Constraints: Do we have to efficiency concerns with respect to time or resource constraints? Decision Factors

Part Three ASSESSMENT ARCHITECTURE (YOUR EXAMPLE)

Person (prior knowledge and experience) Task characteristics Context (test, simulation, game) Fixed Variables + + Assumptions and Design Rationale Assessment Architecture 36

Person (prior knowledge and experience) Task characteristics Context (test, simulation, game) Fixed Variables + + Performance to be Assessed Assessment Architecture 37 Observed Event(s) What happened? (Raw data, scored information?)

Person (prior knowledge and experience) Task characteristics Context (test, simulation, game) Fixed Variables + + Assessment Architecture 38 Observed Event(s) What happened? (Raw data, scored information?) Translation What does this mean? Judgment of performance

Person (prior knowledge and experience) Task characteristics Context (test, simulation, game) Fixed Variables + + Assessment Architecture 39 Observed Event(s) What happened? (Raw data, scored information?) Translation What does this mean? Assessment Validation Inferences What are the potential causes of the observed events? Lack of Knowledge? Context? Characteristics of the task? Not sure?

Potential Course of Actions Repeat Same Trial Get more evidence or information Perform Similar Task Ask a questio n No intervention Move OnEnd TaskInstructional Remediation Intervene Give Elaborated Feedback Worked Example More Information Modify Task New Task With Reduced Difficulty Add Scaffoldin g