Randy Bennett Frank Jenkins Hilary Persky Andy Weiss Scoring Simulation Assessments Funded by the National Center for Education Statistics,

Slides:



Advertisements
Similar presentations
Inquiry-Based Instruction
Advertisements

IB Portfolio Tasks 20% of final grade
This is a subject-based presentation for you as PV to adapt to your subject. The subject content at present is English for Speakers of other Languages.
Analyzing Student Work
Assessment Assessment should be an integral part of a unit of work and should support student learning. Assessment is the process of identifying, gathering.
Computer-Based Performance Assessments from NAEP and ETS and their relationship to the NGSS Aaron Rogat Educational Testing Service.
Advances in the PARCC Mathematics Assessment August
A Parent’s Guide to Standardized Testing in Georgia
Consultation on Senior Cycle Science Anna Walshe Brendan Duane
Lecture 1 Introduction to the ABAP Workbench
Highlights from PIRLS and TIMSS 2011 Jack Buckley National Center for Education Statistics Washington, DC December 11, 2012.
KSC Mathematics Creating Rubrics for Assessment of General Education Mathematics Dick Jardine SUNY General Education Assessment Conference Syracuse, NY.
Educational Outcomes: The Role of Competencies and The Importance of Assessment.
Consistency of Assessment
CHAPTER 3 ~~~~~ INFORMAL ASSESSMENT: SELECTING, SCORING, REPORTING.
1 SESSION 3 FORMAL ASSESSMENT TASKS CAT and IT FORMS OF ASSESSMENT.
SRI Technology Evaluation WorkshopSlide 1RJM 2/23/00 Leverage Points for Improving Educational Assessment Robert J. Mislevy, Linda S. Steinberg, and Russell.
Planning Value of Planning What to consider when planning a lesson Learning Performance Structure of a Lesson Plan.
Common Core State Standards Professional Learning Module Series
Oregon State Mathematics Assessment By Sandy Kralovec.
1 New York State Mathematics Core Curriculum 2005.
Foreign language and English as a Second Language: Getting to the Common Core of Communication. Are we there yet? Marisol Marcin
Michigan Common Core Standards
Framework for K-12 Science Education
1. An Overview of the Data Analysis and Probability Standard for School Mathematics? 2.
New Advanced Higher Subject Implementation Events
Experiences and requirements in teacher professional development: Understanding teacher change Sylvia Linan-Thompson, Ph.D. The University of Texas at.
Sheila Roberts Department of Geology Bowling Green State University.
Building Effective Assessments. Agenda  Brief overview of Assess2Know content development  Assessment building pre-planning  Cognitive factors  Building.
1 6-8 Smarter Balanced Assessment Update English Language Arts February 2012.
Advances in the PARCC Mathematics Assessment August
KING FAHD UNIVERSITY OF PETROLEUM & MINERALS DHAHRAN, SAUDI ARABIA ARE-512 Life Cycle Costing – Cost Data Bank For By G. C. SOFAT S.K. TYAGI Presented.
Qualifications Update: Environmental Science Qualifications Update: Environmental Science.
TEA Science Workshop #3 October 1, 2012 Kim Lott Utah State University.
Teacher Page I. Scientific Inquiry, A. Processes of Scientific Inquiry Missouri science standards 1.4, 1.5, 1.6, 1.7, 1.8, 2.1, 2.7 7th grade science MAP.
1 Issues in Assessment in Higher Education: Science Higher Education Forum on Scientific Competencies Medellin-Colombia Nov 2-4, 2005 Dr Hans Wagemaker.
The present publication was developed under grant X from the U.S. Department of Education, Office of Special Education Programs. The views.
New Advanced Higher Subject Implementation Events Engineering Science Advanced Higher Course Assessment.
Smarter Balanced Assessment Update English Language Arts February 2012.
Achievethecore.org 1 Setting the Context for the Common Core State Standards Sandra Alberti Student Achievement Partners.
Essentials for Quality Gifted Education Making Middle Grades Work.
Overview of CCSS Statistics and Probability Math Alliance September 2011.
Student assessment Assessment tools AH Mehrparvar,MD Occupational Medicine department Yazd University of Medical Sciences.
Standard Setting Results for the Oklahoma Alternate Assessment Program Dr. Michael Clark Research Scientist Psychometric & Research Services Pearson State.
Qualifications Update: Human Biology Qualifications Update: Human Biology.
Copyright © Allyn & Bacon 2008 Intelligent Consumer Chapter 14 This multimedia product and its contents are protected under copyright law. The following.
Anchor Standards ELA Standards marked with this symbol represent Kansas’s 15%
Quantitative Writing: Communicating Data Presenter: Kim Massaro The University of Texas at San Antonio Quantitative Literacy Program Co-Author: Gail Pizzola.
What Are the Characteristics of an Effective Portfolio? By Jay Barrett.
1 TEXT-BASED SYNTHESIS WRITING: INFORMATIVE / ARGUMENT OPINION NARRATIVE Silver Shores Elementary.
Qualifications Update: Chemistry Qualifications Update: Chemistry.
I NTERNATIONAL B ENCHMARKING S TUDY — C ONTENT A LIGNMENT Mary J. Pitoniak, Nancy Glazer, Luis Saldivia Educational Testing Service June 22, 2015 National.
Learning Objectives Understand the concepts of Information systems.
Key Stage 3 National Strategy Aims of session  To develop greater consistency in teacher assessment of ICT.  To develop a common understanding about.
Greenbush. An informed citizen possesses the knowledge needed to understand contemporary political, economic, and social issues. A thoughtful citizen.
Qualifications Update: Physics Qualifications Update: Physics.
Overview of the handbook Chapter 5: Levee inspection, assessment and risk attribution.
Chapter 6 Assessing Science Learning Updated Spring 2012 – D. Fulton.
Implementing the Professional Growth Process Session 3 Observing Teaching and Professional Conversations American International School-Riyadh Saturday,
NAEP What is it? What can I do with it? Kate Beattie MN NAEP State Coordinator MN Dept of Education This session will describe what the National Assessment.
Knowing What Students Know Ganesh Padmanabhan 2/19/2004.
Philippines – Australia Basic Education Assistance for Mindanao Beam Pre-service Workshop “Authentic Assessment”
INTRODUCTION TO THE ELPAC
Animals use their senses for survival
ASSESSMENT OF STUDENT LEARNING
Architecture Components
SCIENCE AND ENGINEERING PRACTICES
9-12 Smarter Balanced Assessment Update
Presentation transcript:

Randy Bennett Frank Jenkins Hilary Persky Andy Weiss Scoring Simulation Assessments Funded by the National Center for Education Statistics, US Department of Education

What is NAEP? National Assessment of Educational Progress The only nationally representative and continuing assessment of what US students know and can do in various subject areas Paper testing program Administered to samples in grades 4, 8, and 12 Scores reported for groups but not individuals

TRE Study Purpose Demonstrate an approach to assessing “problem solving with technology” at the 8 th grade level that: Fits the NAEP context Uses extended performance tasks Models student proficiency in an evidence- centered way

Conceptualizing Problem Solving with Technology Technology Environment Content Domain Searchable Database Text Processor Simulation Tools Dynamic Displays Spread- sheet Commun- ication Tools Biology Ecology Physics Balloonxxxxxxx Economics History

What do the Example Modules Attempt to Measure? By scientific-inquiry skill, we mean being able to find information about a given topic, judge what information is relevant, plan and conduct experiments, monitor one’s efforts, organize and interpret results, and communicate a coherent interpretation. By computer skill, we mean being able to carry out the largely mechanical operations of using a computer to find information, run simulated experiments, get information from dynamic visual displays, construct a table or graph, sort data, and enter text.

Scoring the TRE Modules Develop initial scoring specifications during assessment design Represent what is being measured as a graphical model Proposal for how the components of proficiency are organized in the domain of problem solving in technology-rich environments

TRE Student Model

Connecting Observations to the Student Model Three-step process Feature extraction Feature evaluation Evidence accumulation

Feature Extraction All student actions are logged in a transaction record Feature extraction involves pulling out particular observations from the student transaction record Example: the specific experiments the student chose to run for each of the Simulation problems

A Portion of the Student Transaction Record #ActionValueTime 1ChooseValues30 2SelectMass9035 3TryIt37 4MakeTable55 5SelectedTabVarsPayload Mass60 6MakeGraph68 7VertAxisAltitude75 8HorizAxisHelium83

Feature Evaluation Each extraction needs to be judged as to its correctness Feature evaluation involves assigning scores to these observations

A Provisional Feature- Evaluation Rule Quality of experiments used to solve Problem 1 IF the list of payload masses includes the low extreme (10), the middle value (50), and the high extreme (90) with or without additional values, THEN the best experiments were run. IF the list omits one or more of the above required values but includes at least 3 experiments having a range of 50 or more, THEN very good experiments were run. IF the list has only two experiments but the range is at least 50 OR the list has more than two experiments with a range equal to 40, THEN good experiments were run. IF the list has two or fewer experiments with a range less than 50 OR has more than two experiments with a range less than 40, THEN insufficient experiments were run.

An Example of a “Best” Solution

An Example of an “Insufficient” Solution

Evidence Accumulation Feature evaluations (like item responses) need to be combined into summary scores that support the inferences we want to make from performance Evidence accumulation entails combining the feature scores in some principled manner Bayesian inference networks Offer a very general, formal, statistical framework for reasoning about interdependent variables in the presence of uncertainty

An Evidence Model Fragment for Exploration Skill in Simulation 1

Using Evidence to Update the Student Model

TRE Student Model

Conclusion TRE illustrates: Measuring problem-solving with technology, with emphasis on the integration of the two skill sets Using extended tasks like those encountered in advanced academic and work environments Modeling student performance in a way that explicitly accounts for multidimensionality and for uncertainty

Conclusion Important remaining issues Measurement Tools to evaluate model fit not well-developed Extended performance tasks have limited generalizability Logistical Adequate school technology not yet universal Cost Task production and scoring are labor-intensive

Randy Bennett Frank Jenkins Hilary Persky Andy Weiss Scoring Simulation Assessments Funded by the National Center for Education Statistics, US Department of Education