Design Review “Paper Towels Assessment” Education 334X Technology-Based Student Achievement Assessment 2001.10.22 ‘Alim, Josie, Mike G, Bradley, Nina,

Slides:



Advertisements
Similar presentations
Inquiry-Based Instruction
Advertisements

Second Information Technology in Education Study (SITES) A Project of the International Association for the Evaluation of Educational Achievement (IEA)
Assessment types and activities
Performance Assessment
Project-Based vs. Text-Based
SCIENCE & TESTING. CMT-SCIENCE Given for the first time in Spring 2008 in grades 5 & 8 Consists of multiple choice and open ended questions Based on student.
Elementary School Standards-Based Classroom.
Assessment in the MYP CENTURY MIDDLE SCHOOL. What is Assessment? Assessment is integral to all teaching and learning. MYP assessment requires teachers.
Copyright © Allyn & Bacon (2007) Data and the Nature of Measurement Graziano and Raulin Research Methods: Chapter 4 This multimedia product and its contents.
Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI International Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI.
VALIDITY.
Does Schema-Based Instruction and Self-Monitoring Influence Seventh Grade Students’ Proportional Thinking? Asha Jitendra, University of Minnesota Jon R.
Seminar /workshop on cognitive attainment ppt Dr Charles C. Chan 28 Sept 2001 Dr Charles C. Chan 28 Sept 2001 Assessing APSS Students Learning.
Inquiry.
Paper Towels Assessment Development and Evaluation Plan Presented by the Paper Towels Design Team 4 December, 2001.
Inquiry. Inquiry is a term that we often hear when we are talking about science teaching. How do you define “inquiry”?
Principles of High Quality Assessment
Big Ideas and Problem Solving in Junior Math Instruction
Interactive Science Notebooks: Putting the Next Generation Practices into Action
The 5 E Instructional Model
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
A Framework for Inquiry-Based Instruction through
Standards-Based Science Instruction. Ohio’s Science Cognitive Demands Science is more than a body of knowledge. It must not be misperceived as lists of.
SLB /04/07 Thinking and Communicating “The Spiritual Life is Thinking!” (R.B. Thieme, Jr.)
Classroom Assessment A Practical Guide for Educators by Craig A
SOL Changes and Preparation A parent presentation.
1 Issues in Assessment in Higher Education: Science Higher Education Forum on Scientific Competencies Medellin-Colombia Nov 2-4, 2005 Dr Hans Wagemaker.
Understanding Meaning and Importance of Competency Based Assessment
Invention Convention Seth Krivohlavek Angie Deck.
Assessment Professional Learning Module 5: Making Consistent Judgements.
Week 5 Lecture 4. Lecture’s objectives  Understand the principles of language assessment.  Use language assessment principles to evaluate existing tests.
CT 854: Assessment and Evaluation in Science & Mathematics
1 Assessment Professional Learning Module 5: Making Consistent Judgements.
Assessment Tools.
Effective Grading Strategies Alison Morrison-Shetlar Faculty Center for Teaching and Learning Adapted from the book Effective Grading by Barbara Walvoord.
Classroom Assessment (1) EDU 330: Educational Psychology Daniel Moos.
VALUE/Multi-State Collaborative (MSC) to Advance Learning Outcomes Assessment Pilot Year Study Findings and Summary These slides summarize results from.
Assessment Information from multiple sources that describes a student’s level of achievement Used to make educational decisions about students Gives feedback.
Assessment. Workshop Outline Testing and assessment Why assess? Types of tests Types of assessment Some assessment task types Backwash Qualities of a.
Standards-Based Science Assessment. Ohio’s Science Cognitive Demands Science is more than a body of knowledge. It must not be misperceived as lists of.
Maryland College and Career Readiness Conference Summer 2015.
Major Science Project Process A blueprint for experiment success.
Minnesota Manual of Accommodations for Students with Disabilities Training January 2010.
Review: Cognitive Assessments II Ambiguity (extrinsic/intrinsic) Item difficulty/discrimination relationship Questionnaires assess opinions/attitudes Open-/Close-ended.
National Science Education Standards. Outline what students need to know, understand, and be able to do to be scientifically literate at different grade.
Goals 1. To understand inquiry 2. To learn about inquiry-based science 3. To compare children’s science and scientists’ science. 4. To compare two methods.
Strategies for blended learning in an undergraduate curriculum Benjamin Kehrwald, Massey University College of Education.
Grading based on student centred and transparent assessment of learning outcomes Tommi Haapaniemi
FSM NSTT Teaching Competency Test Evaluation. The NSTT Teaching Competency differs from the three other NSTT tests. It is accompanied by a Preparation.
What is Inquiry in Science?. Goals 1. To understand nature of science as inquiry 2. To learn about inquiry as a model of teaching 3. To compare inquiry.
Designing a Culminating Task Presented by Anne Maben UCLA Science & Literacy Coach Based on the model by Jay McTighe, Maryland Assessment Consortium.
Chapter 6 Assessing Science Learning Updated Spring 2012 – D. Fulton.
What does the Research Say About . . .
Inquiry-Based Instruction
Assessment of Learning 1
Classroom Assessment A Practical Guide for Educators by Craig A
Using Cognitive Science To Inform Instructional Design
Computational Reasoning in High School Science and Math
Science, Technology, and Engineering
EarthComm Inquiry: Preparing Students to be Critical Thinkers and Science Literate Citizens Key Points: Today’s talk is about a high school Earth science.
Prepared by: Toni Joy Thurs Atayoc, RMT
Research Methods.
PLATON: Promoting Learning Approaches
Design Review “Paper Towels Assessment”
Critically Evaluating an Assessment Task
Assessment for Learning
Learning Assessment Learning Teaching Dr. Md. Mozahar Ali
TAKS, Inquiry, Standards and Assessment
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
Presentation transcript:

Design Review “Paper Towels Assessment” Education 334X Technology-Based Student Achievement Assessment ‘Alim, Josie, Mike G, Bradley, Nina, Yunn Chyi, Johnnie, Jim

What does this assessment measure? Conceptual framework & locate assessment Conceptual analysis of –task –response format –scoring system What does prior research say? Observational Data: summary of notes from live observation and implications for simulation

In the conceptual framework Content knowledge Constrained Process knowledge Open Rich Lean Paper Towels

Conceptual Analysis DeclarativeProceduralSchematicStrategic Performance Task What is… weight, volume & relationship between them saturation area & length, purpose/use of different tools How to… read/write saturate control variables measure volume, weight etc. determine amount of water soaked up Why… saturate control variables use care in procedures recognize when has reached valid/reliable result recognize when unexpected factors affect result Response Format Which towel absorbs the most and which the least? Steps in procedureHow did you know? Completely wet? Same size? Scoring System Correct resultMethod for getting towel wet Determine result Care in procedure Saturation Determine result Care in procedure

For this type of performance assessment, prior research has found: Student notebook scores are exchangeable for direct observation scores. Notebook scores were slightly more reliable for students with prior experience with hands-on instruction. There is very high interrater reliability. Teachers can be trained to consistently judge performance. There is somewhat higher reliability for observed hands-on performance than for notebooks. There was moderate correlation between this type of assessment and the multiple choice test suggesting that different aspects of science achievement are being measured. There is low to moderate intertask reliability. This may be a concern because a number of tasks are needed in order to be able to generalize the performance to the domain. Sources: Baxter & Shavelson ’94- Evaluation of Procedure-Based Scoring for Hands-On Science Assessment Shavelson, Gao & Baxter ’93 - Sampling Variability of Performance Assessments

n = 1

Observation Consideration: Bias due to familiarity with test and scoring criteria. Single variable procedures only for method. Equalization primary concern for Keri. First Trial: Used dropper to saturate equal pieces. Counted drops and compared. Second Trial: Weighed equal pieces in scale. Third Trial: Used set amount of water in beaker to saturate. Compared amount of water left in beaker.

Valid for what kinds of instructional goals? To develop scientific reasoning, critical thinking, and inquiry/curiosity (3) (2) To design and conduct a sound scientific experiments, including identifying and controlling variables (develop procedural understanding) (4) To apply previously learned skills to new situations (6) To develop skills and appreciation for careful observation and measurement (5) “Paper Towels” is a Performance Assessment designed to support a hands-on, inquiry-based approach to learning and teaching that is essential in the primary grades. (1)

What is the utility of this hands-on performance assessment? Provides richer feedback regarding the student’s ability to solve problems using scientific reasoning The experience is instructional and non- intimidating Allows you to characterize students’ procedural strengths and weaknesses (good feedback to teacher) It’s much more collaborative and fun! More costly than multiple- choice, paper-pencil test (new and additional hands- on materials) Requires more time (cost) to score and be trained to score; requires trained scorers (harder to score) Requires more set-up time Requires more class time to execute the test The lab-book depends on literacy skills It’s weak in content (+) (-)

Reliability Inter-rater Very high,.94 (according to authors) Inter-task Not relevant, only one task is offered Sources of error Very few opportunities for discrepancy Limited to judgment calls: - care taken? - saturated?

What is the construct validity of PT assessment? Instructional GoalsWhat PT measures? 1.To develop reasoning, critical thinking and inquiry skills 1.Effective explanation of principles underlying the performance tasks. 2. Ability to provide a goal-directed, efficient strategies (problem solving) 3. Ability to self-monitor, constantly checking on their thinking and reasoning to determine the result. 2.To design and conduct a sound scientific experiment Ability to assess the nature of the problem and construct a plan (mental model) to carry out the strategy 4.To transfer their knowledgeProblem solving skills (since PT is content lean) 5.Develop skills and appreciation for careful observation and measurement Care in measuring &/or saturation

What is the construct validity of PT assessment? Content validity: The instructional goals of the lesson parallel with what PT is measuring (as shown in the table) Cognitive process PT assess students’ ability to explain the underlying principles, construct a plan (mental model) to initiate the strategy, devise strategies to solve the problem, monitor their thinking and reasoning… while PERFORMING the task (vs. recalling, reciting). These cognitive processes indicate effective learning of the subject matter (Baxtel & Elder, 1996). Performance level Research by Baxter et al (1992) indicates that PA measures correlated less with traditional ability than did a MC achievement test, ie: this PA doesn’t measure dlec Exchangeability for other methods (M-C, C-R, Teacher observation?) Research has shown performance assessment (e.g. PT) is exchangeable for direct observation. As a content lean performance assessment, PT measures different aspects of science achievement (problem solving skills) as Multiple Choice (MC) and Construct Response (CR), which concentrates more on domain knowledge. Hence they are not exchangeable.

Implications for Virtualizing PT Real world mistakes vs. virtual world mistakes - How many can/should we accommodate? Drop size, scale validity, scattered moisture, size of towels… How many roads (scenarios) can we account for? Cross-method results (combining results from multiple experiments) Can human observation really be replaced by computer tracking? Now that we’re not using the notebooks, should we bring back tracking “care” (eg: alarms or flags set for missing “care steps”) Do we still need to give them some hands-on materials? Can we really do this??