Lecturette 1: Looking at Student Work

Slides:



Advertisements
Similar presentations
Performance Assessment
Advertisements

Problem solving skills
Curriculum Development and Course Design
Session Two Review SEF analysis –Benefits –Issues.
Minnesota State Community and Technical College Critical Thinking Assignment Example and Assessment.
Principles of High Quality Assessment
SUNITA RAI PRINCIPAL KV AJNI
Analytical Thinking.
Chapter 3 Needs Assessment
Standards-Based Science Instruction. Ohio’s Science Cognitive Demands Science is more than a body of knowledge. It must not be misperceived as lists of.
Social Psychology. Experimental reports detail the results of experimental research projects. Experimental reports are write-ups of your results after.
 Participants will teach Mathematics II or are responsible for the delivery of Mathematics II instruction  Participants attended Days 1, 2, and 3 of.
Journal Write a paragraph about a decision you recently made. Describe the decision and circumstances surrounding it. How did it turn out? Looking back,
 For years, corporations have used computer-based simulations with employee-training programs, augmenting traditional on-the-job training with virtual.
Using COS Data to Inform Program Improvement at All Levels Every day, we are honored to take action that inspires the world to discover, love and nurture.
Assessment and Testing
P.R.I.D.E. School Professional Day :45 am- 3:30 pm.
DATA ANALYSIS Looking at Student Work November 2013.
Marking and Feedback CPD Follow up to marking. Expectations and ground rules Respect the views of others Give everyone space to make a contribution All.
Ronica Pardesi 30/09/08 DEVELOPING ESSENTIAL NUMERACY SKILLS IN THE FOUNDATION PHASE MISCONCEPTIONS AND THE METHODS OF REMEDIATION.
Therapeutic Use of Groups superKAT :). Group 3 or more people who are together for some period of time with common goals or share a common purpose.
Key Updates. What has changed? National Curriculum Early Years baseline assessment SATS Teacher Assessments Assessment without levels, expected standards.
LEADERSHIP & TEACHER DEVELOPMENT
Conceptual Change Theory
AP CSP: Making Visualizations & Discovering a Data Story
Chapter 33 Introduction to the Nursing Process
The New Learning Needs Analysis Tool
Articulating Your Practice C3 - Session #3
Internal Assessment 2016 IB Chemistry Year 2 HL.
Lecturette 2: Steps to Improve Data Use
Scientific Method Miss Martel Free Powerpoint Templates.
Clear & Convincing Evidence
Course name: Weekly Planning
Articulating Your Practice C3 - Session #3
Academic Rubric Slides
The Scientific Method: How to solve just about anything
Component 4: Effective and Reflective Practitioner Session 2
Counseling with Depth of Knowledge
Assessment Design Essential Question Key Understandings
Component 4 Effective and Reflective Practitioner
Self Manager What life skills have you used today? Have you been a….
Analyzing Student Work
Reasoning and Problem Solving Maths Café
Scientific Inquiry Unit 0.3.
Reading Research Papers
Analyze Student Work Sample 2 Instructional Next Steps
Student Assessment and Evaluation
Revisit Differentiation and Reflection Standard Prepare for Submission
From Notes to Research Paper. Pathways to Success
IB BIOLOGY INTERNAL ASSESSMENT
Procedures & Collecting Data
Analyzing Student Work Sample 2 Instructional Next Steps
Lecturette 2: Mining Classroom Data
8th Grade Science Content Standard
Lecturette 2: Steps to Improve Data Use
I Think Therefore I Am…A Scientist!
IB BIOLOGY INTERNAL ASSESSMENT
Information Gathering
EDTE 408 Principles of Teaching
Lecturette 1: Looking at Student Work
Research Proposal and Report
The Writing Process.
Copy into binder – Class notes The Scientific Method
Designing & Conducting Formative Evaluation
Biological Science Applications in Agriculture
Student Assessment and Evaluation
Chapter 4 Summary.
Assessment In Mathematics
8th Grade Science Content Standard
Lecturette 1: Looking at Student Work
Presentation transcript:

Lecturette 1: Looking at Student Work Student work provides us with patterns with which to inform our instruction. These patterns show input, output and conceptual errors. Using multiple assessments, we can identify the patterns and improve student learning. Copyright 2004 NCCRESt www.nccrest.org

Jeffrey Watson explains… “Using data is [repetitive].” Educators must have the time and tools to: cycle through steps involving inquiry, acquire data, manage, analyze, apply and evaluate. (Johnson, J.H. Data-driven School Improvement. ERIC Clearinghouse on Educational Management, 1997) Jeffrey Watson Explains: There is a cycle for using data. You look at, gather, analyze, apply, evaluate and do it again and again. There is no end to the cycle. The evaluation stage should push you to analyze data in a new way, or see the need to gather new or different data. Copyright 2004 NCCRESt www.nccrest.org

Conclusions reached through data analysis should be based on multiple sources and measures Do not draw your conclusions on one source of data. If a measure provides you with information that causes concern, conduct another analysis. Only with multiple sources can you reach a valid conclusion. Copyright 2004 NCCRESt www.nccrest.org

What to multiple sources reveal about student learning and performance? What does this work reveal about student learning and performance? If we analyze multiple measurements over time, we should see patterns in particular students’ works. Patterns are a starting point for helping students get the skills and knowledge they are missing. Once patterns are identified, we are able to reassess our situation and move forward with instruction. Copyright 2004 NCCRESt www.nccrest.org

How to Analyze Multiple Sources of Data for Patterns Describe Interpret Reflect How to Analyze Multiple Sources of Data for Patterns: When analyzing more than one source of data, perform the same examinations on each: Describe: What is each form of data measuring? Interpret: What is each form of data telling you? Reflect: Are the measurements describing and interpreting in a similar manner? Copyright 2004 NCCRESt www.nccrest.org

Are the results expected? Why or why not? Assessments aren’t perfect. Be aware of the results. Are they accurately assessing the situation? Do you need to reassess? Do you need to use a different assessment? Facilitator Instructions: Sometimes assessments give inaccurate results. Make sure you convey that to the participants. Participants should use their best judgment when analyzing data. Perhaps the environment was compromised (the AC wasn’t working) or the directions were incomplete. Use this as an opportunity to remind them that this is why you use multiple assessments to analyze student performance. Copyright 2004 NCCRESt www.nccrest.org

Are there any surprises. What results are unexpected Are there any surprises? What results are unexpected? What anomalies exist? Surprises? Assessments sometimes reveal surprises. You may have thought that your students understood a particular issue, but the assessment shows otherwise. Take this opportunity to use another assessment to reevaluate this item and be sure that your students are learning what they need to know. Copyright 2004 NCCRESt www.nccrest.org

What areas did the students perform best in What areas did the students perform best in? What weaknesses are evident? Strengths and Weaknesses: An assessment’s data of students’ strengths and weaknesses are a good way to evaluate your instructional strategies. Copyright 2004 NCCRESt www.nccrest.org

What questions are raised from the data? At this point, what else do you need to know? Do you need further data? Are you ready to move forward? Copyright 2004 NCCRESt www.nccrest.org

How good is good enough? How Good is Good Enough? Assessments show student errors and if you track patterns, then you can evaluate the severity of the problem. Is the error significant? Is good, good enough? Copyright 2004 NCCRESt www.nccrest.org

Of what kinds of student errors do assessments inform us? Input Output Conceptual Of what kinds of student errors do assessments inform us? The three kinds of thinking that assessments measure are recall, evaluation of product, or conceptual thinking. Copyright 2004 NCCRESt www.nccrest.org

Of what kinds of student errors do assessments inform us? Input Output Conceptual Of what kinds of student errors do assessments inform us? Input errors: Students have input errors when the assignment or task has lack of meaning or misunderstands the instructions. Are the instructions clear? Is there too much information? Is there chaos? Facilitator Instructions: Ask participants for examples of input errors. Copyright 2004 NCCRESt www.nccrest.org

Of what kinds of student errors do assessments inform us? Input Output Conceptual Of what kinds of student errors do assessments inform us? Output errors: Students have output errors when they don’t complete the task or did not complete it to the accurate specifications. Is there enough background knowledge? Is the environment conducive for the task? Are there resources available? Facilitator Instructions: Ask participants for examples of output errors. Copyright 2004 NCCRESt www.nccrest.org

Of what kinds of student errors do assessments inform us? Input Output Conceptual Of what kinds of student errors do assessments inform us? Conceptual errors: Students have conceptual errors when they don’t complete the task incorrectly. Some errors include: not enough background experience, incomplete frameworks for understanding, semantic errors, procedural issues, lack of problem solving skills, or processing errors. Facilitator Instructions: Ask participants for examples of conceptual errors. Copyright 2004 NCCRESt www.nccrest.org

Identifying Patterns in Data Allow the groups time to identify patterns and note which ones they find…. (Activity 2) Facilitator Instructions: Explain Activity 2 here. Use the timecard PowerPoints [A2.3 timecards.ppt] (Section 24) to help move the activity along. Copyright 2004 NCCRESt www.nccrest.org