PI: Christina Hendricks

Slides:



Advertisements
Similar presentations
Co-Teaching as a Model of Student Teaching: Common Trends and Levels of Student Engagement Co-Teaching as a Model of Student Teaching: Common Trends and.
Advertisements

Introduction to Psychology
Developing the Teaching Portfolio Carol Tresolini, Ph. D
JENNA PORTER DAVID JELINEK SACRAMENTO STATE UNIVERSITY Statistical Analysis of Scorer Interrater Reliability.
CWU Writing Assessment. Why? To improve student learning.
Making Your Assessments More Meaningful Flex Day 2015.
Ritual or reality: do student evaluations have any effect on teacher thinking and practices? Presentation at the Australasian Higher Education Evaluation.
Longitudinal Experiments Larry V. Hedges Northwestern University Prepared for the IES Summer Research Training Institute July 28, 2010.
Program Review and General Education Assessment at the University at Albany: Past, Present and Future Barbara Wilkinson Assistant Director for Assessment.
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
Chapter 7 Correlational Research Gay, Mills, and Airasian
Learners’ Internal Management of Cognitive Processing in Online Learning Chun-Ying Chen Department of Electronic Commerce Transworld Institute of Technology,
Welcome to Psychology 437 Advanced Psychology Lab: Research Methods in Personality Psychology web page:
An Introduction to HLM and SEM
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
The ZEN of Writing (and a few other matters) EDU 8603.
Technology Tools to Support Writing: Preliminary WWWOW Grant Results Presented by Jane Cook & Chris La Casse CWP Summer Institute, July 2007.
Discover the Future of L2 Writing Through Blogs and Wikis.
Near East University Department of English Language Teaching Advanced Research Techniques Correlational Studies Abdalmonam H. Elkorbow.
NorthSky Nonprofit Network Creating Customer Satisfaction Surveys Presented by Christine A. Ameen, Ed.D. Ameen Consulting & Associates
Think. Learn. Succeed. Preparing International Students to Meet Academic Writing Challenges Melissa Allen Coordinator of Support Services for Non-Native.
Measuring Complex Achievement
Educators’ Attitudes about the Accessibility and Integration of Technology into the Secondary Curriculum Dr. Christal C. Pritchett Auburn University
Peer review systems, e.g. SWoRD [1], need intelligence for detecting and responding to problems with students’ reviewing performance E.g. problem localization.
NTU Natural Language Processing Lab. 1 Investment and Attention in the Weblog Community Advisor: Hsin-Hsi Chen Speaker: Sheng-Chung Yen.
Objectives 2.1Scatterplots  Scatterplots  Explanatory and response variables  Interpreting scatterplots  Outliers Adapted from authors’ slides © 2012.
Presenter: Wan-Ning Chen Professor: Ming-Puu Chen Date: November 11, 2009 Ramaswamy, S., Harris, I., & Tschirner, U. (2001). Student peer teaching: An.
Printed by The Aftercare and School Observation System: Characteristics of out-of-home contexts and young children’s behavior problems.
What is a Volunteer Peer Reviewer? Cathy Owens Swift Project Manager – MERLOT Editorial Boards Editor – Business Editorial Board Professor of Marketing.
Chapter 6: Analyzing and Interpreting Quantitative Data
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
26134 Business Statistics Week 4 Tutorial Simple Linear Regression Key concepts in this tutorial are listed below 1. Detecting.
Teaching Peer Review of Writing in a Large First-Year Electrical and Computer Engineering Class: Comparison of Two Methods Michael Ekoniak Molly Scanlon.
Project VIABLE - Direct Behavior Rating: Evaluating Behaviors with Positive and Negative Definitions Rose Jaffery 1, Albee T. Ongusco 3, Amy M. Briesch.
Choosing and using your statistic. Steps of hypothesis testing 1. Establish the null hypothesis, H 0. 2.Establish the alternate hypothesis: H 1. 3.Decide.
OBJECTIVE INTRODUCTION Emergency Medicine Milestones: Longitudinal Interrater Agreement EM milestones were developed by EM experts for the Accreditation.
Analysis of AP Exam Scores
Analyzing Small Group Research DR. Joann Keyton
Anthony Williams, Maria Northcote, Jason Morton and John Seddon
SAT Prep Lesson #3 SAT vs. ACT.
Using MERLOT To Support Retention, Tenure, and Promotion Reviews
Suketu Bhavsar Director, Kellogg Honors College
Scholarship of Teaching and Learning
APA Format What you need to know
Coding Manual and Process
Assessing Learning Outcomes:
Analyzing and Interpreting Quantitative Data
Teacher slide Objectives: To consider how A level choices affect your options when applying to universities. You need: A level choices – student worksheet.
Title: Validating a theoretical framework for describing computer programming processes 29 November 2017.
Kylee’s Career Path How to become an Editor.
© The University of Edinburgh
& Anglophone Writing Assessments
Using Peer Review to Improve Student Writing
Natalie Robinson Centre for Evidence-based Veterinary Medicine
Preparing Tables and Figures: Some Basics
Maria Aguirre, Ph.D. Milagros Bravo, Ph.D. AlACiMa
The Efficacy of Student Evaluations of Teaching Effectiveness
Writing the Introduction
Sociology Outcomes Assessment
Mapping the ACRL Framework and Nursing Professional
January 2019 Designing Upper Economics Electives with a significant writing component Helen Schneider The University of Texas at Austin.
What do they mean and how can I use them?
15.1 The Role of Statistics in the Research Process
Enhancing Small Group Teaching and Group Projects
Developing a Rubric for Assessment
Inter Rater Reliability
Class Components ►Monday/Tuesday
Educational Testing Service
Developing Honors College Admissions Rubric to Ensure Student Success
Getting the most out of interactive and developmental data
Presentation transcript:

PI: Christina Hendricks Longitudinal Analysis of Peer Feedback in a Writing-Intensive Course: A Pilot Study PI: Christina Hendricks Co-PI: Jeremy Biesanz University of British Columbia-Vancouver Funded by the UBC Institute for the Scholarship of Teaching and Learning SoTL Seed Fund Festival of Learning, June 2016 Slides licensed CC-BY 4.0

Literature on peer feedback Receiving peer feedback improves writing (Paulus, 1999; Cho & Schunn, 2007; Cho & MacArthur, 2010; Crossman & Kite, 2012) Giving peer feedback improves writing (Cho & Cho, 2011; Li, Liu & Steckelberg, 2010)

Few studies look at “dose-response curve” GAPS: Most studies look at revisions to a single essay, not changes across different essays PFB PFB Draft 1 Draft 2 Draft 3 Few studies look at “dose-response curve” PFB PFB PFB PFB Essay 1 Essay 2 Essay 3 Essay 4 Essay …n

Pilot study research questions How do students use peer comments given and received for improving different essays rather than drafts of the same essay? Are students more likely to use peer comments given and received for improving their writing after more than one or two peer feedback sessions? How many sessions are optimal? Does the quality of peer comments improve over time?

Interdisciplinary, full year course for first-years http://artsone.arts.ubc.ca Interdisciplinary, full year course for first-years 18 credits (English, History, Philosophy) Students write 10-12 essays (1500-2000 words) Peer feedback tutorials every week (4 students) Friedrich Nietzsche, public domain, Wikimedia Commons Jane Austen, public domain on Wikimedia Commons Osamu Tezuka, public domain on Wikimedia Commons Toni Morrison, Wikimedia Commons, licensed CC BY-SA 2.0

Data for pilot study 2013-2014 10 essays by 12 participants (n=120) Comments by 3 peers on essays (n=1218) Comments by instructor (n=3291) All coded with same rubric

E.g., STREV 2: could use more textual evidence to support your claims Coding Rubric Categories (plus subcategories, for 11 options) Strength of argument Organization Insight Style & Mechanics Numerical value 1: Significant problem 2: Moderate problem 3: Positive comment/praise Change for future Number of “1” comments total: 239 out of over 4000 1’s by students: 35 1’s by instructor: 204 E.g., STREV 2: could use more textual evidence to support your claims

Inter-coder reliability 3 coders: Daniel Munro & Kosta Prodanovic (undergrads, former Arts One) Jessica Stewart (author, editor) Fleiss’ Kappa Intra-class correlation Student comments (n=141) All categories: 0.61 (moderate) Most used categories: 0.8 (excellent) 0.96 (excellent) Essays (n=120) 0.71 (adequate) How much agreement do we observe relative to how much we would expect to see by chance? -- takes into account the frequency of the type of code occurring in the data -- some codes are more frequent, so you’d expect those to have more apparent agreement -1 to +1 0 = amount of agreement we’d expect to see by chance -1 is complete disagreement 0.6 is moderate agreement; 0.8 is substantial -- Kappa includes just the category Many of the mostly used categories have agreement in 0.8 range Reliability on degree: intra class correlation (ICC) of 0.96 -- to what extent is the average across the three raters reliable: average of all the numbers each gave—how does this correlate with the average of everyone who could possibly do this—get no benefit for adding more people -- average is 2.5 -- 1’s are pretty infrequent -- people agree on whether a 2 or a 3 (40% are 2s, 60% are 3s) Change for future

Looking at trends in comments over time

INSTRUCTOR Comments Number of 2 comments over time -.28** Strength Style Organiz. -.04* Insight

STUDENT comments Number of 2 comments over time Strength Organiz. Style -.16** Insight

INSTRUCTOR Comments Number of 3 comments Strength .31*** Style .19** Organiz. .11** Insight .08** These numbers are linear trend over time, not autoregressive

STUDENT Comments Number of 3 comments over time Organiz. Style Strength Insight

How does essay quality change over time?

Essay quality improves linearly b = .038 t(107) = 2.1 p = .037 Essays rated on a 7-point scale

More complex analyses

Cross-lagged panel design with auto-regressive structure Essay Quality Time 1 B Essay Quality Time 2 … N E A D Comments Time 1 Comments Time 2 … N C

Path A: Instructor Comments Essay Quality Time 1 B Essay Quality Time 2 … N E A D Comments Time 1 Comments Time 2 … N C What this says, basically, is that the coders’ ratings of essay quality are pretty similar to the instructor’s comments on essay quality, in these categories at least Significant relationships Ratings of 1 in Strength (-.12*) & Org. (-.23**) Ratings of 2 in Strength (-.06*) & Style (-.08*) Ratings of 3 in Str, (.11*), Insight (.35*), Style (.15*) *p < .05, **p< .01, ***p< .001, ****p < .0001 *****p <.00001

Path A: Student comments Essay Quality Time 1 B Essay Quality Time 2 … N E A D Comments Time 1 Comments Time 2 … N C Significant relationships Ratings of 2 in Insight (-.53*) Ratings of 3 in Organization (.13*) *p < .05, **p< .01, ***p< .001, ****p < .0001 *****p <.00001

Path C: instructor comments Essay Quality Time 1 B Essay Quality Time 2 … N E A D Comments Time 1 Comments Time 2 … N C Significant effects don’t show up if split out by category Comments ratings of 1 (.29**) Comments ratings of 2 (.23*) Comments ratings of 3 (.21, p=.057) *p < .05, **p< .01, ***p< .001, ****p < .0001 *****p <.00001

Path C: student comments Essay Quality Time 1 B Essay Quality Time 2 … N E A D Comments Time 1 Comments Time 2 … N C This could just be saying that students tend to give the same sorts of comments to the same people, but also that things aren’t changing that much from one essay to another. Significant relationships Comments rated 2 in Strength (.22*) & Style (.33**) Comments rated 3 in Style (.31*) *p < .05, **p< .01, ***p< .001, ****p < .0001 *****p <.00001

Path D: Student & Instructor comments Essay Quality Time 1 B Essay Quality Time 2 … N E A D Comments Time 1 Comments Time 2 … N C Significant relationship ONLY if combine student & instructor comments, & only for comments rated 1 (all categories combined): (.05, p=.06)

Research question 1 How do students use peer comments given and received for improving different essays rather than drafts of same essay? Very little significant evidence of relationships in Path D No difference between comments given & received

Research question 2 Are students more likely to use peer comments given and received for improving their writing after more than one or two peer feedback sessions? How many sessions are optimal? No evidence that there is any change over time in path D No difference between comments given or received

Research question 3 Does the quality of peer comments improve over time? No evidence of change over time in path A Essay Quality Time 1 B Essay Quality Time 2 … N E A D Comments Time 1 Comments Time 2 … N C

Research Question 3, cont’d Student/instructor agreement on average numerical ratings on each essay tends to go down over time (-.04**) student ratings increase at only half the rate (.16*) that instructor’s ratings increase (.33*****) *p < .05, **p< .01, ***p< .001, ****p < .0001 *****p <.00001

Research Question 3, cont’d Correlations on number of comments, students & instructor No change in these relationships over time Comment value 1 Comment value 2 Comment value 3 Strength           0.23* Organization 0.21*     0.17* Insight  0.17* Style No change in this over time, though *p < .05, **p< .01, ***p< .001, ****p < .0001 *****p <.00001

Some conclusions Pilot study: feasible for larger sample? Yes, if: instructors code essay quality rather than coders “chunk” essays for cross-lagged analyses have easy collection of comments

References Cho, K., & MacArthur, C. (2010). Student revision with peer and expert reviewing, Learning and Instruction. 20, 328-338. Cho, Y. H., & Cho, K. (2011). Peer reviewers learn from giving comments. Instructional Science, 39, 629-643. Cho, K. & Schunn, C. D. (2007). Scaffolded writing and rewriting in the discipline: A web-based reciprocal peer review system. Computers & Education, 48, 409–426 Crossman, J. M., & Kite, S. L. (2012). Facilitating improved writing among students through directed peer review, Active Learning in Higher Education, 13, 219-229. Li, L., Liu, X., & Steckelberg, A. L. (2010). Assessor or assessee: How student learning improves by giving and receiving peer feedback. British Journal of Educational Technology, 41(3), 525–536. Paulus, T. M. (1999). The effect of peer and teacher feedback on student writing. Journal of Second Language Writing, 8, 265-289.

University of British Columbia-Vancouver Thank you! Christina Hendricks University of British Columbia-Vancouver Website: http://blogs.ubc.ca/christinahendricks Blog: http://blogs.ubc.ca/chendricks Twitter: @clhendricksbc Slides available: https://is.gd/PeerFeedbackPilot_FOL16 Capitals needed underscore Slides licensed CC-BY 4.0