Using Computational Linguistics to Support Students and Teachers during Peer Review of Writing Diane Litman Professor, Computer Science Department Senior.

Slides:



Advertisements
Similar presentations
Understanding the ELA/Literacy Evidence Tables. The tables contain the Reading, Writing and Vocabulary Major claims and the evidences to be measured on.
Advertisements

SENIOR SEMINARS Specifics & Example Performances CEPR Center for Educational Policy Research.
TEKS : Write multiple brief responses to teacher-provided, open-ended questions to make connections within and across genres (e.g., literary-literary,
Relating research to practice Heather King Department of Education King’s College London.
Summary-Response Essay
Annie Michaelian Jill Okurowski Stephen Toto. Tri-State Quality Review Rubric.
EQuIP Rubric and Quality Review Curriculum Council September 26, 2014.
Qualitative Grading Notes compiled by Mary D’Alleva January 18 th, 2005 Office of Faculty Development.
© 2012 Aptima, Inc. The Science of Game-based Training Effectiveness 29 March 2012 Krista Langkamer Ratwani Kara L. Orvis.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
The Art of Helping Your Students Help Themselves
Lecturing with Digital Ink Richard Anderson University of Washington.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Common Core State Standards Professional Learning Module Series
Unit 2 – Making It Real Learning Objectives Reflect on your classroom instruction to identify the literacy components you are already using and how they.
Classroom Discussions: Math Discourse in Action
Writing and Sequencing Assignments Christopher J. Bilodeau, Ph.D. Dickinson College, Carlisle, PA, USA
Dr. MaLinda Hill Advanced English C1-A Designing Essays, Research Papers, Business Reports and Reflective Statements.
Servant Leadership Paper The student will concentrate on their individual workplace or business as the focus of a 5-7 page research paper discussing Servant.
Chapter 4 Evaluating and Creating Interactive and Content- Based Assessment.
Group 8 ‘GudBoyz’ teaching writing to L2 learners Agus Prayogo Asih Nurakhir Nico Ouwpoly Sutarno.
The Common Core Curriculum By Dean Berry, Ed. D. Gregg Berry, B.A.
Anne Zeman, Ed.D., Director, Curriculum/Professional Learning Don Azevada, Program Specialist, History/Social Science Ray Pietersen, Program Specialist,
Honors Level Course Implementation Guide [English Language Arts]
NSW Curriculum and Learning Innovation Centre Draft Senior Secondary Curriculum ENGLISH May, 2012.
Automatically Predicting Peer-Review Helpfulness Diane Litman Professor, Computer Science Department Senior Scientist, Learning Research & Development.
Exploring a topic in depth... From Reading to Writing The drama Antigone was written and performed 2,500 years ago in a society that was very different.
English Language Arts Level 7 #44 Ms. Walker
Interactive Dialogue Systems Professor Diane Litman Computer Science Department & Learning Research and Development Center University of Pittsburgh Pittsburgh,
Servant Leadership Thesis & Research Questions Students are to develop at least five questions about Servant Leadership that can be answered through the.
Improving Learning from Peer Review with NLP and ITS Techniques (July 2009 – June 2011) Kevin Ashley Diane Litman Chris Schunn.
ELD Transition Sessions
Natural Language Processing for Writing Research: From Peer Review to Automated Assessment Diane Litman Senior Scientist, Learning Research & Development.
circle Adding Spoken Dialogue to a Text-Based Tutorial Dialogue System Diane J. Litman Learning Research and Development Center & Computer Science Department.
ENGLISH LANGUAGE ARTS MODULE 4 FEBRUARY 2013 Reading Common Core Focus: Text Complexity.
Summary-Response Essay Responding to Reading. Reading Critically Not about finding fault with author Rather engaging author in a discussion by asking.
PRINCIPAL SESSION 2012 EEA Day 1. Agenda Session TimesEvents 1:00 – 4:00 (1- 45 min. Session or as often as needed) Elementary STEM Power Point Presentation.
1 Top 10 fixes 1.Define terms and give more explanation of findings and concepts (minimum 2-3 pages) 2.Add more citations to support claims (5 minimum).
Peer review systems, e.g. SWoRD [1], need intelligence for detecting and responding to problems with students’ reviewing performance E.g. problem localization.
Student Peer Review An introductory tutorial. The peer review process Conduct study Write manuscript Peer review Submit to journal Accept Revise Reject.
Old-school teachingReform teachingPeer interaction ??
Exploring a topic in depth... From Reading to Writing Questioning the world around us is a natural instinct. We read articles in the paper. We read information.
Using Artificial Intelligence to Support Peer Review of Writing Diane Litman Department of Computer Science, Intelligent Systems Program, & Learning Research.
Unit 2 – Making It Real Learning Objectives Reflect on your classroom instruction to identify the literacy components you are already using and how they.
By: Akilah Philips Barry Harris Jeremy Taulton Naairah Lott Tessa Anderson Tiara Edwards.
MH502: Developing Mathematical Proficiency, Geometry and Measurement, K-5 Seminar 1 September 28, 2010.
 An article review is written for an audience who is knowledgeable in the subject matter instead of a general audience  When writing an article review,
Natural Language Processing for Enhancing Teaching and Learning at Scale: Three Case Studies Diane Litman Professor, Computer Science Department Co-Director,
Common Core.  Find your group assignment.  As a group, read over the descriptors for mastery of this standard. (The writing standards apply to more.
Ontology-Based Argument Mining and Automatic Essay Scoring Nathan Ong, Diane Litman, Alexandra Brusilovsky University of Pittsburgh First Workshop on Argumentation.
SE 521 Dr. Lola Taylor Unit one. INTRODUCTION This week's seminar will focus on getting acquainted and understanding the expectations for the course.
GCSE English Language 8700 GCSE English Literature 8702 A two year course focused on the development of skills in reading, writing and speaking and listening.
How to Write an Informative Paper. The purpose of an informative essay is to present information in an organized and coherent fashion, but without expressing.
Implementing the Common Core State Standards Monday, January 23rd - 4pm EST Deconstructing the Common Core Standards: Analyzing for Content, Level of Cognition.
Speech and Natural Language Technology for Educational Applications Diane Litman Professor, Computer Science Department Senior Scientist, Learning Research.
© 2011 Board of Regents of the University of Wisconsin System, on behalf of the WIDA Consortium Introducing the Protocol for Review of Instructional.
INTRODUCTION TO THE WIDA FRAMEWORK Presenter Affiliation Date.
Teaching Peer Review of Writing in a Large First-Year Electrical and Computer Engineering Class: Comparison of Two Methods Michael Ekoniak Molly Scanlon.
Prosodic Cues to Disengagement and Uncertainty in Physics Tutorial Dialogues Diane Litman, Heather Friedberg, Kate Forbes-Riley University of Pittsburgh.
Essay Writing Tips for Writing Essays POL No Simple Answers… Unfortunately there is no simple answer to how to write an essay. However, there are.
Natural Language Processing for Enhancing Teaching and Learning
Learning Assessment Techniques
Developing Critical Thinking using Information Texts
The Stranger Timed Write Exam
Temporal Argument Mining for Writing Assistance
Aspect-based sentiment analysis
Blogging in the Classroom
January 2019 Designing Upper Economics Electives with a significant writing component Helen Schneider The University of Texas at Austin.
Lesson 21: Timed writing About this lesson
Presentation transcript:

Using Computational Linguistics to Support Students and Teachers during Peer Review of Writing Diane Litman Professor, Computer Science Department Senior Scientist, Learning Research & Development Center Director, Intelligent Systems Program University of Pittsburgh Pittsburgh, PA USA Joint work with Professors K. Ashley, A. Godley & C. Schunn 1

Peer Review Research is a Goldmine for Computational Linguistics New Educational Technology! Learning Science at Scale! Can we automate human coding?

Outline SWoRD (Computer-Supported Peer Review) Supporting Students with Review Scaffolding Keeping Teachers Well-informed Summary and Current Directions

SWoRD: A web-based peer review system [Cho & Schunn, 2007] Authors submit papers (or diagrams) Peers submit reviews Authors provide back-reviews to peers

Pros and Cons of Peer Review Pros Quantity and diversity of review feedback Students learn by reviewing Useful for MOOCs Cons Reviews are often not stated in effective ways Reviews and papers do not focus on core aspects Information overload for students and teachers

Outline SWoRD (Computer-Supported Peer Review) Supporting Students with Review Scaffolding Keeping Teachers Well-informed Summary and Current Directions

The P roblem Reviews are often not stated effectively Example: no localization – Justification is sufficient but unclear in some parts. Our Approach: detect and scaffold – Justification is sufficient but unclear in the section on African Americans.

Detecting Key Properties of Text Reviews Computational Linguistics to extract attributes from text, e.g. – Regular expressions (e.g. “the section about”) – Domain lexicons (e.g. “federal”, “American”) – Syntax (e.g. demonstrative determiners) – Overlapping lexical windows (quotation identification) Machine Learning to predict whether reviews contain properties correlating with feedback implementation – Localization – Solutions – Thesis statements

Paper Review Localization Model [Xiong, Litman & Schunn, 2010]

Localization in Diagram Reviews Although the text is minimal, what is written is fairly clear. Study 17 doesn’t have a connection to anything, which makes it unclear about it’s purpose.

Diagram Review Localization Model [Nguyen & Litman, 2013] Pattern-based detection algorithm – Numbered ontology type, e.g. citation 15 – Textual component content, e.g. time of day hypothesis – Unique component, e.g. the con-argument – Connected component, e.g. support of 2nd hypothesis – Numerical regular expression, e.g. H1, #10 11

Learned Localization Model 12 Localized? yes no Pattern Algorithm = yes yesno yes Pattern Algorithm = no #domainWord> 2 #domainWord ≤ 2 windowSize > 16 windowSize ≤ 16 windowSize ≤ 12 windowSize > 12 #domainWord ≤ 0 #domainWord > 0

Localization Scaffolding 13 Localization model applied System scaffolds (if needed) Reviewer makes decision

A First Classroom Evaluation [Nguyen, Xiong & Litman, 2014] Computational linguistics extracts attributes in real-time Prediction models use attributes to detect localization Scaffolding if < 50% of comments predicted as localized Deployment in undergraduate Research Methods

Results: Can we Automate? Diagram reviewPaper review AccuracyKappaAccuracyKappa Majority baseline61.5% (not localized) 050.8% (localized) 0 Our models81.7% %0.46 Comment Level Review Level Diagram reviewPaper review Total scaffoldings17351 Incorrectly triggered10

Results: New Educational Technology Reviewer responseREVISEDISAGREE Diagram review54 (48%)59 (52%) Paper review13 (30%)30 (70%) Response to Scaffolding Why are reviewers disagreeing? No correlation with true localization ratio (diagrams )

A Deeper Look: Revision Performance # and % of comments (diagram reviews) NOT Localized → Localized2630.2% Localized → Localized2630.2% NOT Localized → NOT Localized3338.4% Localized → NOT Localized11.2% Comment localization is either improved or remains the same after scaffolding

A Deeper Look: Revision Performance # and % of comments (diagram reviews) NOT Localized → Localized2630.2% Localized → Localized2630.2% NOT Localized → NOT Localized3338.4% Localized → NOT Localized11.2% Open questions Are reviewers improving localization quality? Interface issues, or rubric non-applicability?

Other Results: Non-Scaffolded Revision Number (pct.) of comments of diagram reviews Scope=InScope=OutScope=No NOT Loc. → Loc %787.5%312.5% Loc. → Loc %112.5%1666.7% NOT Loc. → NOT Loc %00%520.8% Loc. → NOT Loc.11.2%00%0 Localization continues after scaffolding is removed

Outline SWoRD (Computer-Supported Peer Review) Supporting Students with Review Scaffolding Keeping Teachers Well-informed Summary and Current Directions

Observation: Teachers rarely read peer reviews Challenges faced by teachers – Reading all reviews (scalability issues) – Simultaneously remembering reviews across students to compare and contrast (cognitive overload) – Knowing where to start (cold start) 21

Solution: RevExplore SWoRD RevExplore: An Interactive Analytic Tool for Peer-Review Exploration for Teachers [Xiong, Litman, Wang & Schunn, 2012] 22 Peer-review content

RevExplore Example Writing assignment : “ Whether the United States become more democratic, stayed the same, or become less democratic between 1865 and ” Reviewing dimensions : – Flow, logic, insight Goal – Discover student group difference in writing issues 23

K-means clustering Peer rating distribution Target groups: A & B 24 RevExplore Example Step 1 -- Interactive student grouping

25 RevExplore Example Step 2 – Automated topic-word extraction Click “Enter”

26 RevExplore Example Step 2 – Automated topic-word extraction

27 RevExplore Example Step 3 – Group comparison by topic words Group A receives more praise than group B Group A’s writing issues are location- specific – Paragraph, sentence, page, add, … Group B’s are general – Hard, paper, proofread, …

28 RevExplore Example Step 3 – Group comparison by topic words Double click

Evaluating Topic-Word Analytics [Xiong & Litman, 2013] User study (extrinsic evaluation) – 1405 free-text reviews of 24 history papers – 46 recruited subjects Research questions – Are topic words useful for peer-review analytics? – Does the topic-word extraction method matter? – Do results interact with analytic goal, grading rubric, and user demographics? 29

Topic Signatures in RevExplore Domain word masking via topic signatures [Lin & Hovy, 2000; Nenkova & Louis, 2008] – Target corpus: Student papers – Background corpus: English Gigaword – Topic words: Words likely to be in target corpus (chi-square) Comparison-oriented topic signatures – User reviews are divided into groups High versus low writers (SWoRD paper ratings) High versus low reviewers (SWoRD helpfulness ratings) – Target corpus: Reviews of user group – Background corpus: Reviews of all users 30

Comparing Student Reviewers MethodReviews by helpful studentsReviews by less helpful students Topic SignaturesArguments, immigrants, paper, wrong, theories, disprove, theory Democratically, injustice, page, facts 31

Comparing Student Reviewers MethodReviews by helpful studentsReviews by less helpful students Topic SignaturesArguments, immigrants, paper, wrong, theories, disprove, theory Democratically, injustice, page, facts FrequencyPaper, arguments, evidence, make, also, could, argument paragraph Page, think, argument, essay 32

Experimental Results Topic words are effective for peer-review analytics – Objective metrics (e.g. correct identification of high vs. low student groups) – Subjective ratings (e.g. “how often did you refer to the original reviews?”) Topic signature method outperforms frequency Interactions with: – Analytic goal (i.e. reviewing vs. writing groupings) – Reviewing dimensions (i.e. grading rubric) – User demographics (e.g. prior teaching experience) 33

Outline SWoRD (Computer-Supported Peer Review) Supporting Students with Review Scaffolding Keeping Teachers Well-informed Summary and Current Directions

Summary Computational linguistics for peer review to improve both student reviewing and writing Scaffolding useful feedback properties – reviews are often not stated in effective ways Incorporation of argument diagramming – reviews and papers do not focus on core aspects Topic-word analytics for teachers – teacher information overload Deployments in university and high school classes 35

Current Directions Additional measures of review quality – Solutions to problems [Nguyen & Litman, 2014] – Argumentation [Falakmasir et al., 2014; Ong et al., 2014] – Impact on paper revision [Zhang & Litman, 2014] New scaffolding interventions Teacher dashboard – Review and paper revision quality – Topic-word analytics – Helpfulness guided review summarization Talk at 2pm at Oxford tomorrow [Xiong & Litman, submitted]

Thank You! Questions? Further Information – –

Computational Linguistics & Educational Research Learning Language (reading, writing, speaking) Automatic Essay Grading

Computational Linguistics & Educational Research Learning Language (reading, writing, speaking) Using Language (teaching in the disciplines) Tutorial Dialogue Systems (e.g. for STEM) Automatic Essay Grading

Computational Linguistics & Educational Research Learning Language (reading, writing, speaking) Using Language (teaching in the disciplines) Processing Language (e.g. from MOOCs ) Tutorial Dialogue Systems (e.g. for STEM) Automatic Essay Grading Peer Review

Author creates Argument Diagram Peers review Argument Diagrams Author revises Argument Diagram Author writes paper Peers review papers Author revises paper AI: Guides preparing diagram & using it in writing AI: Guides reviewing Phase II: Writing Phase I: Argument Diagramming ArgumentPeer Project Joint work with Kevin Ashley and Chris Schunn

Current Directions: SWoRD in High School Fall 2012 – Spring 2013 – English, History, Science, Math – low SES, urban schools – 9 to 12 grade Classroom contexts – Little writing instruction – Variable access to technology Challenge: different review characteristics Joint work with Kevin Ashley, Amanda Godley, Chris Schunn DomainPraise%Critique%Localized%Solution% College28%62%53%63% High School15%52%36%40%

Common Themes NLP for supporting writing research at scale – Educational technology – Learning science Many opportunities and challenges – Characteristics of student writing Prior NLP software often trained on newspaper texts – Model desiderata Beyond accuracy – Interactions between NLP and Educational Technologies Robustness to noisy predictions Implicit feedback for lifelong computer learning 43