The interaction plateau CPI 494, April 9, 2009 Kurt VanLehn 1.

Slides:



Advertisements
Similar presentations
Effective Instructional Conversations + TuTalk Instruction Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Advertisements

Focus on Instructional Support
Chapter 11 user support. Issues –different types of support at different times –implementation and presentation both important –all need careful design.
an hour and a half There is an hour and a half for this paper, which means that, allowing for: reading the questions, making your selection, writing the.
Scaffolding Students’ Comprehension of Text Article written by Kathleen F. Clark & Michael F. Graves Summarized by Kristine Barrett.
COMPOSITIONOPEN RESPONSE  Assess writing  A starting point to initiate a student’s own thoughts on the topic  No “correct” answer (but MUST address.
Introduction to The Scientific Method Biology 155 Spring 2010 B. L. Krilowicz.
Debating the Issue of Tutoring Interactivity: Intuition vs. Experimentation Tanner Jackson It’s a MAD MAD MAD MAD Morning.
(c) Frey & Fisher, 2008 TEACHER RESPONSIBILITY STUDENT RESPONSIBILITY Focused Instruction Guided Instruction “ I do it ” “ We do it ” “ You do it together.
Robert E. Lee Constructed Response The Killer Angels.
Dialogue in Intelligent Tutoring Systems Dialogs on Dialogs Reading Group CMU, November 2002.
Student simulation and evaluation DOD meeting Hua Ai 03/03/2006.
QA on Anderson et al Intro to CTAT CPI 494 & 598 Jan 27, 2009 Kurt VanLehn.
Viewing AutoTutor as a step- based tutoring system CPI 494, April 2, 2009 Kurt VanLehn.
Strategies for Research Methods Tutoring using Latent Semantic Analysis David AllbrittonPeter Wiemer-Hastings Elizabeth ArnottJesse Efron DePaul University.
Sims for soft skills CPI 494, April 23, 2009 Kurt VanLehn.
Providing Tutoring Service through Accumulating Interaction Data Chi-Jen LIN Fo Guang University, Taiwan.
Essay assessors CPI 494, April 14, 2009 Kurt VanLehn.
Simulation-based tutoring systems CPI 494/598, April 16 Kurt VanLehn.
QA on AutoTutor 2004 paper CPI 494, March 31, 2009 Kurt VanLehn.
NLP: Why? How much? How? Peter Wiemer-Hastings. Why NLP? Intro: once upon a time, I was a grad student and worked on MUC. Learned: –the NLP was as good.
Click to edit the title text format An Introduction to TuTalk: Developing Dialogue Agents for Learning Studies Pamela Jordan University of Pittsburgh Learning.
Socratic Seminar “The unexamined life is not worth living.”
Author: Fang Wei, Glenn Blank Department of Computer Science Lehigh University July 10, 2007 A Student Model for an Intelligent Tutoring System Helping.
 Today, we will Conduct an Investigation Take some notes Write a short reflection about what we learned  Please Do Now: get your lab notebook get out.
Smarter Balanced Assessment Consortium Parent Information Session Colchester Public Schools 1.
Meta-Cognition, Motivation, and Affect PSY504 Spring term, 2011 January 25, 2010.
In Vivo Experimentation Lecture 1 for the IV track of the 2012 PSLC Summer School Philip Pavlik Jr. University of Memphis.
Total of ALL forces acting on an object
An Implementation of Vicarious Learning with Deep-Level Reasoning Questions in Middle School and High School Classrooms Barry Gholson, Art Graesser, and.
What does Socratic mean? Socratic comes from the name Socrates Socrates Classical Greek philosopher who developed a Theory of Knowledge.
circle A Comparison of Tutor and Student Behavior in Speech Versus Text Based Tutoring Carolyn P. Rosé, Diane Litman, Dumisizwe Bhembe, Kate Forbes, Scott.
Paper 2 Source Skills. Candidates’ weaknesses (according to examiners’ reports) Not supporting your answers with source detail Simply reproducing knowledge.
Click to edit the title text format Methodology & Basics of Authoring TuTalk Dialogue Agents Pamela Jordan University of Pittsburgh Learning Research and.
Relationship between Physics Understanding and Paragraph Coherence Reva Freedman November 15, 2012.
Evaluation of Tutoring Systems Kurt VanLehn PSCL Summer School 2006.
Speech Analysing Component in Automatic Tutoring Systems Presentation by Doris Diedrich and Benjamin Kempe.
Click to edit the title text format An Introduction to TuTalk: Developing Dialogue Agents for Learning Studies Pamela Jordan University of Pittsburgh Learning.
circle Adding Spoken Dialogue to a Text-Based Tutorial Dialogue System Diane J. Litman Learning Research and Development Center & Computer Science Department.
1 While waiting for the talk to start, try to find 4 mistakes in this student essay. Question: Suppose you are running in a straight line at constant speed.
Evidence-based Practice Chapter 3 Ken Koedinger Based on slides from Ruth Clark 1.
Writing a Thesis for a Literary Analysis Grade 11 English.
Using NLP to Support Scalable Assessment of Short Free Text Responses Alistair Willis Department of Computing and Communications, The Open University,
ITS and SCORM Xiangen Hu, Andrew Olney, Eric Mathews, Art Graesser The University of Memphis.
1 USC Information Sciences Institute Yolanda GilFebruary 2001 Knowledge Acquisition as Tutorial Dialogue: Some Ideas Yolanda Gil.
READING STRATEGIES THAT WORK A Report to the Carnegie Corporation READING NEXT A Vision for Action and Research in Middle and High School Literacy © 2004.
Chap#11 What is User Support?
Assessment embedded in step- based tutors (SBTs) CPI 494 Feb 12, 2009 Kurt VanLehn ASU.
Bell Work (woohoo!) Pick up a SOAPPSTone and complete it. Be ready to discuss in 5 minutes. Remember what we learned about POV this week. Be as thorough.
Spoken Dialog Systems Diane J. Litman Professor, Computer Science Department.
The Effect of Generation and Interaction on Robust Learning Robert G.M. Hausmann Kurt VanLehn Pittsburgh Science of Learning Center Learning Research and.
2 4 6 Classic Problem What is the Rule? Wason’s Problem
Improving (Meta)cognitive Tutoring by Detecting and Responding to Uncertainty Diane Litman & Kate Forbes-Riley University of Pittsburgh Pittsburgh, PA.
KEY ISSUES IN SCIENCE LEARNING Confronting “Intuitive” Theories conceptual change vs. new information Developing Scientific Reasoning hypothesis testing.
Investigate Plan Design Create Evaluate (Test it to objective evaluation at each stage of the design cycle) state – describe - explain the problem some.
Irene-Angelica Chounta, Bruce M. McLaren Carnegie Mellon University
Claims Evidence Reasoning CER Anchor Charts
Presenter: Guan-Yu Chen
Dialogue-Learning Correlations in Spoken Dialogue Tutoring
Constraint-based tutoring
Intelligent Tutoring Systems
The Behavior of Tutoring Systems
Vincent Aleven & Kirsten Butcher
The Effect of Generation on Robust Learning
Chapter 11 user support.
AICE General Paper What IS this class?.
Dr. Karen Terrell University of Massachusetts Dartmouth May 28, 2019
AICE English General Paper
Scientific Method and CER
Presentation transcript:

The interaction plateau CPI 494, April 9, 2009 Kurt VanLehn 1

2 Schematic of a natural language tutoring systems, AutoTutor Step start T: Tell T: ElicitS: Correct Step end S: Incorrect T: Hint or prompt Remediation: Only if out of hints

3 Schematic of other natural language tutors, e.g., Atlas, Circsim-Tutor, Kermit-SE Step start T: Tell T: ElicitS: Correct Step end S: Incorrect T: What is…? S: I don’t know. T:Well, what is… S:… T:… Remediation: Only if out of hints Often called a KCD: Knowledge construction dialogue

4 Hypothesized ranking of tutoring, most effective first A.Expert human tutors B.Ordinary human tutors C.Natural language tutoring systems D.Step-based tutoring systems E.Answer-based tutoring systems F.No tutoring

5 Hypothesized effect sizes

6 Bloom’s (1984) 2-sigma: 4 weeks of human tutoring vs. classroom Classroom

7 Hypothesized effect sizes Classroom Kulik (1984) meta-analysis of CAI vs. classroom  0.4 sigma

8 Hypothesized effect sizes Classroom Many intelligent tutoring systems: e.g., Andes (VanLehn et al, 2005), Carnegie Learning’s tutors…

9 My main claim: There is an interaction plateau

10 A problem and its steps uSuppose you are running in a straight line at constant speed. You throw a pumpkin straight up. Where will it land? 1.Initially, you and the pumpkin have the same horizontal velocity. 2.Your throw exerts a net force vertically on the pumpkin. 3.Thus causing a vertical acceleration. 4.Which leaves the horizontal velocity unaffected. 5.So when the pumpkin falls, it has traveled the same distance horizontally as you have. 6.Thus, it lands in your hands

11 A dialogue between a human tutor (T) and human student (S) uSuppose you are running in a straight line at constant speed. You throw a pumpkin straight up. Where will it land? S: Behind me. –T: Hmm. Let’s think about that. Before you toss the pumpkin and are just carrying it, do you and the pumpkin have the same speed? S: Yes –T: Good. When you toss it up, is the net force on it exactly vertical? S: I’m not sure. T: You exert a force on the pumpkin, right? Etc.

12 Schematic of dialogue about a single step Step start T: Tell T: ElicitS: Correct Step end S: Incorrect T: Hint, or prompt, or explain, or analogy, or … Remediation:

13 Comparisons of expert to novice human tutors Step start T: Tell T: ElicitS: Correct Step end S: Incorrect T: Hint, or prompt, or explain, or analogy, or … Novices Experts Experts may have a wider variety

14 Schematic of an ITS handling of a single step Step start T: Tell S: Correct Step end S: IncorrectT: Hint Only if out of hints

15 Major differences uLow-interaction tutoring (e.g., CAI) –Remediation on answer only uStep-based interaction (e.g., ITS) –Remediation on each step –Hint sequence, with final “bottom out” hint uNatural tutoring (e.g., human tutoring) –Remediation on each step, substep, inference… –Natural language dialogues –Many tutorial tactics

16 Conditions (VanLehn, Graesser et al., 2007) uNatural tutoring –Expert Human tutors »Typed »Spoken –Natural language dialogue computer tutors »Why2-AutoTutor (Graesser et al.) »Why2-Atlas (VanLehn et al.) uStep-based interaction –Canned text remediation uLow interaction –Textbook

17 Human tutors (a form of natural tutoring) Step start T: Tell T: ElicitS: Correct Step end S: Incorrect T: Hint, or prompt, or explain, or analogy, or …

18 Why2-Atlas (a form of natural tutoring) Step start T: Tell T: ElicitS: Correct Step end S: Incorrect A Knowledge Construction Dialogue

19 Why2-AutoTutor (a form of natural tutoring) Step start T: Tell T: ElicitS: Correct Step end S: Incorrect Hint or prompt

20 Canned-text remediation (a form of step-based interaction) Step start T: Tell T: ElicitS: Correct Step end S: Incorrect Text

21 Experiment 1: Intermediate students & instruction

22 Experiment 1: Intermediate students & instruction No reliable differences

23 Experiment 2: AutoTutor > Textbook = Nothing Reliably different

24 Experiments 1 & 2 (VanLehn, Graesser et al., 2007) No significant differences

25 Experiment 3: Intermediate students & instruction Deeper assessments

26 Experiment 3: Intermediate students & instruction No reliable differences

27 Experiment 4: Novice students & intermediate instruction Relearning

28 Experiment 4: Novice students & intermediate instruction All differences reliable

29 Experiment 5: Novice students & intermediate (but shorter) instruction Relearning Add

30 Experiment 5: Novice students & intermediate instruction No reliable differences

31 Experiment 5: Low-pretest students only Aptitude- treatment interaction?

32 Experiment 5, Low-pretest students only Spoken human tutoring > canned text remediation

33 Experiments 6 and 7 Novice students & novice instruction Was the intermediate text over the novice students’ heads?

34 Experiments 6 and 7 Novice students & novice instruction No reliable differences

35 Interpretation Experiments 1 & 4 Experiments 3 & 5 Experiments 6 & 7 High-pretest Low-pretest Intermediates High-pretest Low-pretest Novices Content complexity = Can follow reasoning only with tutor’s help (ZPD) predict: Tutoring > Canned text remediation = Can follow reasoning without any help predict: Tutoring = Canned text remediation

36 Original research questions uCan natural language tutorial dialog add pedagogical value? –Yes, when students must study content that is too complex to be understood by reading alone uHow feasible is a deep linguistic tutoring system? –We built it. It’s fast enough to use. uCan deep linguistic and dialog techniques add pedagogical value?

37 When content is too complex to learn by reading alone: Deep>Shallow? Why2-Atlas is not clearly better than Why2-AutoTutor

38 When to use deep vs. shallow? Shallow linguisticDeep linguistic Sentence understanding LSA, Rainbow, RappelCarmel: parser, semantics… Essay/Discourse understanding LSAAbduction, Bnets Dialog management Finite state networksReactive planning Natural language generation TextPlan-based Use both Use deep Use locally smart FSA Use equivalent texts

39 Results from all 7 experiments (VanLehn, Graesser et al., 2007) uWhy2: Atlas = AutoTutor uWhy2 > Textbook –No essays –Content differences uHuman tutoring = Why2 = Canned text remediation –Except when novice students worked with instruction designed for intermediates, then Human tutoring > Canned text remediation

40 Other evidence for the interaction plateau (Evens & Michael, 2006) No significant differences

41 Other evidence for the interaction plateau (Reif & Scott, 1999 ) No significant differences

42 Other evidence for the interaction plateau (Chi, Roy & Hausmann, in press) No significant differences

43 Still more studies where natural tutoring = step-based interaction uHuman tutors 1.Human tutoring = human tutoring with only content-free prompting for step remediation (Chi et al., 2001) 2.Human tutoring = canned text during post-practice remediation (Katz et al., 2003) 3.Socratic human tutoring = didactic human tutoring (Rosé et al., 2001a 4.Socratic human tutoring = didactic human tutoring (Johnson & Johnson, 1992) 5.Expert human tutoring = novice human tutoring (Chae, Kim & Glass, 2005) uNatural language tutoring systems 1.Andes-Atlas = Andes with canned text (Rosé et al, 2001b) 2.Kermit = Kermit with dialogue explanations (Weerasinghe & Mitrovic, 2006)

44 Hypothesis 1: Exactly how tutors remedy a step doesn’t matter much Step start T: Tell T: ElicitS: Correct Step end S: Incorrect What’s in here doesn’t matter much

45 Main claim: There is an interaction plateau Hypothesis 1

46 Hypothesis 2: Cannot eliminate the step remediation loop Step start T: Tell T: ElicitS: Correct Step end S: Incorrect Must avoid this

47 Main claim: There is an interaction plateau Hypothesis 2

48 Conclusions uWhat does it take to make computer tutors as effective as human tutors? –Step-based interaction –Bloom’s 2-sigma results may have been due to weak control conditions (classroom instruction) –Other evaluations have also used weak controls uWhen is natural language useful? –For steps themselves (vs. menus, algebra…) –NOT for feedback & hints (remeditation) on steps

49 Future directions for tutoring systems research uMaking step-based instruction ubiquitous –Authoring & customizing –Novel task domains uIncreasing engagement

50 Final thought uMany people “just know” that more interaction produces more learning. u“It ain’t so much the things we don’t know that get us into trouble. It’s the things we know that just ain’t so.” –Josh Billings (aka. Henry Wheeler Shaw)