An introduction to intelligent interactive instructional systems

Slides:



Advertisements
Similar presentations
Online Course Design Online Course Design EVALUATION AND REVISION
Advertisements

Formative Feedback Inside Classroom Outside Classroom.
Assessment Adapted from text Effective Teaching Methods Research-Based Practices by Gary D. Borich and How to Differentiate Instruction in Mixed Ability.
The Word Identification Strategy The Learning Strategy Series.
Session Outcomes Explain how assessment contributes to the learning process Use a model of feedback to enhance student learning Identify a range of feedback.
Chapter 1 - An Introduction to Computers and Problem Solving
Chapter 2 - Problem Solving
Chapter 2 - Problem Solving
QA on Anderson et al Intro to CTAT CPI 494 & 598 Jan 27, 2009 Kurt VanLehn.
All you need to know about ITS (in 60 slides) Created by Kurt VanLehn © Presented by Vincent Aleven VanLehn, K. (2006). The behavior of tutoring systems.
QA on “The behavior of tutoring systems” CPI 494 Feb 3, 2009 Kurt VanLehn.
Introduction to teaching and assessing so students will learn more using learner-centered teaching Phyllis Blumberg Warm-up activity How can instructor’s.
Intelligent Tutoring Systems Traditional CAI Fully specified presentation text Canned questions and associated answers Lack the ability to adapt to students.
31 st October, 2012 CSE-435 Tashwin Kaur Khurana.
DR EBTISSAM AL-MADI Computers in Dental Education.
Adapted from Growing Success (Ontario Schools) by K. Gibson
How does an interactive learning environment affect the students’ learning? Marina Issakova University of Tartu, Institute of Computer Science Estonia.
Robert W. Arts, Ph.D. Professor of Education & Physics University of Pikeville Pikeville, KY Presented at the Spring Meeting of the Kentucky Association.
Khan Academy Implementation Models Making the Best Use of Khan Academy with Your Students 1.
Recitation Week #1 Chem Queens College Spring 2010.
Lecture 1 What is Modeling? What is Modeling? Creating a simplified version of reality Working with this version to understand or control some.
Rika Yoshii, Ph.D. and Jacquelyn Hernandez CSIS Department California State University, San Marcos Send us suggestions and requests to.
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
Sudoku Hands-on Training Masters Project Presentation Yiqi Gao March 19, 2014.
GRE General Facts And Self-defense Tips. Registration and General Information How do I register for the GRE? –Call: 1 – 800 – GRE – CALL –Register on-line.
Evaluation of Tutoring Systems Kurt VanLehn PSCL Summer School 2006.
New Teachers’ Induction January 20, 2011 Office of Curriculum and Instruction.
Wiley eGrade. What is eGrade? Web-based software that enables instructors to automate the process of assigning and grading homework and quiz assignments.
1. Go to Assign Assist Assess Student Guide to Assignments in Conference School 2. Go to the Log in page. 3. If.
All you need to know about ITS (in 60 slides) © Kurt VanLehn For journal article version, see distrib/journal/ITSintroAbstract.htm.
Intro: FIT1001 Computer Systems S Important Notice for Lecturers This file is in skeleton form only Lecturers are expected to modify / enhance.
Noboru Matsuda Human-Computer Interaction Institute
Office of School Improvement Differentiated Webinar Series Formative Assessment – Feedback February 28,2012 Dr. Dorothea Shannon, Thomasyne Beverly, Dr.
Teresa Farran Lecturer ICT & Mathematics Education Using an ILS to support learning of numeracy and basic algebra.
The Andes Intelligent Tutoring System: Five years of evaluations Kurt VanLehn Pittsburgh Science of Learning Center (PSLC) University of Pittsburgh.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. A Coursework Support System for Offering Challenges and Assistance by Analyzing Students’ Web Portfolios.
Tutoring Groups School of Electrical Engineering Systems.
1 USC Information Sciences Institute Yolanda GilFebruary 2001 Knowledge Acquisition as Tutorial Dialogue: Some Ideas Yolanda Gil.
Adapted from “Best Practices for Student Learning, Assessment in Online Courses”“Best Practices for Student Learning, Assessment in Online Courses”
Practical Coach Discussion Session #60. Learning Objectives To understand the value of coaching in a manager’s work To learn how to determine when to.
Problem-Solving Approach of Allied Health Learning Community.
A Parent’s Guide to Formative Assessment Communication is Key! Education is shared between the home and the school. Good communication is important as.
Learning through Practice S.Sathiakumar School of Electrical and Information Engineering University of Sydney Problem solving should be a major part of.
711: Intelligent Tutoring Systems Week 1 – Introduction.
Assessment embedded in step- based tutors (SBTs) CPI 494 Feb 12, 2009 Kurt VanLehn ASU.
ASSESSMENT AND PLANNING. Special Education  The term ‘special education’ means specially designed instruction, at no cost to parents, to meet the unique.
ASSESSMENT AND CORRECTION MATHEMATICS EDUCATION: ECED 4251 Rosalind Duplechain, PhD University of West Georgia College of Education Introduction to D&C.
“Teaching”…Chapter 11 Planning For Instruction
Professional Development
The Interactive Strategies Approach to Early Literacy Intervention (ISA) Michelle Eackles RDG 692 Best Practices in Early Literacy Instruction Diane M.
Chapter 2 - VB 2005 by Schneider- modified by S. Jane '081 Chapter 2 - Problem Solving 2.1 Program Development Cycle 2.2 Programming Tools.
Instructional Design Ryan Glidden. Instructional Design The process of planning, creating, and developing instructional or educational resources.
George Goguadze, Eric Andrès Universität des Saarlandes Johan Jeuring, Bastiaan Heeren Open Universiteit Nederland Generation of Interactive Exercises.
Algorithms and Pseudocode
“Intelligent User Interfaces” by Hefley and Murray.
Teaching Roles for Instructional Software Eric Sharp EDMS 6474.
July 8, 2015 UMI Exemplars Fence Painting o Suggested Grade Span – Grades 3-5 o Grade(s) in Which Task was Piloted – Grade 4 Problem Solving.
February 19, 2013 EXPLICIT INSTRUCTION.  After achieving a working knowledge and components of explicit instruction, teachers will self-assess their.
CEIT 225 Instructional Design Prof. Dr. Kürşat Çağıltay
Tia Juana Malone, English Professor Ruth Ronan, Course Developer Assessment Strategies That Promote Student Engagement.
Joshua Miller Blytheville Middle School 7 th Grade Math Student Engagement.
How can we use CATs in tutoring? Lee Helbert-Gordon Director, Institutional Research & Student Success Center Prairie State College.
How to use the assessment process to improve the afterschool program.
Constraint-based tutoring
Learning companions 1 CPI 494, Kurt VanLehn March 26, 2009.
The Behavior of Tutoring Systems
Teaching Java with the assistance of harvester and pedagogical agents
Vincent Aleven & Kirsten Butcher
Julie Booth, Robert Siegler, Ken Koedinger & Bethany Rittle-Johnson
Curriculum Coordinator: Janet Parcell Mitchell January 2016
Presentation transcript:

An introduction to intelligent interactive instructional systems Kurt VanLehn ASU

Outline Tutoring systems Other interactive instructional systems Step loop User interface Interpreting student steps Suggesting good steps Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination Other interactive instructional systems

Intelligent “tutoring” system is a misnomer Almost all are used as seatwork/homework coaches The instructor still… Lectures Leads whole class, small group & lab activities Assigns grades; defends grades Can assign homework / seatwork problems or delegate to the tutoring system The instructor no longer… Grades homework / seatwork Tests? For-profit web-based homework grading services are growing rapidly

If students enter only the answer, call it answer-based tutoring 30° 25 X = 40° y° y° x° Answer 45° What is the value of x?

If students enter steps that derive the answer, call it step-based tutoring 40+30+y=180 70+y=180 y=110 x+45+y=180 x+45+110=180 x=180-155 x=25 45° x° y° 30° 40° What is the value of x? Step Step Step Step Step Answer Step

Def: Feedback is a comment on one of the student’s steps 40+30+y=180 y = 250 45° x° y° 30° 40° What is the value of x? Oops! Check your arithmetic. OK

Feedback is often given as a hint sequence 40+30+y=180 y = 250 45° x° y° 30° 40° What is the value of x? Oops! Check your arithmetic. OK

Hints become more specific 40+30+y=180 y = 250 45° x° y° 30° 40° What is the value of x? You seem to have made a sign error. OK

Try taking a smaller step. Hints segue from commenting on the student’s step to suggesting a better step 40+30+y=180 y = 250 45° x° y° 30° 40° What is the value of x? Try taking a smaller step. OK

and become more specific 40+30+y=180 y = 250 45° x° y° 30° 40° What is the value of x? Try doing just one arithmetic operation per step. OK

Enter 70+y=180, and keep going from there. Def: A bottom-out hint is the last hint, which tells the student what to enter. 40+30+y=180 y = 250 45° x° y° 30° 40° What is the value of x? Enter 70+y=180, and keep going from there. OK

Try doing just one arithmetic operation per step. Def: A next step help request is another way to start up a hint sequence. 40+30+y=180 help 45° x° y° 30° 40° What is the value of x? Try doing just one arithmetic operation per step. OK

Delayed (as opposed to immediate) feedback occurs when the solution is submitted x+45+y=180 x+45+250=180 x=180–250 x= –70 45° x° y° 30° 40° What is the value of x?

Delayed (as opposed to immediate) feedback occurs when the solution is submitted Oops! Check your arithmetic. OK 40+30+y=180 y=250 x+45+y=180 x+45+250=180 x=180–250 x= –70 45° x° y° 30° 40° What is the value of x? Can an angle measure be negative? OK

Both step-based tutors and answer-based tutors have a task loop Tutor and/or student select a task Tutor poses it to the student Student does the task and submits an answer If answer-based tutor, then work offline If step-based tutor, then work online The step-loop = Do step; get feedback/hints; repeat Repeat

Technical terms/concepts (so far) Answer-based tutoring system (= CAI, CBI, …) Step-based tutoring system (= ITS, ICAI…) Step Next-step help Feedback Immediate Delayed Hint sequence Bottom-out hint Task loop Step loop

Andes user interface Read a physics problem Draw vectors Type in equations Type in answer

Andes feedback and hints “What should I do next?” “What’s wrong with that?” Green means correct Red means incorrect Dialogue & hints

SLQ-Tutor (Addison Wesley) Problem Step Step Step Submit! Feedback The database that the problem refers to

Cognitive Algebra I Tutor (Carnegie Learning) Step: Enter an equation Problem Step: Divide both sides Step: Label a column Step: Define an axis Step: Fill in a cell Step: Plot a point

AutoTutor The task Student input is the 2nd half of the step Let me give you an example. We have an AutoTutor for Newtonian Physics. Answer is about a paragraph of information. It may take 100 turns between students and tutors to answer this question. Each tutor turn + student turn in the dialogue is a step Student input is the 2nd half of the step

Introduction: Summary Main ideas Task loop over tasks Step loop over steps of a task Feedback can be immediate or delayed But it focuses on steps Hint sequence Types of tutoring systems Step-based tutors (ITS) – both loops Answer-based tutors (CBT, CAI, etc) – task loop only

Initial framework Step loop Task selection Assessment User interface Interpreting student actions Suggesting good actions Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination

Initial framework Step loop Task selection Assessment User interface Forms, with boxes to be filled Dialogue Simulation Etc. Interpreting student steps Suggesting good steps Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination

Initial framework Step loop Task selection Assessment User interface Interpreting student steps Equations Typed natural language Actions in a simulation Etc. Suggesting good steps Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination

Initial framework Step loop Task selection Assessment User interface Interpreting student steps Suggesting good steps Any correct path vs. shortest path to answer Which steps can be skipped? Recognize the student’s plan and suggest its next step? Etc. Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination

Initial framework Step loop Task selection Assessment User interface Interpreting student steps Suggesting good steps Feedback and hints Give a hint before the student attempts a step? Immediate vs. delayed feedback? feedback on request? How long a hint sequence? When to bottom out immediately? Etc. Task selection Assessment Authoring and the software architecture Evaluations Dissemination

Initial framework Step loop Task selection Assessment User interface Interpreting student steps Suggesting good steps Feedback and hints Task selection Keeping the student in the “zone of proximal development” (ZPD) Mastery learning: Keep giving similar tasks until student master them Choosing a task that suits the learner’s style/attributes Etc. Assessment Authoring and the software architecture Evaluations Dissemination

Initial framework Next Step loop Task selection Assessment User interface Interpreting student steps Suggesting good steps Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination Next

Assessment vs. Evaluation “Assessment” of students What does the student know? How motivated/interested is the student? “Evaluation” of instructional treatments Was the treatment implemented as intended? Did it produce learning gains in most students? Did it produce motivation gains in most students? What is the time cost? Other costs?

Assessment consists of fitting a model to data about the student Single factor model: A single number representing competence/knowledge Probability of a correct answer on a test item = f(competence(student), difficulty(item)) Knowledge component model: One number per knowledge component representing its mastery Probability of a correct answer on a test item = f(mastery(KC1), mastery(KC2), mastery(KC3), …) where KCn are the ones applied in a correct solution

Example: Answer-based assessment of algebraic equation solving skill Test item: Solve 3+2x=10 for x KC5: Subtract from both sides & simplify 3+2x=10  2x=7 KC8: Divide both sides & simplify 2x=7  x=3.5 Single factor model If answer is correct, increment competence else decrement Knowledge component model If answer correct, increment mastery of KC5 & KC8 If answer incorrect, decrement mastery of KC5 & KC8 Weakest one is most likely to be the failure, so decrement it more

Step-based assessment of algebraic equation solving skill Solve 3+2x=10 for x Step1: Step2: Single factor model: Whenever a step is answered correctly without hints, increment competence else decrement Knowledge component model: Whenever a step is answered correctly without hints, increment its KC’s mastery else decrement 2x = 7 x = 3.5

Task selection uses assessments Single factor model Choose a task that is the right level of difficulty i.e., in the ZPD (zone of proximal development) of the student Knowledge component model Choose a task whose solution uses mostly mastered KCs, and only a few KCs that need to be mastered

Other assessment issues Other decisions, besides task selection, that can use assessment? Assessment of motivation or interest? Assessment of learning styles? Disabilities? Diagnosis of misconceptions? Bugs? 8 7 - 1 9 6 2 8 6 7 - 1 9

Should a “Skillometer” displays knowledge component mastery to the student?

Initial framework Next Step loop Task selection Assessment User interface Interpreting student steps Suggesting good steps Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination Next

Authoring Author creates new tasks Who can be an author? Author generates all solutions? System generates all solutions? Same taste as author? Can author add new problem-solving knowledge? Who can be an author? Instructors? Professional authors? Knowledge engineers?

Software architecture & engineering Client-server issues Platform independence Integration with learning management systems E.g., Blackboard, WebAssign, many others Cheating, privacy Quality assurance Software bugs Content & pedagogy bugs

Initial framework Next Step loop Task selection Assessment User interface Interpreting student steps Suggesting good steps Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination Next

Types of evaluations Analyses of expert human tutors What do they do that the system should emulate? Formative evaluation What behaviors of the system need to be fixed? Have students talk aloud, interviews; teachers… Summative evaluation Is the system more effective than what it replaces? Two condition experiment: System vs. control/baseline Pre-test and post-test (+ other assessments) Hypothesis testing Why is the system effective? Multi-condition experiments: System ±feature(s)

Example: Summative evaluation of the Andes Physics tutor University physics (mechanics) 1 semester 2 Conditions: Homework done with… Andes physics tutor Pencil & paper Same teachers (sometimes), text, exams, labs Results (2000-2003) in terms of effect sizes Experimenter’s post-test: d=1.2 Final exam: d=0.3 d = (mean_Andes_score – mean_control_score) ÷ pooled_standard_deviation

Aptitude-treatment interaction (ATI) Control instruction Experimental instruction Good learner (e.g., highly motivated, well prepared…) Large gains Poor learner Small gains

Open-response, problem solving exams scores Andes Exam score Control Grade-point average

Ideal tutoring system adapts to the student’s needs High Bored, & irritated Assistance provided Large learning gains Struggling Low Assistance needed Assistance provided = task selection, feedback, hints, user interface…

No-so-good tutoring system helps only some students High Not-so-good tutor Assistance provided Bored, & irritated Large learning gains Struggling Low Assistance needed Assistance provided = task selection, feedback, hints, user interface…

Initial framework Next Step loop Task selection Assessment User interface Interpreting student steps Suggesting good steps Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination Next

Dissemination = getting the system into widespread use Routes Post and hope Open source Commercialization Issues Instructor acceptance Instructor training Student acceptance Marketing

Outline Next Tutoring systems Other interactive instructional systems Step loop User interface Interpreting student steps Suggesting good steps Feedback and hints Task selection Assessment Authoring and the software architecture Evaluations Dissemination Next Other interactive instructional systems

Other intelligent interactive instructional systems Teachable agent Student deliberately teaches the system, which is then assessed (in public) Learning companion Student works while system encourages Peer learner Student and system work & learn together To be discovered…