Evaluation of the Advice Generator of an Intelligent Learning Environment Maria Virvou, Katerina Kabassi Department of Informatics University of Piraeus.

Slides:



Advertisements
Similar presentations
Computer Systems & Architecture Lesson 2 4. Achieving Qualities.
Advertisements

Initializing Student Models in Web-based ITSs: a Generic Approach Victoria Tsiriga & Maria Virvou Department of Informatics University of Piraeus.
Evaluation of User Interface Design
Evaluation of a Virtual Reality Game for Education Maria Virvou, Constantinos Manos, George Katsionis, Kalliopi Tourtoglou Department of Informatics University.
System Development Life Cycle (SDLC)
Copyright 1999 all rights reserved The HCI Design Process n User Interfaces are not just built by sitting down and drawing up designs for them n Just like.
A cognitive theory for affective user modelling in a virtual reality educational game George Katsionis, Maria Virvou Department of Informatics University.
Relating Error Diagnosis and Performance Characteristics for Affect Perception and Empathy in an Educational Software Application Maria Virvou, George.
On the motivation and attractiveness scope of the virtual reality user interface of an educational game Maria Virvou, George Katsionis & Konstantinos Manos.
Ch 11 Cognitive Walkthroughs and Heuristic Evaluation Yonglei Tao School of Computing and Info Systems GVSU.
Usability Inspection n Usability inspection is a generic name for a set of methods based on having evaluators inspect or examine usability-related issues.
Marakas: Decision Support Systems, 2nd Edition © 2003, Prentice-Hall Chapter Chapter 7: Expert Systems and Artificial Intelligence Decision Support.
1 CS 430 / INFO 430 Information Retrieval Lecture 24 Usability 2.
SIMS 213: User Interface Design & Development Marti Hearst Tues Feb 13, 2001.
EXPERT SYSTEMS Part I.
© Lethbridge/Laganière 2001 Chapter 7: Focusing on Users and Their Tasks1 7.1 User Centred Design (UCD) Software development should focus on the needs.
User interface design Designing effective interfaces for software systems Objectives To suggest some general design principles for user interface design.
Task analysis 1 © Copyright De Montfort University 1998 All Rights Reserved Task Analysis Preece et al Chapter 7.
Usability and Evaluation Dov Te’eni. Figure ‎ 7-2: Attitudes, use, performance and satisfaction AttitudesUsePerformance Satisfaction Perceived usability.
VR-ENGAGE: A Virtual Reality Educational Game that Incorporates Intelligence Maria Virvou, Constantinos Manos, George Katsionis, Kalliopi Tourtoglou Department.
Virtual Reality edutainment: cost-effective development of personalised software applications Maria Virvou, Konstantinos Manos & George Katsionis Department.
Non-functional requirements
Chapter 11 Management Decision Making
Design, goal of design, design process in SE context, Process of design – Quality guidelines and attributes Evolution of software design process – Procedural,
Requirements Gathering and Task analysis. Requirements gathering and task analysis 4 Requirements gathering is a central part of systems development understanding.
Usability Methods: Cognitive Walkthrough & Heuristic Evaluation Dr. Dania Bilal IS 588 Spring 2008 Dr. D. Bilal.
1. Learning Outcomes At the end of this lecture, you should be able to: –Define the term “Usability Engineering” –Describe the various steps involved.
Learner Modelling in a Multi-Agent System through Web Services Katerina Kabassi, Maria Virvou Department of Informatics, University of Piraeus.
Predictive Evaluation
1 Shawlands Academy Higher Computing Software Development Unit.
LESSON 8 Booklet Sections: 12 & 13 Systems Analysis.
Chapter 8: Systems analysis and design
المحاضرة الثالثة. Software Requirements Topics covered Functional and non-functional requirements User requirements System requirements Interface specification.
System Design: Designing the User Interface Dr. Dania Bilal IS582 Spring 2009.
Evaluation of Adaptive Web Sites 3954 Doctoral Seminar 1 Evaluation of Adaptive Web Sites Elizabeth LaRue by.
Interacting with IT Systems Fundamentals of Information Technology Session 5.
Course Instructor: K ashif I hsan 1. Chapter # 2 Kashif Ihsan, Lecturer CS, MIHE2.
Programming Project (Last updated: August 31 st /2010) Updates: - All details of project given - Deadline: Part I: September 29 TH 2010 (in class) Part.
Multimedia Specification Design and Production 2013 / Semester 1 / week 9 Lecturer: Dr. Nikos Gazepidis
COSC 3461: Module 6 UI Design in the Workplace. 2 What is User-Centered Design? Three major components: –Iterative design –Early focus on users and tasks.
Object-Oriented Software Engineering Practical Software Development using UML and Java Chapter 7: Focusing on Users and Their Tasks.
Evaluation approaches Text p Text p
Diagnosing Language Transfer in a Web-based ICALL that Self-Improves its Student Modeler Victoria Tsiriga & Maria Virvou Department of Informatics, University.
SEG3120 User Interfaces Design and Implementation
1 ISA&D7‏/8‏/ ISA&D7‏/8‏/2013 The Analysis Phase System Requirements Models and Modelling of requirements Stakeholders as a source of requirements.
Evaluation of User Interface Design 4. Predictive Evaluation continued Different kinds of predictive evaluation: 1.Inspection methods 2.Usage simulations.
Software Architecture
An evaluation agent that simulates students’ behavior in Intelligent Tutoring Systems Maria Virvou, Konstantinos Manos, George Katsiwnis Department of.
Chapter 8 Usability Specification Techniques Hix & Hartson.
Welcome to the Usability Center Tour Since 1995, the Usability Center has been a learning environment that supports and educates in the process of usability.
The Software Development Process
EVALUATION PROfessional network of Master’s degrees in Informatics as a Second Competence – PROMIS ( TEMPUS FR-TEMPUS-JPCR)
GCSE ICT Systems Analysis. Systems analysis Systems analysis is the application of analytical processes to the planning, design and implementation of.
Introduction to Evaluation without Users. Where are you at with readings? Should have read –TCUID, Chapter 4 For Next Week –Two Papers on Heuristics from.
Introduction to Evaluation “Informal” approaches.
Fall 2002CS/PSY Predictive Evaluation (Evaluation Without Users) Gathering data about usability of a design by a specified group of users for a particular.
“Intelligent User Interfaces” by Hefley and Murray.
A simulated student-player in support of the authoring process in a knowledge-based authoring tool for educational games Maria Virvou, Konstantinos Manos.
Usability Engineering Dr. Dania Bilal IS 587 Fall 2007.
6. (supplemental) User Interface Design. User Interface Design System users often judge a system by its interface rather than its functionality A poorly.
Methodologies By Akinola Soyinka.
Software engineering USER INTERFACE DESIGN.
Informatics 121 Software Design I
Usability Testing: An Overview
Alfred Kobsa University of California, Irvine
Alfred Kobsa University of California, Irvine
Usability Techniques Lecture 13.
Evaluation.
COMP444 Human Computer Interaction Usability Engineering
Presentation transcript:

Evaluation of the Advice Generator of an Intelligent Learning Environment Maria Virvou, Katerina Kabassi Department of Informatics University of Piraeus

Evaluation of educational software Formative evaluation involving human tutors. A formative evaluation occurs during design and early development of a project. Oriented to the immediate needs of developers. Increase the likelihood that the final product will achieve its stated goals. The evaluation should take place before the implementation to minimize the cost of early design errors.

Intelligent Learning Environment Protected learning environment for novice users of graphical user interfaces. File manipulation program such as windows 98/NT explorer. Monitors users’ actions and reasons about them. In case of an error the system provides spontaneous advice.

Human Plausible Reasoning (HPR) Collins, A., Michalski R.:1989 A descriptive theory on human plausible inference. The theory consists of: –Representation of plausible inference patterns, such as deductions, inductions and analogies. –A set of parameters, e.g. similarity, typicality. –a system relating the different plausible inference patterns and the different certainty parameters.

IFM’s architecture Intelligent Tutoring Systems’ architecture IFM’s components –Domain representation, –Advice generator, –User modeller, –User interface.

Operation of the system 1.The user issues a command. 2.The system reasons about user’s action. 3.If action is considered unexpected, the system generates alternative commands. Otherwise, command is executed. 4.The user is not obligated to follow the system’s advice. The user can: execute his/her initial action execute a new action.

Certainty parameters Degree of typicality of the usage of a command in the set of the total number of executed commands (  ) Degree of similarity of a command or an object to another command or object, respectively (  ) Frequency of an error set in the set of all errors (  ) Dominance of an error in the set of all errors (  ) Degree of certainty (  )  = 0.4 *  *  *  * 

Evaluation IFM’s aim: Provision of plausible advice similar to human tutors’ advice. HPR for simulating the human plausible reasoning of a human advisor. Evaluation’s aim: How close is IFM’s reasoning to human tutors’ reasoning? Trying to reveal what the human tutors’ way of thinking was.

Methods of evaluation Heuristic evaluation –The expert reviewers critique to determine conformance with a short list of design heuristics –Questionnaire Cognitive walkthrough –The experts simulate users walking through the interface to carry out typical tasks –Comment on real-life examples

Results of the Questionnaires. 60% of the human tutors thought that the similarity between objects or commands was the most important aspect when generating advice. 55% believed that the identification of the most frequent error of an individual student was the second more important aspect that should be taken into account. 55% of the human tutors believed that the frequency of an observed error was to be taken into account, but not in first priority. Finally, 60% believed that the frequency of execution of a command was the last aspect to be taken into account.

Additional Results Similarity between two commands. 60% of the human tutors believed that the relative position of commands in the graphical representation was more important than the similarity of their result when executed. This proportion was even greater (60% - 85%) when the human tutors commented on the real-life protocols.

Users’ Protocols Human tutors were asked to comment on real-life examples The protocols collected by an empirical study (early stages of the life-cycle) Human tutors were also given information provided by the user model. Example were given as input to IFM. Comparison of IFM’s advice to human tutors’ comments.

A real-life example Undesired action: A user issues a command for creating a Microsoft Word Document and claims s/he didn’t intend to have done so. 55% of the human tutors suggested –Create a bitmap image, because this command was next to the one issued IFM suggested 1.Create a text document, because this command’s result was similar to that of the command issued (Human tutors’ second alternative). 2.Create a sound file, because this was his/her more usual action (Human tutors’ third alternative).

Analysis of the results Connection of the results with the certainty parameters used by the advice generator. IFM could successfully generate the alternative actions to be suggested to the student but not in the right order. The categorization of the alternative actions was not completely successful and had to be refined. Refinement of the formula for the calculation of the certainty.

Refinement of the formula  = 0.4 *  *  *  *   = 0.4 *  *  *  *  The identification of the most frequent error of a particular student was the second factor to be taken into account (55%) - The weight of dominance is 0.3, instead of 0.1. The fact that a student has repeated an error many times must have been taken into account, but not in first priority (60%) - The weight of frequency was fixed to 0.2, instead of 0.3.

Calculation of the similarity Similarity of objects –the similarity of their names (70%). –their relative distance in the graphical representation. Similarity of commands –their relative distance in the graphical representation (60%). –the similarity of the commands’ result (40%). IFM was successful in this calculation

Conclusions The evaluation aimed –How successful HPR was at reproducing human tutors reasoning. –Usability of the system (future plans). The evaluation revealed –The learning environment was quite successful at generating alternative commands –The order of presentation of the alternative commands to students was not similar to the the one of the majority of human experts. –The evaluation contributed to the refinement of the adaptation of IFM into the learning environment.

Comments & Questions Maria Virvou: Katerina Kabassi:

Simple example 1.Cut(A:\program1.pas) 2.Paste(A:\project2\) The systems finds the action suspect. Replacement of file A:\project2\program1.pas System’s advice: Change directory to A:\project1\ project1 program1.pas A:\ project2 program1.pas program2.pas

Questionnaires When you suggest an alternative action to a user who has just executed an action that you think as unintended, which one of the following issues do you consider and in what order of significance? –Object or command similarity (an object is a file or a folder and a command is for example, cut or copy). –The user’s error frequency (e.g. The user has repeatedly selected an unintended command, which is neighbouring to the one s/he meant.) –In case of error diagnosis, whether it is the most common user’s mistake related to the other types of error he commits.