Erfaringer med Remote Usability Testing? Jan Stage Professor, PhD Forskningsleder i Informationssystemer (IS)/Human-Computer Interaction (HCI) Aalborg.

Slides:



Advertisements
Similar presentations
Chapter 14: Usability testing and field studies
Advertisements

Methodology and Explanation XX50125 Lecture 1: Part I. Introduction to Evaluation Methods Part 2. Experiments Dr. Danaë Stanton Fraser.
A GUIDE TO CREATING QUALITY ONLINE LEARNING DOING DISTANCE EDUCATION WELL.
Chapter 14: Usability testing and field studies. Usability Testing Emphasizes the property of being usable Key Components –User Pre-Test –User Test –User.
HCI Methods for Pathway Visualization Tools Purvi Saraiya, Chris North, Karen Duca* Virginia Tech Dept. of Computer Science, Center for Human-Computer.
The USE Project: Usability Evaluation and Software Design: Bridging the Gap University of Copenhagen Aalborg University Use Case Evaluation (UCE): A Method.
Job Analysis & Its Components Job analysis: Job analysis: The process of describing and recording many aspects or elements of the job. The outcome of job.
Instant Data Analysis (IDA): Evaluating Usability in a Day Jesper Kjeldskov Mikael B. Skov Jan Stage.
Think-aloud usability experiments or concurrent verbal accounts Judy Kay CHAI: Computer human adapted interaction research group School of Information.
Jesper Kjeldskov & Jan Stage Department of Computer Science Aalborg University Denmark New Techniques for Usability Evaluation of Mobile Systems.
Feedback from Usability Evaluation to User Interface Design: Are Usability Reports Any Good? Christian M. Nielsen 1 Michael Overgaard 2 Michael B. Pedersen.
Jesper Kjeldskov Mikael B. Skov Jan Stage HCI-Lab Department of Computer Science Aalborg University Denmark Does Time Heal? A Longitudinal Study of Usability.
1 1 Slide © 2015 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
From Controlled to Natural Settings
Task analysis 1 © Copyright De Montfort University 1998 All Rights Reserved Task Analysis Preece et al Chapter 7.
1 Assuring the Quality of your COSF Data. 2 What factors work to improve the quality of your data? What factors work to lessen the quality of your data?
Three Hours a Week?: Determining the Time Students Spend in Online Participatory Activity Abbie Brown, Ph.D. East Carolina University Tim Green, Ph.D.
By: Mahmood Sedaghatian  Anticipatory Set  Transformation Geometry  Project Description  Timeline  Resources  Grading  Technology Used  Learning.
© 2004 Keynote Systems Customer Experience Management (CEM) Bonny Brown, Ph.D. Director, Research & Public Services.
Usability Methods: Cognitive Walkthrough & Heuristic Evaluation Dr. Dania Bilal IS 588 Spring 2008 Dr. D. Bilal.
Chapter 5 Formulating the research design
UNIVIRTUAL FOR INSTRUCTIONAL DESIGN Versione 00 del 29/07/2009.
Copyright John Worthington Perceptions of Young Children’s Learning and Computer Use Discussion Seminar Canberra Tuesday 12 March.
Impact of Moodle usage practices on students’ performance in the context of a blended learning environment Filippidi Andromahi,
Introduction to SDLC: System Development Life Cycle Dr. Dania Bilal IS 582 Spring 2009.
Hyun-Myoung Kim* and Young-Soo Kim** Dept. of Biology Education, Seoul National University, Korea Application of Internet Services to Promote.
Sarah Drummond Dept. Computer Science University of Durham, UK MSc Research An Investigation into Computer Support for Cooperative Work in Software Engineering.
RUP Implementation and Testing
© 2015 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Quality Assurance: Looking for Quality Data 1 I know it is in here somewhere Presented by The Early Childhood Outcomes Center Revised January 2013.
Patterns of Usage of Internet Resources by Persons with Post-Coronary Arterial Bypass Graft (CABG) Kamisha Hamilton, Michelle Rogers, Anita Ground, RN,
Multimedia Specification Design and Production 2013 / Semester 1 / week 9 Lecturer: Dr. Nikos Gazepidis
Direct and indirect effects of online learning on distance education 指導教授 : 陳 明 溥 研 究 生 : 許 良 村 Shin, N. & Chan, K.Y.(2004).Direct and indirect effects.
Gruppearbejde Mikael B. Skov Jan Stage. 2 Gruppearbejde Hvad er jeres produkter, og hvad gør I vedr. evaluering? Hvordan kan specifikt evaluere disse?
Human Computer Interaction
Using A Digital Campus to Support Electronic Learning In Lebanon Presenters: Shumin Chuang Professor: Ming-Puu Chen 2008/6/26 王堯興王堯興 Schaik, P., Barker,
What is Usability? Usability Is a measure of how easy it is to use something: –How easy will the use of the software be for a typical user to understand,
©2010 John Wiley and Sons Chapter 6 Research Methods in Human-Computer Interaction Chapter 6- Diaries.
Research PHE 498. Define Research Research can be considered as systematic inquiry: A process that needs to be followed systematically to derive conclusions.
© 2012 Cengage Learning. All Rights Reserved. This edition is intended for use outside of the U.S. only, with content that may be different from the U.S.
COMP5047 Pervasive Computing: 2012 Think-aloud usability experiments or concurrent verbal accounts Judy Kay CHAI: Computer human adapted interaction research.
I Power Higher Computing Software Development The Software Development Process.
Evaluation Plan New Jobs “How to Get New Jobs? Innovative Guidance and Counselling 2 nd Meeting Liverpool | 3 – 4 February L Research Institute Roula.
Population Census Data Dissemination through Internet H. Furuta Lecturer/Statistician SIAP 1 Training Course on Analysis and Dissemination of Population.
Design Process … and some design inspiration. Course ReCap To make you notice interfaces, good and bad – You’ll never look at doors the same way again.
Can social information change behaviour? The results of a study with student & national trust volunteers Facilitator: Ben Lee Presenters: Professor Oliver.
United Nations Oslo City Group on Energy Statistics OG7, Helsinki, Finland October 2012 ESCM Chapter 8: Data Quality and Meta Data 1.
Thomas Kern | The system documentation as binding agent for and in between internal and external customers April 24th, 2009 | Page 1 The system documentation.
Copyright ©2005 by South-Western, a division of Thomson Learning. All rights reserved Chapter 18 1 Controlling: Purpose and Process MANAGEMENT Meeting.
Searching and Using Electronic Literature III. Research Design.
Usability Engineering Dr. Dania Bilal IS 587 Fall 2007.
1 Usability Analysis n Why Analyze n Types of Usability Analysis n Human Subjects Research n Project 3: Heuristic Evaluation.
STAT MINI- PROJECT NKU Executive Doctoral Co-hort August, 2012 Dot Perkins.
Table 1. The Five Medical Laboratory Case Studies Jan Martin et al. Transnational Evaluation of a New Continuing Professional Development Activity for.
Demographic Full Count Review Presentation to the FSCPE March 26, 2001 Washington D.C.
Component D: Activity D.3: Surveys Department EU Twinning Project.
School of Engineering and Information and Communication Technology KIT305/607 Mobile Application Development Week 7: Usability (think-alouds) Dr. Rainer.
Regulative support for collaborative scientific inquiry learning Presenter: Hou, Ming-Hsien Professor: Ming-Puu Chen Date: August 19, 2008 Manlove, S.,
Quality Assurance: Looking for Quality Data
Monitoring and Evaluation Systems for NARS Organisations in Papua New Guinea Day 3. Session 8. Routine monitoring.
RACIAL VIOLENCE ARCHIVE
Evaluation.
Enacted: Generating data in research events
EXAMPLES OF E-LEARNING MATERIALS EIT Meeting – Cyprus September 2007
Involving Families Early Childhood Outcomes Center.
Kasper Hornbæk Department of Computer Science University of Copenhagen
Chapter 10 Problem-Solving in Groups
Embry-Riddle Aeronautical University, College of Aviation
Periodic Workflow Review
Presentation transcript:

Erfaringer med Remote Usability Testing? Jan Stage Professor, PhD Forskningsleder i Informationssystemer (IS)/Human-Computer Interaction (HCI) Aalborg Universitet, Institut for Datalogi, HCI-Lab

Institut for Datalogi 2 Undersøgelse 1 Undersøgelse 2 Oversigt

Institut for Datalogi 3 Undersøgelse 1: synkron eller asynkron Metode Resultater Konklusion Undersøgelse 2 Oversigt

Institut for Datalogi Empirical Study 1 Four methods: LAB – RS – AE – AU Test subjects: 6 in each condition (18 users and 6 with usability expertise), all students at Aalborg University System: client (Mozilla Thunderbird 1.5) 9 defined tasks (typical functions) Setting, procedure and data collection in accordance with method Data analysis: 24 outputs were analysed by three persons in random and different order Generated their individual lists of usability problems with their own categorizations (also for the AE and AU conditions) These were merged into an overall problem list through negotiation 4

Institut for Datalogi 5 Results: Task Completion No significant difference in task completion Significant difference in task completion time The users in the two asynchronous conditions spent considerably more time We do not know the reason

Institut for Datalogi 6 Results: Usability Problems Identified A total of 46 usability problems No significant difference between LAB and RS AE/AU identified significantly fewer problems, also critical problems No significant difference between AE and AU in terms of problems identified

Institut for Datalogi 7 Conclusion RS is the most widely described and used remote method. The performance is virtually equivalent to LAB (or slightly better) AE and AU perform surprisingly well Experts do not perform significantly better than users Video analysis (LAB and RS) required considerably more evaluator effort than the user-based reporting (AU and AE) Users can actually contribute to usability evaluation – not with the same quality, but reasonably well, and there are plenty of them

Institut for Datalogi 8 Undersøgelse 1 Undersøgelse 2: hvilken asynkron metode Metode Resultater Konklusion Oversigt

Institut for Datalogi 9 Empirical Study 2 Purpose: examine and compare remote asynchronous methods Focus on usability problems identified Comparable with the previous study Selection of asynchronous methods based on literature survey

Institut for Datalogi 10 The 3 Remote Asynchronous Methods User-reported critical incident (UCI) Well-defined method (Castillo et al. CHI 1998) Forum-based online reporting and discussion (Forum) Assumption: through collaboration participants may give input which increases data quality and richness (Thompson, 1999) A source for collecting qualitative data in a study of auto logging (Millen, 1999): the participants turned out to report detailed usability feedback Diary-based longitudinal user reporting (Diary) Used on a longitudinal basis for participants in a study of auto logging to provide qualitative information (Steves et al. CSCW 2001) First day: same tasks as the other conditions (first part of diary delivered) Four more days: new tasks (same type) sent daily (complete diary delivered) Conventional user-based laboratory test (Lab) Included as benchmark

Institut for Datalogi 11 Participants: 40 test subjects, 10 for each condition Students, age 20 to 30 Distributed evenly: gender and tech/non-tech education Setting: LAB: in our usability lab Remote asynchronous: in the participants’ homes Participants in the remote asynchronous conditions received the software and installed it on their computer Training material for the remote asynchronous conditions Identification and categorisation of usability problems A minimalist approach that was strictly remote and asynchronous (via ) Empirical Study (1)

Institut for Datalogi 12 Tasks: Nine fixed tasks The same across the four conditions to ensure that all participants used the same parts of the system Typical tasks (same as previous study) Data collection in accordance with the method LAB: video recordings UCI: web-based system for generating problem descriptions while solving tasks Forum: after solving tasks, one week for posting and discussing problems Diary: a diary with no imposed structure; first part after the first day Empirical Study (2)

Institut for Datalogi 13 All data collected before the data analysis started 3 evaluators did the whole data analysis The 40 data sets were analysed by the 3 evaluators In random order: by a draw In different order between them The user input from the three remote conditions was transformed into usability problem descriptions Each evaluator generated his/her own individual lists of usability problems with their own severity ratings A problem list for each condition A complete problem list (joined) These were merged into an overall problem list through negotiation Data Analysis

Institut for Datalogi 14 Results: Task Completion Time Considerable variation in task completion times Participants in the remote conditions worked in their home at a time they selected For each task there was a hint that allowed them to check if they had solved the task correctly As we have no data on the task solving process in the remote conditions, we cannot explain this variation

Institut for Datalogi 15 LAB: significantly better than the 3 remote conditions UCI-Forum: no significant difference UCI-Diary: significant overall: Diary – also significant on cosmetic Forum-Diary: significant overall: Diary – not significant on any level Results: Usability Problems Identified Lab N=10 UCI N=10 Forum N=10 Diary N=10 Task completion time in minutes: Average (SD) (6.3)34.45 (14.33)15.45 (5.83) Tasks 1-9: (28.34) Usability problems:#%#%#%#% Critical (21) Serious (17) Cosmetic (24) Total (62)

Institut for Datalogi 16 Results: Evaluator Effort The sum for all evaluators involved in each activity Time for finding test subjects is not included (8h, common for all) Task specifications from an earlier study. Preparation in the remote conditions: work out written instructions Considerable differences between the remote conditions for analysis and merging of problem lists Lab (46) UCI (13) Forum (15) Diary (29) Preparation6:002:40 Conducting test10:001:00 1:30 Analysis33:182:523:569:38 Merging problem lists11:451:411:424:58 Total time spent61:038:139:1818:46 Avg. time per problem1:200:380:370:39

Institut for Datalogi 17 Conclusion The three remote methods performed significantly below the classical lab test in terms of the number of usability problems identified The Diary was the best remote method – it identified half of the problems found in the Lab condition UCI and Forum performed similarly for critical problems but worse for serious problems UCI and Forum took 13% of the lab test. Diary took 30% The productivity of the remote methods was considerably higher

Institut for Datalogi 18

Institut for Datalogi Interaktionsdesign og usability-evaluering Master i IT Videreuddannelse under IT-Vest Fagpakke i Interaktionsdesign og usability-evaluering starter 1/2-12 Optager bachelorer, men også indgang for datamatikere Information: 19