CSCD 487/587 Human Computer Interface Winter 2013 Lecture 19 Evaluation.

Slides:



Advertisements
Similar presentations
Introducing evaluation. The aims Discuss how developers cope with real-world constraints. Explain the concepts and terms used to discuss evaluation. Examine.
Advertisements

Chapter 14: Usability testing and field studies
©2011 1www.id-book.com Introducing Evaluation Chapter 12.
Agile Usability Testing Methods
CS305: HCI in SW Development Evaluation (Return to…)
IS214 Recap. IS214 Understanding Users and Their Work –User and task analysis –Ethnographic methods –Site visits: observation, interviews –Contextual.
WHAT IS INTERACTION DESIGN?
User Interface Design Notes p7 T120B pavasario sem.
Chapter 14: Usability testing and field studies. 2 FJK User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept.
SIMS 213: User Interface Design & Development Marti Hearst Tues, Feb 25, 2003.
Usability presented by the OSU Libraries’ u-team.
AJ Brush Richard Anderson
Heuristic Evaluation. Sources for today’s lecture: Professor James Landay: stic-evaluation/heuristic-evaluation.ppt.
Evaluation Methodologies
Heuristic Evaluation IS 485, Professor Matt Thatcher.
Heuristic Evaluation Evaluating with experts. Discount Evaluation Techniques  Basis: Observing users can be time- consuming and expensive Try to predict.
Evaluating with experts
1 CS 430 / INFO 430 Information Retrieval Lecture 24 Usability 2.
SIMS 213: User Interface Design & Development Marti Hearst Tues Feb 13, 2001.
An evaluation framework
Usability 2004 J T Burns1 Usability & Usability Engineering.
Testing and Modeling Users Kristina Winbladh & Ramzi Nasr.
Evaluation: Inspections, Analytics & Models
RESEARCH METHODS IN EDUCATIONAL PSYCHOLOGY
Discount Usability Engineering Marti Hearst (UCB SIMS) SIMS 213, UI Design & Development March 2, 1999.
Review an existing website Usability in Design. to begin with.. Meeting Organization’s objectives and your Usability goals Meeting User’s Needs Complying.
Evaluation in HCI Angela Kessell Oct. 13, Evaluation Heuristic Evaluation Measuring API Usability Methodology Matters: Doing Research in the Behavioral.
1. Learning Outcomes At the end of this lecture, you should be able to: –Define the term “Usability Engineering” –Describe the various steps involved.
©2011 1www.id-book.com Analytical evaluation Chapter 15.
Heuristic Evaluation “Discount” Usability Testing Adapted from material by Marti Hearst, Loren Terveen.
류 현 정류 현 정 Human Computer Interaction Introducing evaluation.
Predictive Evaluation
Chapter 8: Systems analysis and design
Evaluation Framework Prevention vs. Intervention CHONG POH WAN 21 JUNE 2011.
Chapter 11: An Evaluation Framework Group 4: Tony Masi, Sam Esswein, Brian Rood, & Chris Troisi.
Evaluation of Adaptive Web Sites 3954 Doctoral Seminar 1 Evaluation of Adaptive Web Sites Elizabeth LaRue by.
Part 1-Intro; Part 2- Req; Part 3- Design  Chapter 20 Why evaluate the usability of user interface designs?  Chapter 21 Deciding on what you need to.
Multimedia Specification Design and Production 2013 / Semester 1 / week 9 Lecturer: Dr. Nikos Gazepidis
Human Computer Interaction
Usability Evaluation June 8, Why do we need to do usability evaluation?
Click to edit Master subtitle style USABILITY and USER INTERFACE DESIGN Application.
Computer Science Department California Polytechnic State University San Luis Obispo, CA, U.S.A. Franz J. Kurfess CPE/CSC 484: User-Centered Design and.
Introducing Evaluation: why, what, when, where Text p Text p 317 – 323;
SEG3120 User Interfaces Design and Implementation
Level 2 Prepared by: RHR First Prepared on: Nov 23, 2006 Last Modified on: Quality checked by: MOH Copyright 2004 Asia Pacific Institute of Information.
CS2003 Usability Engineering Usability Evaluation Dr Steve Love.
Evaluation of User Interface Design 4. Predictive Evaluation continued Different kinds of predictive evaluation: 1.Inspection methods 2.Usage simulations.
COMP5047 Pervasive Computing: 2012 Think-aloud usability experiments or concurrent verbal accounts Judy Kay CHAI: Computer human adapted interaction research.
Chapter 12: Introducing Evaluation. The aims To illustrate how observation, interviews and questionnaires that you encountered in Chapters 7 and 8 are.
Chapter 12: Introducing Evaluation. The aims To illustrate how observation, interviews and questionnaires that you encountered in Chapters 7 and 8 are.
CSCI 4163 / CSCI 6904 – Winter Housekeeping  Clarification about due date for reading comments/questions  Skills sheet  Active listening handout.
Evaluating a UI Design Expert inspection methods Cognitive Walkthrough
June 5, 2007Mohamad Eid Heuristic Evaluation Chapter 9.
Design Process … and some design inspiration. Course ReCap To make you notice interfaces, good and bad – You’ll never look at doors the same way again.
By Godwin Alemoh. What is usability testing Usability testing: is the process of carrying out experiments to find out specific information about a design.
Oct 211 The next two weeks Oct 21 & 23: Lectures on user interface evaluation Oct 28: Lecture by Dr. Maurice Masliah No office hours (out of town) Oct.
Fall 2002CS/PSY Predictive Evaluation (Evaluation Without Users) Gathering data about usability of a design by a specified group of users for a particular.
Introducing Evaluation Chapter 12. What is Evaluation?  Assessing and judging  Reflecting on what it is to be achieved  Assessing the success  Identifying.
THE MARKETING RESEARCH PROCESS CHAPTER 29 Mrs. Simone Seaton Marketing Management.
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
Introduction to System Evaluation IS 588 Dania Bilal & Lorraine Normore Spring 2007.
SIE 515 Design Evaluation Lecture 7.
Introducing Evaluation
WHAT IS INTERACTION DESIGN?
Usability Techniques Lecture 13.
PostPC Computing Heuristic Evaluation Prof. Scott Kirkpatrick, HUJI
Imran Hussain University of Management and Technology (UMT)
Introducing Evaluation
Evaluation: Inspections, Analytics & Models
Presentation transcript:

CSCD 487/587 Human Computer Interface Winter 2013 Lecture 19 Evaluation

Overview Evaluation is integral to the design process Evaluators collect information about users' or potential users' experiences when interacting with your interface They do this in order to improve its design Evaluation focuses on both usability of system – How easy it is to learn and to use and On users' experience when interacting with system – How satisfying, enjoyable, or motivating

Overview Many different evaluation methods Which to use depends on the goals of the evaluation Evaluations can occur in a range of places such as laboratories, people's homes, outdoors, and work settings Evaluations usually involve observing participants and measuring their performance – In usability testing, experiments, or field studies There are other methods, however, that do not involve participants, such as modeling user behavior Tend to be approximations of what users might do when interacting with an interface … not measuring users

Why You Need to Evaluate Designers should not presume that everyone is like them or that following set guidelines would guarantee them good usability Evaluation is needed to check that users can use product and like it

Why You Need to Evaluate Bruce Tognazzini – Usability Consultant “Iterative design, with its repeating cycle of design and testing, is the only validated methodology in existence that will consistently produce successful results If you don’t have user-testing as a part of your design process you are going to throw buckets of money down the drain”

Why You Need to Evaluate Toganazzini’s 5 resons to evaluate 1. Problems are fixed before product is shipped, not after 2. The team can concentrate on real problems, not imaginary ones 3. Engineers code is sharply reduced 4. Time to market is sharply reduced 5. Upon 1 st release, your sales department has a rock solid design it can sell without having to pepper their pitches with how well next release will work

Why You Need to Evaluate Evaluation as part of the Design Process design implementationevaluation

What to Evaluate It’s a Continuous Process Users and their tasks Observe, measure, and analyze their performance with the system Basically anything for which we need feedback Important for designer to check and make sure that they understand user requirements

Where Do We Evaluate Laboratories – Formal experiments On-line – Gather feedback from on-line user statistics Homes – Users can provide feedback in a natural setting

When to Evaluate Depends on what we are evaluating Requirements Gather them, evaluate, Did we capture them? Iterate Evaluate again May do this several times Design Evaluate as develop interface and product Known as Formative Evaluation Product Completion Evaluate to assess user satisfaction Known as Sumative Evaluation

When to Evaluate New product Use mockups, sketches, and other low fidelity prototyping techniques are used to represent design ideas Upgrading existing products Compare user performance and attitudes and contrast new products with the previous versions Evaluation is a key ingredient for a successful design

Evaluation Types 1. Controlled Settings Labs or Living Labs User's activities are controlled Do usability testing and experiments 2. Natural Settings On-line communities and public places No control over user activities Field studies

Evaluation Types 3. Settings Not Involving Users Experts and consultants research Predict and model aspects of interfaces You Choose a method based on …. Deciding how much control is needed to find out how an interface is used

Controlled Settings Evaluators can control what users do Also reduce outside influences and distractions Usability Testing Purpose - Evaluate User Interfaces Experiments, observations, interviews and questionnaires In a controlled setting Goal Determine whether an interface is usable by intended population

Controlled Settings Usability Testing continued Has been done for years as part of Human Computer Interaction Findings of test summarized in a usability specification Changes in design can be specified according to usability spec Example Paper prototype session viewed in class one example

Natural Settings Evaluate People in Natural Settings Purpose Identify opportunities for new technology Establish requirements for a new design New context for existing product Field Studies Can be in person or virtual On-line communities Chat rooms Goals Examine social processes, collaborations and cooperation between users

Natural Settings Field Studies While techniques, such as focus groups, usability tests, and surveys, can lead to valuable insights, most powerful tool in the toolbox is the 'field study' Field studies get teams immersed in environment of their users and allow them to observe critical details for which there is no other way of discovering Examples: Watch users in shopping malls, watched system administrators, watched paper flow through manufacturing facility What do you think is biggest downside ? Biggest downside to field studies is the cost to the organization, Scheduling the visits, taking team members out of the office for several days, and finishing the analysis can have a huge impact on a project's resources Measuring man or dog?

Field Studies The Power of Field Studies Even a short field study, such as two or three half-day visits, can yield tremendous value. From these we can learn: Terminology and processes – What do users do and how do they talk about it? – While users can describe a process or share terms in an interview format, watching them work points out subtleties that they are unaware of

Field Studies The Power of Field Studies Even a short field study, such as two or three half- day visits, can yield tremendous value. From these we can learn: Context – What are the external forces that will impact the design? – Do the user's requirements change when they are rushed or up against a deadline? – People have trouble describing the context of their work, however it's easy for outsiders to observe

Field Studies The Power of Field Studies Even a short field study, such as two or three half- day visits, can yield tremendous value. From these we can learn: Similarities and differences – Visiting multiple sites can allow a team to collect a rich amount of information about commonalities that appear across environments, along with variations that will impact design decisions – Such as providing switches, options, and optional features

Research Not with Users Researcher has to model how interface is likely to be used Predict user behavior Identify usability problems Example Heuristic evaluation of an interface Applies knowledge of user behavior Context system will be used Steps through a scenario and answers set of questions

Heuristic Evaluation Originally used in evaluation of screen-based applications Heuristic evaluation is discount usability engineering method for quick, cheap, and easy evaluation of a user interface design Heuristic evaluation is most popular usability inspection method Heuristic evaluation is done as systematic inspection of a user interface design for usability

Heuristic Evaluation Jakob Nielsen Most usability engineering methods will contribute substantially to the usability of an interface … …if they are actually used.

Heuristic Evaluation Goal of heuristic evaluation is to find usability problems in design so that they can be integrated into iterative design process Heuristic evaluation involves having a small set of evaluators examine interface and judge its compliance with recognized usability principles (the "heuristics") Dates back to the 1990's

Heuristic Evaluation How does it work? Evaluators use a checklist of basic usability heuristics Evaluators go through an interface twice 1 st Pass get a feel for the flow and general scope 2 nd Pass refer to checklist of usability heuristics and focus on individual elements The findings of evaluators are combined and assessed evaluation/

Heuristic Evaluation Usability Heuristics (original, unrevised list) Simple and natural dialogue Speak the users’ language Minimize the users’ memory load Consistency Feedback Clearly marked exits Shortcuts Precise and constructive error messages Prevent errors Help and documentation COMMENTS?

Heuristic Evaluation Debriefing session Conducted in brain-storming mode Evaluators rate the severity of all problems identified Use a 0 – 4, absolute scale 0 I don’t agree that this is a prob at all 1 Cosmetic prob only 2 Minor prob – low priority 3 Major prob – high priority 4 Usability catastrophe – imperative to fix

Heuristic Evaluation How does H.E. Differ from User Testing? Evaluators have checklists Evaluators are not target users Evaluators decide on their own how they want to proceed Observer can answer evaluators’ questions about the domain or give hints for using the interface Evaluators say what they didn’t like and why; observer doesn’t interpret evaluators’ actions

Heuristic Evaluation What are the shortcomings of H.E.? Identifies usability problems without indicating how they are to be fixed “Ideas for appropriate redesigns have to appear magically in the heads of designers on the basis of their sheer creative powers.” Cannot expect it to address all usability issues when evaluators are not domain experts / actual users Have a short video of Heuristic Evaluation …

On-line Analytics Web Analytics Measurement, collection and reporting of user behavior Understand and optimize web usage Tools help analyze log files of interaction data and web traffic analytics-expert-heuristic-evaluations/

Web Analytics What do they tell you? Data such as – Number of visitors to a website, – Where they are from, – Which pages they view, and – Which links they click Measured, collected, analyzed and reported Web analytics are relevant to usability practitioners in that they can provide insight into the large-scale behavior of website users to understand and improve (optimize) the website What wouldn't they tell you?

Web Analytics What wouldn't they tell you? Web analytics cannot provide answers – About user motivations or underlying needs and goals – Web analytics may indicate that users are abandoning a checkout process at a particular point, but they cannot be used to explain why this is happening – Usability testing of issues found through web analytics brings deeper understanding needed to fix these usability problems – Additionally, in-person observations of users can lead to insight that informs what metrics are worthwhile to collect, and how to interpret them.

Summary User evaluation is a critical part of user centered design Uses many of the techniques we already discussed Plus, can use Heuristics as a quick and cheap way to explore user interface problems Web statistics through website analytics can provide valuable data but is no substitute for more in-depth user testing

34 End Two more Weeks of Classes !!!