Evaluation How do we test the interaction design? Several Dimensions

Slides:



Advertisements
Similar presentations
©2011 1www.id-book.com An evaluation framework Chapter 13.
Advertisements

Chapter 13: An evaluation framework
Chapter 14: Usability testing and field studies
Evaluation of User Interface Design
Imran Hussain University of Management and Technology (UMT)
Data gathering. Overview Four key issues of data gathering Data recording Interviews Questionnaires Observation Choosing and combining techniques.
6.811 / PPAT: Principles and Practice of Assistive Technology Wednesday, 16 October 2013 Prof. Rob Miller Today: User Testing.
CS305: HCI in SW Development Evaluation (Return to…)
WHAT IS INTERACTION DESIGN?
Chapter 14: Usability testing and field studies. 2 FJK User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept.
Observing Users Paul Bogen, LaShon Johnson, Jehoon Park.
Data gathering.
Chapter 14: Usability testing and field studies. Usability Testing Emphasizes the property of being usable Key Components –User Pre-Test –User Test –User.
1 User Centered Design and Evaluation. 2 Overview Why involve users at all? What is a user-centered approach? Evaluation strategies Examples from “Snap-Together.
Evaluation. formative 4 There are many times throughout the lifecycle of a software development that a designer needs answers to questions that check.
Empirical Methods in Human- Computer Interaction.
Evaluation Methodologies
User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept. Cal Poly San Luis Obispo FJK 2005.
Observing users.
Observing users. The aims Discuss the benefits & challenges of different types of observation. Describe how to observe as an on-looker, a participant,
An evaluation framework
Evaluation Methods April 20, 2005 Tara Matthews CS 160.
Usability 2004 J T Burns1 Usability & Usability Engineering.
An evaluation framework
From Controlled to Natural Settings
ICS 463, Intro to Human Computer Interaction Design: 8. Evaluation and Data Dan Suthers.
Chapter 14: Usability testing and field studies
Evaluation Framework Prevention vs. Intervention CHONG POH WAN 21 JUNE 2011.
Chapter 11: An Evaluation Framework Group 4: Tony Masi, Sam Esswein, Brian Rood, & Chris Troisi.
Data gathering. Overview Four key issues of data gathering Data recording Interviews Questionnaires Observation Choosing and combining techniques.
Multimedia Specification Design and Production 2013 / Semester 1 / week 9 Lecturer: Dr. Nikos Gazepidis
Data Collection Methods
Chapter 12 Observing Users Li, Jia Li, Wei. Outline What and when to observe Approaches to observation How to observe How to collect data Indirect observation.
Human Computer Interaction
Evaluation Paradigms & Techniques IS 588 Spring 2008 Dr. D. Bilal.
Chapter 12/13: Evaluation/Decide Framework Question 1.
Level 2 Prepared by: RHR First Prepared on: Nov 23, 2006 Last Modified on: Quality checked by: MOH Copyright 2004 Asia Pacific Institute of Information.
Chapter 13. Reviewing, Evaluating, and Testing © 2010 by Bedford/St. Martin's1 Usability relates to five factors of use: ease of learning efficiency of.
Quantitative and Qualitative Approaches
Chapter 15 Qualitative Data Collection Gay, Mills, and Airasian
Virtual University - Human Computer Interaction 1 © Imran Hussain | UMT Imran Hussain University of Management and Technology (UMT) Lecture 40 Observing.
© An Evaluation Framework Chapter 13.
CSCI 4163 / CSCI 6904 – Winter Housekeeping  Clarification about due date for reading comments/questions  Skills sheet  Active listening handout.
Chapter 12/13: Evaluation/Decide Framework. Why Evaluate? Why: to check that users can use the product and that they like it. Designers need to check.
Computer Science Department California Polytechnic State University San Luis Obispo, CA, U.S.A. Franz J. Kurfess CPE/CSC 484: User-Centered Design and.
Marketing Research Approaches. Research Approaches Observational Research Ethnographic Research Survey Research Experimental Research.
Observing users. The aims Discuss the benefits & challenges of different types of observation. Describe how to observe as an on-looker, a participant,
Usability Evaluation, part 2. REVIEW: A Test Plan Checklist, 1 Goal of the test? Specific questions you want to answer? Who will be the experimenter?
EVALUATION PROfessional network of Master’s degrees in Informatics as a Second Competence – PROMIS ( TEMPUS FR-TEMPUS-JPCR)
AVI/Psych 358/IE 340: Human Factors Data Gathering October 6, 2008.
AVI/Psych 358/IE 340: Human Factors Data Gathering October 3, 2008.
Evaluation Methods - Summary. How to chose a method? Stage of study – formative, iterative, summative Pros & cons Metrics – depends on what you want to.
AVI/Psych 358/IE 340: Human Factors Evaluation October 31, 2008.
DECIDE: An evaluation framework. DECIDE: a framework to guide evaluation D D etermine the goals. E E xplore the questions. C C hoose the evaluation approach.
Observing users. What and when to observe Goals & questions determine the paradigms and techniques used. Observation is valuable any time during design.
Um ambiente para avaliação. Objetivos Explicar conceitos e termos da avaliação Descrever paradigmas de avaliação e técnicas utilizadas no design de.
Oct 211 The next two weeks Oct 21 & 23: Lectures on user interface evaluation Oct 28: Lecture by Dr. Maurice Masliah No office hours (out of town) Oct.
Data gathering (Chapter 7 Interaction Design Text)
Chapter 13: An evaluation framework. The aims are: To discuss the conceptual, practical and ethical issues involved in evaluation. To introduce and explain.
SIE 515 Design Evaluation Lecture 7.
Lecture3 Data Gathering 1.
Introducing Evaluation
From Controlled to Natural Settings
Evaluation Paradigms & Techniques
Observing users.
From Controlled to Natural Settings
Human-Computer Interaction: Overview of User Studies
COMP444 Human Computer Interaction Evaluation
Presentation transcript:

Evaluation How do we test the interaction design? Several Dimensions Qualitative vs. Quantitative assessments Conceptual vs. Physical Design

Why Evaluate Five good reasons: Problems fixed before product is shipped Team can concentrate on real (not imaginary) problems Engineers can develop code instead of debating about their personal preferences Time to market is sharply reduced Solid, tested design will sell better

When to Evaluate Formative Evaluations Summative Evaluations Conducted during requirements specification and design Consider alternatives Summative Evaluations Assess the success of a finished product Determine whether product satisfies requirements

What to Evaluate A huge variety of User Interaction features can (and should) be evaluated, such as: Sequence of links in a web search Enjoyment experienced by game users System response time Signal detection performance in data analysis Gould’s principles: Focus on users and their tasks Observe, measure, and analyze user performance Design iteratively

Qualitative Assessment Informal Simply ask users how they like the system Listen to “hallway” conversations about systems Formal Develop survey instruments to ask specific questions, e.g. How long did it take you to become comfortable? Which task is the most difficult to accomplish? Hold focus group discussions about system

Quantitative Assessment Identify Usability Criteria (from Requirements) to test Design human performance experiments to test these, e.g. Measure response time or time to complete a task Measure error rate or incidence of “dead end” This can be used during the design process to compare alternative designs

An Evaluation Framework Evaluation must be an intentional, planned process ad hoc evaluations are of very little value The details of the particular framework can vary from team to team What is important is that the framework be crafted in advance, and that all team members understand the framework

Evaluation Paradigms Evaluation Paradigms are the beliefs and practices (perhaps underpinned by theory) that guide a user study We’ll discuss four core evaluation paradigms: Quick and Dirty evaluation Usability Testing Field Studies Predictive Evaluation

Quick and Dirty Informal feedback from users Can be conducted at any stage Emphasis is on speed, not quality Often consultants are used as surrogate users

Usability Testing Measuring typical users’ performance on carefully prepared tasks that are typical for the system Metrics can include such things as Error rate and time to completion Observations/recordings/logs of interaction Questionnaires Strongly controlled by the evaluator

What is usability? An Operational Definition As well as Efficient Effective Safe Easy To learn To remember To use Productive As well as Satisfying Enjoyable Pleasing Motivating Fulfilling

Field Studies Done in natural settings Try to learn what users do and how Artifacts are collected Video, notes, sketches, &c Two approaches: As an outsider looking on Qualitative techniques used to gather data Which may be analyzed qualitatively or quantitatively As an insider Easier to capture role of social environment

Predictive Evaluation Uses models of typical users Heuristic or theoretical Users themselves need not be present Cheaper, faster Tried and true heuristics can be useful E.g. speak the users’ language

Evaluation Techniques Observing users Asking users their opinions Asking experts their opinions Testing users’ performance Modeling users’ task performance to predict efficacy of the interface

Techniques vs. Paradigms Models used to predict efficacy N/A Modeling user’s performance Can measure performance, but difficult Test typical users, typical tasks User testing Heuristics early in design Provide critiques Asking experts May interview Questionnaires & interviews Discussions, focus groups Asking users Central technique Video and logging See how users behave Observing Users Predictive Field Studies Usability Testing Quick and Dirty Techniques Evaluation Paradigms

DECIDE Determine the overall goals that the evaluation addresses Explore the specific questions to be answered Choose the evaluation paradigm and techniques Identify the practical issues that must be addressed Decide how to deal with the ethical issues Evaluate, interpret, and present the data

Determine the goals What are the high-level goals of the evaluation? Who wants the evaluation and why? Should guide the evaluation, i.e.: Check that evaluators understood users’ needs Identify the metaphor under the design Ensure that interface is consistent Investigate degree to which technology influences working practices Identify how the interface of an existing product could be engineered to improve its usability

Explore the questions This amounts to hierarchical question development: “Is the user interface good?” “Is the system easy to learn?” “Are functions easy to find?” “Is the terminology confusing?” “Is response time too slow?” “Is login time too slow?” “Is calculation time too slow?” …

Choose the evaluation paradigm and techniques Choosing one or more evaluation paradigms Can use different paradigms in different stages Can use multiple paradigms in a single stage Combinations of techniques can be used to obtain different perspectives

Identify the practical issues Practical issues must be considered BEFORE beginning any evaluation Users Adequate number of representative users must be found Facilities and equipment How many cameras? Where? Film? Schedule and budget Both always less than would be ideal Expertise Assemble the correct evaluation team

Decide how to deal with ethical issues Experiments involving humans must be conducted within strict ethical guidelines Tell participants the goals and what will happen Explain that personal information is confidential They’re free to stop at any time Pay subjects when possible: formal relationship Avoid using quotes that reveal identity Ask users’ permission to quote them, show them the report Example: Yale shock experiment 1961-2

Evaluate, interpret and present the data What data to collect, how to analyze them Questions need to be asked Reliability: is it reproducible? Validity: measures what it’s supposed to Biases: biases cause distortion Scope: how generalizable? Ecological validity: how important is the evaluation environment – does it match the real environment of interest?

Stopped Here

Observing Users Ethnography – observing the social environment and recording observations which help to understand the function and needs of the people in it Users can be observed in controlled laboratory conditions or in natural environments in which the products are used – i.e. the field

Goals, questions and paradigms Goals and questions should guide all evaluation studies Ideally, these are written down Goals help to guide the observation because there is always so much going on

What and when to observe Insider or outsider? Laboratory or field? Control vs. realism What times are critical times (especially for field observations)?

Approaches to observation Quick and dirty Informal Insider or outsider Observation in usability testing Formal Video, interaction logs, performance data Outsider Observation in field studies Outsider, participant, or ethnographer (participant or not?)

How to observe Techniques of observation and data gathering vary

In controlled environments Decide location Temporary lab in user’s environment? Remote laboratory? Equipment Hard to know what user is thinking “Think Aloud” technique But speaking can alter the interaction Having two subjects work together can help

In the field Who is present? What is happening? What are their roles? What is happening? Include body language, tone When does activity occur? Where is it happening? Why is it happening How is the activity organized?

Participant observation and ethnography In this case, the observer/evaluator must be accepted into the group Honesty about purpose is important both ethically and to gain trust Disagreement in the field about the distinction between ethnography and participant observation Do ethnographers begin with any assumptions?

Data collection Notes plus still camera Audio recording plus still camera Video

Indirect observation: tracking users’ activities Diaries Interaction logging

Analyzing, interpreting, and presenting data Observation produces large quantities of data of various types How to analyze and interpret depends on the research questions first developed

Qualitative analysis to tell a story The ensemble of data (notes, video, diaries, &c) are used to help designers, as a team, understand the users There is much room for evaluator bias in these techniques

Qualitative analysis for categorization A taxonomy can be developed into which user’s behaviors can be placed This can be done by different observers, with the discrepancies used as a measure of observer bias

Quantitative data analysis Observations, interaction logs, and results are gathered and quantified Counted Measured Analysis using statistical reasoning can be used to draw conclusions What is statistical significance? What is a T-test?

Case Study: Evaluating a Data Representation

Feeding the findings back into design Ideally, the design team will participate in post-evaluation discussions of qualitative data Reports to designers should include artifacts, such as quotes, anecdotes, pictures, video clips Depending on the design team, quantitative data may or may not be compelling