1 User Centered Design and Evaluation. 2 Overview My evaluation experience Why involve users at all? What is a user-centered approach? Evaluation strategies.

Slides:



Advertisements
Similar presentations
Chapter 14: Usability testing and field studies
Advertisements

Chapter 4 Design Approaches and Methods
CS305: HCI in SW Development Evaluation (Return to…)
Saul Greenberg User Centered Design Why User Centered Design is important Approaches to User Centered Design.
Chapter 14: Usability testing and field studies. 2 FJK User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept.
User Testing & Experiments. Objectives Explain the process of running a user testing or experiment session. Describe evaluation scripts and pilot tests.
USABILITY AND EVALUATION Motivations and Methods.
The art and science of measuring people l Reliability l Validity l Operationalizing.
Semester in review. The Final May 7, 6:30pm – 9:45 pm Closed book, ONE PAGE OF NOTES Cumulative Similar format to midterm (probably about 25% longer)
Methodology Overview Dr. Saul Greenberg John Kelleher.
Writing a Research Protocol Michael Aronica MD Program Director Internal Medicine-Pediatrics.
Evaluation Adam Bodnar CPSC 533C Monday, April 5, 2004.
1 User Centered Design and Evaluation. 2 Overview Why involve users at all? What is a user-centered approach? Evaluation strategies Examples from “Snap-Together.
Empirical Methods in Human- Computer Interaction.
Midterm Exam Review IS 485, Professor Matt Thatcher.
Evaluation Methodologies
Inspection Methods. Inspection methods Heuristic evaluation Guidelines review Consistency inspections Standards inspections Features inspection Cognitive.
Design and Evaluation of Iterative Systems n For most interactive systems, the ‘design it right first’ approach is not useful. n The 3 basic steps in the.
Evaluation Methods April 20, 2005 Tara Matthews CS 160.
ICS 463, Intro to Human Computer Interaction Design: 9. Experiments Dan Suthers.
Fig Theory construction. A good theory will generate a host of testable hypotheses. In a typical study, only one or a few of these hypotheses can.
SBD: Usability Evaluation
류 현 정류 현 정 Human Computer Interaction Introducing evaluation.
Evaluation: Controlled Experiments Chris North cs3724: HCI.
Chapter 1: Introduction to Statistics
RESEARCH A systematic quest for undiscovered truth A way of thinking
Predictive Evaluation
Ch 14. Testing & modeling users
Evaluating a Research Report
Human Computer Interaction
Usability testing. Goals & questions focus on how well users perform tasks with the product. – typical users – doing typical tasks. Comparison of products.
Usability Evaluation June 8, Why do we need to do usability evaluation?
Today: Our process Assignment 3 Q&A Concept of Control Reading: Framework for Hybrid Experiments Sampling If time, get a start on True Experiments: Single-Factor.
Testing & modeling users. The aims Describe how to do user testing. Discuss the differences between user testing, usability testing and research experiments.
Assumes that events are governed by some lawful order
Midterm Stats Min: 16/38 (42%) Max: 36.5/38 (96%) Average: 29.5/36 (78%)
Evaluation of User Interface Design 4. Predictive Evaluation continued Different kinds of predictive evaluation: 1.Inspection methods 2.Usage simulations.
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
Review of Research Methods. Overview of the Research Process I. Develop a research question II. Develop a hypothesis III. Choose a research design IV.
The product of evaluation is knowledge. This could be knowledge about a design, knowledge about the user or knowledge about the task.
Usability Engineering Dr. Dania Bilal IS 582 Spring 2006.
1 MP2 Experimental Design Review HCI W2014 Acknowledgement: Much of the material in this lecture is based on material prepared for similar courses by Saul.
SBD: Analyzing Requirements Chris North CS 3724: HCI.
Usability Evaluation, part 2. REVIEW: A Test Plan Checklist, 1 Goal of the test? Specific questions you want to answer? Who will be the experimenter?
1 Human-Computer Interaction Usability Evaluation: 2 Expert and Empirical Methods.
EVALUATION PROfessional network of Master’s degrees in Informatics as a Second Competence – PROMIS ( TEMPUS FR-TEMPUS-JPCR)
Evaluation Methods - Summary. How to chose a method? Stage of study – formative, iterative, summative Pros & cons Metrics – depends on what you want to.
Chapter 15: Analytical evaluation. Aims: Describe inspection methods. Show how heuristic evaluation can be adapted to evaluate different products. Explain.
Introduction to Validity True Experiment – searching for causality What effect does the I.V. have on the D.V. Correlation Design – searching for an association.
Empirical Evaluation Chris North cs5984: Information Visualization.
How Psychologists Do Research Chapter 2. How Psychologists Do Research What makes psychological research scientific? Research Methods Descriptive studies.
Evaluation / Usability. ImplementDesignAnalysisEvaluateDevelop ADDIE.
PSYCH 610 Entire Course (UOP) For more course tutorials visit  PSYCH 610 Week 1 Individual Assignment Research Studies Questionnaire.
User-centered approaches to interaction design By Haiying Deng Yan Zhu.
School of Engineering and Information and Communication Technology KIT305/607 Mobile Application Development Week 7: Usability (think-alouds) Dr. Rainer.
Moshe Banai, PhD Editor International Studies of Management and Organization 1.
User Interface Evaluation
SIE 515 Design Evaluation Lecture 7.
Understanding Results
Usability Evaluation, part 2
SBD: Analyzing Requirements
Evaluation techniques
Scientific Method Steps
Evaluation.
HCI Evaluation Techniques
CSM18 Usability Engineering
Testing & modeling users
Learning about your users (cont.): The field interview
Presentation transcript:

1 User Centered Design and Evaluation

2 Overview My evaluation experience Why involve users at all? What is a user-centered approach? Evaluation strategies Examples from “Snap-Together Visualization” paper

3 Empirical comparison of 2D, 3D, and 2D/3D combinations for spatial data

4 Development and evaluation of a Volume visualization interface

5 Collaborative visualization on a tabletop

6 Why involve users?

7 Understand the users and their problems Visualization users are experts We do not understand their tasks and information needs Intuition is not good enough Expectation management & Ownership Ensure users have realistic expectations Make the users active stakeholders

8 Early focus on users and tasks Empirical measurement: users’ reactions and performance with prototypes Iterative design What is a user-centered approach?

9 Focus on Tasks Users’ tasks / goals are the driving force –Different tasks require very different visualizations –Lists of common visualization tasks can help Shneiderman’s “Task by Data Type Taxonomy” Amar, Eagan, and Stasko (InfoVis05) –But user-specific tasks are still the best

10 Focus on Users Users’ characteristics and context of use need to be supported Users have varied needs and experience –E.g. radiologists vs. GPs vs. patients

11 Understanding users’ work Field Studies - May involve observation, interviewing - At user’s workplace Surveys Meetings / collaboration

12 Design cycle Design should be iterative –Prototype, test, prototype, test, … –Test with users! Design may be participatory

13 Key point Visualizations must support specific users doing specific tasks “Showing the data” is not enough!

14 Evaluation

15 How to evaluate with users? Quantitative Experiments Clear conclusions, but limited realism Qualitative Methods –Observations –Contextual inquiry –Field studies More realistic, but conclusions less precise

16 How to evaluate without users? Heuristic evaluation Cognitive walkthrough –Hard – tasks ill-defined & may be accomplished many ways Allendoerfer et al. (InfoVis05) address this issue GOMS / User Modeling? –Hard – designed to test repetitive behaviour

17 Types of Evaluation (Plaisant) Compare design elements –E.g., coordination vs. no coordination (North & Shneiderman) Compare systems –E.g., Spotfire vs. TableLens Usability evaluation of a system –E.g., Snap system (N & S) Case studies –Real users in real settings E.g., bioinformatics, E-commerce, security

18 Snap-Together Vis Custom coordinated views

19 Questions Is this system usable? –Usability testing Is coordination important? Does it improve performance? –Experiment to compare coordination vs. no coordination

20 Usability testing vs. Experiment Usability testing Aim: improve products Few participants Results inform design Not perfectly replicable Partially controlled conditions Results reported to developers Quantitative Experiment Aim: discover knowledge Many participants Results validated statistically Replicable Strongly controlled conditions Scientific paper reports results to community

21 Usability of Snap-Together Vis Can people use the Snap system to construct a coordinated visualization? Not really a research question But necessary if we want to use the system to answer research questions How would you test this?

22 Critique of Snap-Together Vis Usability Testing + Focus on qualitative results + Report problems in detail + Suggest design changes -Did not evaluate how much training is needed (one of their objectives) Results useful mainly to developers

23 Summary: Usability testing Goals focus on how well users perform tasks with the prototype May compare products or prototypes Techniques: –Time to complete task & number & type of errors (quantitative performance data) –Qualitative methods (questionnaires, observations, interviews) –Video/audio for record keeping

24 Controlled experiments Strives for –Testable hypothesis –Control of variables and conditions –Generalizable results –Confidence in results (statistics)

25 Testable hypothesis State a testable hypothesis –this is a precise problem statement Example: –(BAD) 2D is better than 3D –(GOOD) Searching for a graphic item among 100 randomly placed similar items will take longer with a 3D perspective display than with a 2D display.

26 Controlled conditions Purpose: Knowing the cause of a difference found in an experiment –No difference between conditions except the ideas being studied Trade-off between control and generalizable results

27 Confounding Factors (1) Group 1 Visualization A in a room with windows Group 2 Visualization B in a room without windows What can you conclude if Group 2 performs the task faster?

28 Confounding Factors (2) Participants perform tasks with Visualization A followed by Visualization B. What can we conclude if task time is faster with Visualization A?

29 Confounding Factors (3) Do people remember information better with 3D or 2D displays? Participants randomly assigned to 2D or 3D Instructions and experimental conditions the same for all participants Tavanti and Lind (Infovis 2001)

30 What are the confounding factors? 2D Visualization3D Visualization

31 What is controlled Who gets what condition –Subjects randomly assigned to groups When & where each condition is given How the condition is given –Consistent Instructions –Avoid actions that bias results (e.g., “Here is the system I developed. I think you’ll find it much better than the one you just tried.”) Order effects

32 Order Effects Example: Search for circles among squares and triangles in Visualizations A and B 1.Randomization E.g., number of distractors: 3, 15, 6, 12, 9, 6, 3, 15, 9, 12… 2.Counter-balancing E.g., Half use Vis A 1 st, half use Vis B first

33 Experimental Designs Between- subjects Within- subjects No order effects? +- Participants can compare conditions? -+ Number of participants ManyFew

34 Statistical analysis Apply statistical methods to data analysis –confidence limits : the confidence that your conclusion is correct “p = 0.05” means: –a 95% probability that there is a true difference –a 5% probability the difference occurred by chance

35 Types of statistical tests T-tests (compare 2 conditions) ANOVA (compare >2 conditions) Correlation and regression Many others

36 Snap-Together Vis Experiment Are both coordination AND visual overview important in overview + detail displays? How would you test this?

37 Critique of Snap-Together Vis Experiment + Carefully designed to focus on factors of interest - Limited generalizability. Would we get the same result with non-text data? Expert users? Other types of coordination? Complex displays? - Unexciting hypothesis – we were fairly sure what the answer would be

38 How should evaluation change? Better experimental design –Especially more meaningful tasks Fewer “Compare time on two systems” experiments Qualitative methods Field studies with real users

39 Take home messages Talk to real users! Learn more about HCI!