Ch 13. Asking Users & Experts Team 3:Jessica Herron Lauren Sullivan Chris Moore Steven Pautz.

Slides:



Advertisements
Similar presentations
Chapter 15: Analytical evaluation
Advertisements

Chapter 7 Data Gathering 1.
Data gathering. Overview Four key issues of data gathering Data recording Interviews Questionnaires Observation Choosing and combining techniques.
Asking Users and Experts
ACTIVELY ENGAGING THE STAKEHOLDER IN DEFINING REQUIREMENTS FOR THE BUSINESS, THE STAKEHOLDER, SOLUTION OR TRANSITION Requirements Elicitation.
CS305: HCI in SW Development Continuing Evaluation: Asking Experts Inspections and walkthroughs.
Evaluation (cont.): Heuristic Evaluation Cognitive Walkthrough CS352.
Data gathering.
Asking users & experts.
Chapter 15: Analytical evaluation. 2 FJK User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept. Cal Poly San.
Asking users & experts. The aims Discuss the role of interviews & questionnaires in evaluation. Teach basic questionnaire design. Describe how do interviews,
Heuristic Evaluation. Sources for today’s lecture: Professor James Landay: stic-evaluation/heuristic-evaluation.ppt.
1 User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept. Cal Poly San Luis Obispo FJK 2009.
Heuristic Evaluation IS 485, Professor Matt Thatcher.
Heuristic Evaluation Evaluating with experts. Discount Evaluation Techniques  Basis: Observing users can be time- consuming and expensive Try to predict.
Evaluation Through Expert Analysis U U U
Evaluating with experts
Asking Users and Experts Bobby Kotzev Adrian Sugandhi.
Usability 2004 J T Burns1 Usability & Usability Engineering.
Asking users & experts The aims Discuss the role of interviews & questionnaires in evaluation. Teach basic questionnaire design. Describe how do interviews,
Asking users & experts. Interviews Unstructured - are not directed by a script. Rich but not replicable. Structured - are tightly scripted, often like.
Evaluation: Inspections, Analytics & Models
Chapter 7 GATHERING DATA.
FOCUS GROUPS & INTERVIEWS
1 User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept. Cal Poly San Luis Obispo FJK 2005.
Review an existing website Usability in Design. to begin with.. Meeting Organization’s objectives and your Usability goals Meeting User’s Needs Complying.
©2011 1www.id-book.com Analytical evaluation Chapter 15.
‘Hints for Designing Effective Questionnaires ’
Asking Users and Experts Yujia ZHU Yimeng DOU Asking Users l Interviews l Questionnaires.
Human Computer Interface
1 Asking users & experts and Testing & modeling users Ref: Ch
Discount Evaluation Evaluating with experts. Discount Evaluation Techniques Basis: – Observing users can be time-consuming and expensive – Try to predict.
Data gathering. Overview Four key issues of data gathering Data recording Interviews Questionnaires Observation Choosing and combining techniques.
Ch 14. Testing & modeling users
Chapter 7 Data Gathering 1.
Part 1-Intro; Part 2- Req; Part 3- Design  Chapter 20 Why evaluate the usability of user interface designs?  Chapter 21 Deciding on what you need to.
Interviews. Unstructured - are not directed by a script. Rich but not replicable. Structured - are tightly scripted, often like a questionnaire. Replicable.
Multimedia Specification Design and Production 2013 / Semester 1 / week 9 Lecturer: Dr. Nikos Gazepidis
Usability Evaluation June 8, Why do we need to do usability evaluation?
SEG3120 User Interfaces Design and Implementation
CS3205: HCI in SW Development Evaluation (Return to…) We’ve had an introduction to evaluation. Now for more details on…
Chapter 15: Analytical evaluation. Inspections Heuristic evaluation Walkthroughs.
Chapter 15: Analytical evaluation Q1, 2. Inspections Heuristic evaluation Walkthroughs Start Q3 Reviewers tend to use guidelines, heuristics and checklists.
Analytical evaluation Prepared by Dr. Nor Azman Ismail Department of Computer Graphics and Multimedia Faculty of Computer Science & Information System.
CSCI 4163 / CSCI 6904 – Winter Housekeeping  Clarification about due date for reading comments/questions  Skills sheet  Active listening handout.
Evaluating a UI Design Expert inspection methods Cognitive Walkthrough
Usability 1 Usability evaluation Without users - analytical techniques With users - survey and observational techniques.
Questionnaires Questions can be closed or open Closed questions are easier to analyze, and may be done by computer Can be administered to large populations.
AVI/Psych 358/IE 340: Human Factors Data Gathering October 6, 2008.
AVI/Psych 358/IE 340: Human Factors Data Gathering October 3, 2008.
Asking users & experts. The aims Discuss the role of interviews & questionnaires in evaluation. Teach basic questionnaire design. Describe how do interviews,
Chapter 15: Analytical evaluation. Aims: Describe inspection methods. Show how heuristic evaluation can be adapted to evaluate different products. Explain.
Oct 211 The next two weeks Oct 21 & 23: Lectures on user interface evaluation Oct 28: Lecture by Dr. Maurice Masliah No office hours (out of town) Oct.
Fall 2002CS/PSY Predictive Evaluation (Evaluation Without Users) Gathering data about usability of a design by a specified group of users for a particular.
Data gathering (Chapter 7 Interaction Design Text)
Asking Users and Experts Li Zhang (Jacey) Yuewei Zhou (Joanna)
Lecture 4 Supplement – Data Gathering Sampath Jayarathna Cal Poly Pomona Based on slides created by Ian Sommerville & Gary Kimura 1.
Chapter 7 GATHERING DATA.
SIE 515 Design Evaluation Lecture 7.
Human Computer Interaction Lecture 15 Usability Evaluation
Imran Hussain University of Management and Technology (UMT)
Lecture3 Data Gathering 1.
CS3205: HCI in SW Development Evaluation (Return to…)
Chapter 7 Data Gathering 1.
Chapter 7 GATHERING DATA.
GATHERING DATA.
Chapter 7 GATHERING DATA.
Evaluation.
Evaluation: Inspections, Analytics & Models
Evaluation: Inspections, Analytics, and Models
Presentation transcript:

Ch 13. Asking Users & Experts Team 3:Jessica Herron Lauren Sullivan Chris Moore Steven Pautz

Asking Users and Experts: Introduction We want to find out what users do, want to do, like, or don’t like Several tools & techniques are available Interviews Questionnaires Heuristic Evaluation Cognitive Walkthroughs Strengths/Weaknesses of these tools

Interviews Unstructuredare not directed by a script. Rich but not replicable. Structuredare tightly scripted, often like a questionnaire. Replicable but may lack richness. Semi-structuredguided by a script but interesting issues can be explored in more depth. Can provide a good balance between richness and replicability.

Interview Basics Remember the DECIDE framework –Determine Goals, Explore Questions, Choose eval paradigm, Identify issues, Decide on ethical issues, and Evaluate data. Goals and questions guide all interviews Two Types of Questions: Closed Questions Closed questions have a predetermined answer format, e.g., ‘yes’ or ‘no’. Closed questions are quicker and easier to analyze. Open Questions Open questions do not have a predetermined format, they are trying to determine the feeling of the interviewee.

Avoid Interview Pitfalls  Long questions  Compound sentences - split into two  Jargon & language that the interviewee may not understand  Leading questions that make assumptions e.g., why do you like …? They might not actually like the product.  Be aware of unconscious biases  e.g., gender stereotypes

Interview Components Introductionintroduce yourself, explain the goals of the interview, reassure about the ethical issues, ask to record, present an informed consent form. Warm-Upmake first questions easy & non-threatening. Main Bodypresent questions in a logical order Cool Offinclude a few easy questions to defuse tension at the end Closurethank interviewee, signal the end, e.g, switch recorder off.

The Interview Process Dress in a similar way to participants Check recording equipment in advance Devise a system for coding names of participants to preserve confidentiality. Be pleasant Ask participants to complete an informed consent form

Group interviews Also known as focus groups Typically 3-10 participants Provide a diverse range of opinions Need to be managed to: - ensure everyone contributes - discussion isn’t dominated by one person - the agenda of topics is covered, eliminate digressions.

Other Sources Telephone Interview –Can’t see participant’s body language or nonverbal clues. Online Interviews –Asynchronous: or Synchronous: Chat Room Retrospective Interview –Check with the participant to be sure interviewer correctly understood their remarks.

Analyzing interview data Depends on the type of interview Structured interviews can be analyzed like questionnaires Unstructured interviews generate data like that from participant observation It is best to analyze unstructured interviews as soon as possible to identify topics and themes from the data

Links UPA – (Usability Professionals Association)UPA –Discovering User Needs Discovering User Needs Sage Services - User Research & Usability Techniques Available Through Sage.Sage Services

Questionnaires Questions can be closed or open Closed questions are easiest to analyze, and may be done by computer Can be administered to large populations Paper, & the web used for dissemination Advantage of electronic questionnaires is that data goes into a data base & is easy to analyze Sampling can be a problem when the size of a population is unknown as is common online

Questionnaire Style Questionnaire format can include: - ‘yes’, ‘no’ checkboxes - checkboxes that offer many options - Likert rating scales - semantic scales - open-ended responses

Designing Questionnaires Make questions clear and specific. When possible, ask closed questions and offer a range of answers. Consider including a ‘no-opinion’ option for questions that seeks opinions. Think about the ordering of questions. Avoid complex multiple questions.

Designing Questionnaires When scales are used, make sure the range is appropriate and does not overlap. Make sure the ordering of scales is intuitive and consistent. Avoid jargon. Provide clear instructions on how to complete the questionnaire. Make a balance between white space and the need to keep the questionnaire compact.

Encouraging a good response Ensure the questionnaire is well designed so that participants do not get annoyed. Provide a short overview section. Include a stamped, self-addressed envelope. Contact respondents through a follow-up letter. Offer incentives. Explain why you need the questionnaire to be completed.

Online Questionnaires Advantages: –Responses received quickly. –Copying and postage costs lower. –Data can be transferred immediately to database. –Time required is reduced. –Errors in questionnaire can be corrected easily.

Online Questionnaires Disadvantages: –Sampling is problematic if population size is unknown. –Preventing individuals from responding more than once. –Individuals have also been known to change questions in questionnaires

Developing Web-based Questionnaire Steps: 1.Produce an error-free interactive electronic version from the original paper-based one. 2.Make the questionnaire accessible from all common browsers and readable for different sized monitors and network locations. 3.Make sure information identifying each respondent will be captured and stored confidentially. 4.User-test the survey with pilot studies before distributing.

Asking experts Experts use their knowledge of users & technology to review software usability Expert critiques (crits) can be formal or informal reports Heuristic evaluation is a review guided by a set of heuristics Walkthroughs involve stepping through a pre- planned scenario noting potential problems

Heuristic evaluation Developed by Jacob Nielsen in the early 1990s Based on heuristics distilled from an empirical analysis of 249 usability problems These heuristics have been revised for current technology, e.g., HOMERUN for web Heuristics still needed for mobile devices, wearables, virtual worlds, etc. Design guidelines form a basis for developing heuristics

Nielsen’s heuristics Visibility of system status Match between system and real world User control and freedom Consistency and standards Help users recognize, diagnose, recover from errors Error prevention Recognition rather than recall Flexibility and efficiency of use Aesthetic and minimalist design Help and documentation

HOMERUN High-quality content Often updated Minimal download time Ease of use Relevant to users’ needs Unique to the online medium Netcentric corporate culture

Discount evaluation Heuristic evaluation is referred to as discount evaluation when 5 evaluators are used. Empirical evidence suggests that on average 5 evaluators identify 75-80% of usability problems.

3 stages for doing heuristic evaluation Briefing session to tell experts what to do Evaluation period of 1-2 hours in which: - Each expert works separately - Take one pass to get a feel for the product - Take a second pass to focus on specific features Debriefing session in which experts work together to prioritize problems

Heuristic evaluation of websites MEDLINEplus – medical information websiteMEDLINEplus Keith Cogdill, commissioned by NLM, evaluated this website Based evaluation on the following: –Internal consistency –Simple dialog –Shorcuts –Minimizing the user’s memory load –Preventing errors –Feedback –Internal locus of control

MEDLINEplus suggestions for improvement Arrangement of health topics –Arrange topics alphabetically as well as in categories Depth of navigation menu –Having a higher “fan-out” in the navigation would enhance usability…meaning many short menus rather than a few deep ones.

Turning guidelines into heuristics for the web Navigation, access and information design are the three categories for evaluation Navigation –Avoidance of orphan pages –Avoidance long pages with excessive white space (scrolling problems) –Provide navigation support –Avoidance of narrow, deep, hierarchical meus –Avoidance of non-standard link colors –Provide a consistent look and feel

Turning guidelines into heuristics for the web (cont.) Access –Avoidance of complex URLs –Avoid long download times that annoy users Pictures.wav files Large software files Information Design –Have good graphical design –No outdated or incomplete information –Avoid excessive color usage –Avoid gratuitous use of graphics and animation –Be consistent

Advantages and problems Few ethical & practical issues to consider Can be difficult & expensive to find experts Best experts have knowledge of application domain & users Biggest problems - important problems may get missed - many trivial problems are often identified

Walkthroughs Alternative approach to heuristic evaluation Expert walks through task, noting problems in usability Users are usually not involved Several forms: Cognitive Walkthrough Pluralistic Walkthrough

Cognitive Walkthroughs Involves simulating the user’s problem-solving process at each step in the task Focus on ease of learning Examines user problems in detail, without needing users or a working prototype Can be very time-consuming and laborious Somewhat narrow focus Adaptations may offer improvements

Cognitive Walkthrough Steps One or more experts walk through the design description or prototype with a specific task or scenario For each task, they answer three questions: Will users know what to do? Will users see how to do it? Will users understand from feedback whether the action was correct or not?

Cognitive Walkthrough Steps A record of critical information is compiled Assumptions about what would cause problems and why, including explanations for why users would face difficulties Notes about side issues and design changes Summary of results Design is revised to fix problems presented

Pluralistic Walkthroughs Variation of Cognitive Walkthrough Users, developers, and experts work together to step through a scenario of tasks Strong focus on users’ tasks Provides quantitative data May be limited by scheduling difficulties and time constraints

Pluralistic Walkthrough Steps Scenarios are developed in the form of a series of screens Panel members individually determine the sequence of actions they would use to move between screens Everyone discusses their sequence with the rest of the panel The panel moves to the next scenario

Asking Users and Experts: Summary Various techniques exist for getting user opinions Structured, unstructured, and semi-structured Interviews, focus groups, questionnaires Expert evaluations: heuristic evaluation and walkthroughs Various techniques are often combined for added accuracy and usefulness

Interview: Jakob Nielsen “Father of Web Usability” Pioneered heuristic evaluation Demonstrated cost-effective methods for usability testing & engineering Author of several acclaimed design books