CyLab Usable Privacy and Security Laboratory 1 C yLab U sable P rivacy and S ecurity Laboratory Designing.

Slides:



Advertisements
Similar presentations
The Usability Test Process: Steps, tips, and more! Dr. Jennifer L. Bowie For Digital Rhetoric.
Advertisements

User Interface Evaluation Formative Evaluation. Summative Evaluation Evaluation of the user interface after it has been developed. Typically performed.
Data gathering. Overview Four key issues of data gathering Data recording Interviews Questionnaires Observation Choosing and combining techniques.
Surveys and Questionnaires. How Many People Should I Ask? Ask a lot of people many short questions: Yes/No Likert Scale Ask a smaller number.
6.811 / PPAT: Principles and Practice of Assistive Technology Wednesday, 16 October 2013 Prof. Rob Miller Today: User Testing.
CS305: HCI in SW Development Evaluation (Return to…)
Web E’s goal is for you to understand how to create an initial interaction design and how to evaluate that design by studying a sample. Web F’s goal is.
HCI 특론 (2007 Fall) User Testing. 2 Hall of Fame or Hall of Shame? frys.com.
SEVEN FREQUENTLY ASKED QUESTIONS ABOUT USABILITY TESTING Usability Testing 101.
Chapter 3 Producing Data 1. During most of this semester we go about statistics as if we already have data to work with. This is okay, but a little misleading.
©2010 John Wiley and Sons Chapter 14 Research Methods in Human-Computer Interaction Chapter 14- Working with Human Subjects.
CyLab Usable Privacy and Security Laboratory 1 C yLab U sable P rivacy and S ecurity Laboratory Laboratory.
Data gathering.
CS160 Discussion Section Matthew Kam Apr 14, 2003.
Group Project. Don’t make me think Steve Krug (2006)
Chapter 14: Usability testing and field studies. Usability Testing Emphasizes the property of being usable Key Components –User Pre-Test –User Test –User.
1 User Testing. 2 Hall of Fame or Hall of Shame? frys.com.
Think-aloud usability experiments or concurrent verbal accounts Judy Kay CHAI: Computer human adapted interaction research group School of Information.
User Interface Testing. Hall of Fame or Hall of Shame?  java.sun.com.
Usable Privacy and Security Carnegie Mellon University Spring 2008 Lorrie Cranor 1 Designing user studies February.
Intro to Evaluation See how (un)usable your software really is…
ICS 463, Intro to Human Computer Interaction Design: 8. Evaluation and Data Dan Suthers.
Project Design and Data Collection Methods: A quick and dirty introduction to classroom research Margaret Waterman September 21, 2005 SoTL Fellows
CyLab Usable Privacy and Security Laboratory 1 C yLab U sable P rivacy and S ecurity Laboratory Surveys,
Electronic EDI e-EDI. The EDI has been in use since 1999 using a paper-based system and computerized spreadsheets to collect and manage EDI data. Over.
Spring break survey how much will your plans suck? how long are your plans? how many people are involved? how much did you overpay? what’s your name? how.
CyLab Usable Privacy and Security Laboratory 1 C yLab U sable P rivacy and S ecurity Laboratory Surveys,
Chapter 14: Usability testing and field studies
236: II'nMI Usability Testing. What is Usability Testing? Usability testing: What is it? A way to assess the usability of a design with real-world users,
Data gathering. Overview Four key issues of data gathering Data recording Interviews Questionnaires Observation Choosing and combining techniques.
Part 1-Intro; Part 2- Req; Part 3- Design  Chapter 20 Why evaluate the usability of user interface designs?  Chapter 21 Deciding on what you need to.
Involving Users in Interface Evaluation Marti Hearst (UCB SIMS) SIMS 213, UI Design & Development April 8, 1999.
Evaluating a Research Report
Human Computer Interaction
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley The Resonant Interface HCI Foundations for Interaction Design First Edition.
©2010 John Wiley and Sons Chapter 6 Research Methods in Human-Computer Interaction Chapter 6- Diaries.
Assumes that events are governed by some lawful order
Usability Testing Chapter 6. Reliability Can you repeat the test?
COMP5047 Pervasive Computing: 2012 Think-aloud usability experiments or concurrent verbal accounts Judy Kay CHAI: Computer human adapted interaction research.
September 2007 Survey Development Rita O'Sullivan Evaluation, Assessment, & Policy Connections (EvAP) School of Education, University of North Carolina-Chapel.
The product of evaluation is knowledge. This could be knowledge about a design, knowledge about the user or knowledge about the task.
Usability Evaluation, part 2. REVIEW: A Test Plan Checklist, 1 Goal of the test? Specific questions you want to answer? Who will be the experimenter?
Dobrin / Weisser / Keller: Technical Communication in the Twenty-First Century. © 2010 Pearson Education. Upper Saddle River, NJ, All Rights Reserved.
Prof. James A. Landay University of Washington Autumn 2004 User Testing December 2, 2004.
AVI/Psych 358/IE 340: Human Factors Data Gathering October 3, 2008.
4. Marketing research After carefully studying this chapter, you should be able to: Define marketing research; Identify and explain the major forms of.
Usability Evaluation. Objectives for today Go over upcoming deliverables Learn about usability testing (testing with users) BTW, we haven’t had a quiz.
The effects of Peer Pressure, Living Standards and Gender on Underage Drinking Psychologist- Kanari zukoshi.
Ethics in Evaluation Why ethics? What you have to do Slide deck by Saul Greenberg. Permission is granted to use this for non-commercial purposes as long.
Prof. James A. Landay University of Washington Autumn 2006 User Testing November 30, 2006.
User Testing. CSE490f - Autumn 2006User Interface Design, Prototyping, & Evaluation2 Hall of Fame or Hall of Shame? frys.com.
Usability Evaluation or, “I can’t figure this out...do I still get the donuts?”
PREPARATION OF QUESTIONNAIRES PREPARATION OF QUESTIONNAIRES Chapter - 4 Dr. BALAMURUGAN MUTHURAMAN
11/10/981 User Testing CS 160, Fall ‘98 Professor James Landay November 10, 1998.
Evaluation / Usability. ImplementDesignAnalysisEvaluateDevelop ADDIE.
School of Engineering and Information and Communication Technology KIT305/607 Mobile Application Development Week 7: Usability (think-alouds) Dr. Rainer.
The Information School of the University of Washington Information System Design Info-440 Autumn 2002 Session #20.
Day 8 Usability testing.
User Interface Evaluation
Lecture3 Data Gathering 1.
Usability Testing 3 CPSC 481: HCI I Fall 2014 Anthony Tang.
Usability Evaluation.
Professor John Canny Spring 2003
based on notes by James Landay
User Testing November 27, 2007.
Professor John Canny Spring 2004
Professor John Canny Fall 2004
Empirical Evaluation Data Collection: Techniques, methods, tricks Objective data IRB Clarification All research done outside the class (i.e., with non-class.
Human-Computer Interaction: Overview of User Studies
Presentation transcript:

CyLab Usable Privacy and Security Laboratory 1 C yLab U sable P rivacy and S ecurity Laboratory Designing Experiments Lorrie Faith Cranor September 2009

CyLab Usable Privacy and Security Laboratory How is HCISEC different from HCI?  Is it different? If so, how?  Are different user study methods needed?

CyLab Usable Privacy and Security Laboratory 3 Designing and conducting a user study  Identify purpose and metrics  Decide what type of data to collect  Design tasks  Develop experimental design  Develop detailed plan, artifacts, protocol and scripts  IRB approval  Pilot test and revise  Recruit participants and run study  Analyze data  Report results

CyLab Usable Privacy and Security Laboratory 4 Purpose & metrics  Identify purpose of study – What are you trying to learn? Human-in-the-loop questions may be relevant – What are your hypotheses?  Identify metrics – How will you know if it is better, faster, more useful, more usable, etc. ? – What will you measure? – What will you compare it to? – What is your target improvement, time, score, etc. ?  What qualitative data are you looking for?

CyLab Usable Privacy and Security Laboratory 5 What data to collect Process data  Observations of what users are doing or thinking – Think aloud – Audio recording, video recording, screen capture  Formative evaluation – Collect this data first to help identify problems – Useful results with few users – May be hard generalize from results – Complements bottom-line data to help explain results But think-aloud can impact speed and accuracy Bottom-line data  Summary of what happened – Time – Accuracy, errors  Summative evaluation – Validate an approach, compare multiple approaches – But doesn’t usually tell you what went wrong or how to fix (“too slow” “too many errors”) – May require many users for statistical significance

CyLab Usable Privacy and Security Laboratory 6 What data to collect Subjective user preference  Rate system on a Likert scale – Example: The user interface was easy to use 1 - Strongly disagree 2 - Disagree 3 - Neither agree nor disagree 4 - Agree 5 - Strongly agree – Can be hard to be sure what data means novelty, not realistic setting …  Open ended questions – what they liked, disliked, where they had trouble, best part, worst part, etc. Demographics of your participants  Important to know something about your participants  Basic demographics – Age, gender, income level, education level, occupation/major, location, disabilities  Other factors relevant to your study – Technical experience, native language, willingness to take risks, attitudes about privacy, etc.

CyLab Usable Privacy and Security Laboratory 7 Tasks  What tasks will you ask users to perform to allow you to take needed measurements? What questions will you ask them?  What degree of user interface fidelity do you need to allow you to take needed measurements? – Is a paper prototype or low fidelity prototype be preferable, or is a high-fidelity prototype needed?  Where should the study be done? – Lab study, online study, or field study?

CyLab Usable Privacy and Security Laboratory 8 Experimental design  What kind of experimental design should you use? Within subjects, between subjects, hybrid  How many participants should you have? – What will you need for statistical significance? – What are your constraints in terms of time, budget, etc?  What kind of subjects do you need and how will you recruit them? – Special characteristics, knowledge or skills? Sometimes we recruit a particular type of subject because it is more convenient, even if it doesn’t produce as generalizable results – What incentives will they have to participate?

CyLab Usable Privacy and Security Laboratory 9 Comparing alternatives  Between groups experiment (AB experiment) – Two (or more) groups of test users – Each group uses only 1 of the systems – Requires more participants Make sure each group is similar  Within groups experiment – One group of test users – Each person uses both systems Need to control for learning effect

CyLab Usable Privacy and Security Laboratory 10 Controlling for task order and learning effects  Present tasks in random order  Use Latin squares design – n x n table filled with n different symbols such that each appears exactly once in each row and each column

CyLab Usable Privacy and Security Laboratory 11 Detailed plan  Develop artifacts – Prototypes, questionnaires, screening tools, measuring tools, etc.  Protocol and scripts – Exactly what will participants do? – Will you ask participants to think aloud? – What will experimenter(s) do and say? – Do you need to train participants? – Are warm-up or distracter tasks needed? – Will you make audio or video recordings or do screen captures? – Will the experimenter record specific information? Is there a form or template to facilitate this?  Figure out how you will analyze your data

CyLab Usable Privacy and Security Laboratory 12 IRB approval  All published research studies involving human subjects must have CMU IRB approval  Surveys are exempt, but you must still fill out form and ask IRB to give you exemption  Exempt and low-risk IRB approval usually happens within 2 weeks  High risk usually takes about a month, but may be longer if you have to iterate with IRB  Whenever possible, design study so participants sign informed consent form up front – You will have to convince IRB that there is a good reason not to  Submit your IRB form as early as possible, even if not all your study details are worked out – You can submit an amendment later – Label all recruitment forms and questionnaires as “Example” for more flexibility 

CyLab Usable Privacy and Security Laboratory 13 Ethical considerations  Sometimes tests can be distressing – users have left in tears  You have a responsibility to alleviate – make voluntary with informed consent – avoid pressure to participate – let them know they can stop at any time – stress that you are testing the system, not them – make collected data as anonymous as possible

CyLab Usable Privacy and Security Laboratory 14 Pilot test and revise  Run through the whole protocol with members of your team to work out all the details  Run through it with your friends or people you recruit to debug the protocol and find out how long it will take  Do some preliminary data analysis  Revise – Make sure tasks and questions aren’t confusing – Make sure the study can be done in a reasonable amount of time – Make sure the study measures what you are trying to measure  Repeat

CyLab Usable Privacy and Security Laboratory 15 Recruit participants and run study  Posters, , ads, etc. to recruit study participants  Screen participants, sign them up – Make sure they know where to meet you and how to contact you – Send them a reminder the day of the study  Run the study – Make sure you have reserved lab or appropriate space, if needed – Post signs to help participants find you – Make sure you have enough people there to run the study – Make sure your computers are setup, you have your recording devices, human subject payments, and anything else you need

CyLab Usable Privacy and Security Laboratory 16 Analyze data  Sort & prioritize observations – what was important? – lots of problems in the same area?  Summarize the data – make a list of all critical incidents, positive & negative – try to judge why each difficulty occurred  Determine scores, times, etc.  Code audio or text for quantitative analysis  Run appropriate statistical tests to test your hypotheses  What does data tell you?  Iterate – Should you change your interface and test again? – Should you change your experimental protocol to fix problems, test other data?

CyLab Usable Privacy and Security Laboratory 17 Report results  Detail your methodology – Document your assumptions and threat model – How were participants recruited? – What incentive was provided to participate? – Where did the participants go to participate? – What were participants asked to do before, as part of, and following the experiment? – What information did participants learn along the way and how might this have influenced behaviors later on in the experiment? – If the study was a between-subjects study, how did the experience (treatment) vary between the groups? – Did the order of any tasks change for different participants?  Use proper statistical analysis – If your distribution is not normal, don’t use statistical tests that assume a normal distribution – Account for multiple comparisons – If you can’t demonstrate a statistically significant difference it does not necessarily it mean it doesn’t exist – Statistical correlation does not prove causation  Report the limitations of your study  Cite related work

CyLab Usable Privacy and Security Laboratory 18 Group exercise: study design  The AT&T web mail client identifies suspected phishing s and warns a user if they try to open them. If a user opens the messages anyway, they will see the warning symbol next to all suspicious links. If they click on the link, they will go to a page warning them that the link is suspicious, and asking them if they are sure they want to proceed. Design a user study that will allow you to evaluate the effectiveness of this approach to protecting users from phishing and to come up with recommendations for improving the warning interface. Optionally, you can come up with some design improvements and test them in your user study as well. (Groups will be assigned to design a lab study or a field study.)

CyLab Usable Privacy and Security Laboratory 22 Cylab Usable Privacy and Security Laboratory