IS 4800 Empirical Research Methods for Information Science Class Notes Feb. 29, 2012 Instructor: Prof. Carole Hafner, 446 WVH Tel: 617-373-5116.

Slides:



Advertisements
Similar presentations
Chapter 2 The Process of Experimentation
Advertisements

Chapter 22 Evaluating a Research Report Gay, Mills, and Airasian
Critical Reading Strategies: Overview of Research Process
Experimental design Bubbles!.
Human Computer Interaction
Human Computer Interaction
evaluation techniques
Evaluating UI Designs assess effect of interface on user performance and satisfaction identify specific usability problems evaluate users’ access to functionality.
Writing for Publication
Evaluation techniques Part 2
The art and science of measuring people l Reliability l Validity l Operationalizing.
Chapter 14: Usability testing and field studies. Usability Testing Emphasizes the property of being usable Key Components –User Pre-Test –User Test –User.
©N. Hari Narayanan Computer Science & Software Engineering Auburn University 1 COMP 7620 Evaluation Chapter 9.
Useability.
Evaluation Methodologies
From Controlled to Natural Settings
Guidelines to Publishing in IO Journals: A US perspective Lois Tetrick, Editor Journal of Occupational Health Psychology.
ICS 463, Intro to Human Computer Interaction Design: 8. Evaluation and Data Dan Suthers.
Knowledge is Power Marketing Information System (MIS) determines what information managers need and then gathers, sorts, analyzes, stores, and distributes.
Research problem, Purpose, question
RESEARCH DESIGN.
Descriptive and Causal Research Designs
Chapter 14: Usability testing and field studies
 For the IB Diploma Programme psychology course, the experimental method is defined as requiring: 1. The manipulation of one independent variable while.
CriteriaExemplary (4 - 5) Good (2 – 3) Needs Improvement (0 – 1) Identifying Problem and Main Objective Initial QuestionsQuestions are probing and help.
Ch 14. Testing & modeling users
Evaluation Techniques Material from Authors of Human Computer Interaction Alan Dix, et al.
Human Computer Interaction Chapter 3 Evaluation Techniques.
Evaluating a Research Report
Exploratory Research Design Week 02
Human Computer Interaction
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Lecturer – Prof Jim Warren Lecture 4 - Usability Testing Based on Heim, Chapter.
Report Format and Scientific Writing. What is Scientific Writing? Clear, simple, well ordered No embellishments, not an English paper Written for appropriate.
Evaluation Techniques zEvaluation ytests usability and functionality of system yoccurs in laboratory, field and/or in collaboration with users yevaluates.
IS 4800 Empirical Research Methods for Information Science Class Notes Feb. 22, 2012 Instructor: Prof. Carole Hafner, 446 WVH Tel:
Evaluating Research Articles Approach With Skepticism Rebecca L. Fiedler January 16, 2002.
How to read a scientific paper
Usability Testing Chapter 6. Reliability Can you repeat the test?
COMP5047 Pervasive Computing: 2012 Think-aloud usability experiments or concurrent verbal accounts Judy Kay CHAI: Computer human adapted interaction research.
Chapter 15 Qualitative Data Collection Gay, Mills, and Airasian
Chapter 9 evaluation techniques. Evaluation Techniques Evaluation –tests usability and functionality of system –occurs in laboratory, field and/or in.
Evaluation Techniques Evaluation –tests usability and functionality of system –occurs in laboratory, field and/or in collaboration with users –evaluates.
CS 580 chapter 9 evaluation techniques. Evaluation Tests usability and functionality of system Occurs in laboratory, field and/or in collaboration with.
CENG 394 Introduction to Human-Computer Interaction
AMSc Research Methods Research approach IV: Experimental [1] Jane Reid
IS 4800 Empirical Research Methods for Information Science Class Notes Feb. 24, 2012 Instructor: Prof. Carole Hafner, 446 WVH Tel:
Marketing Research Approaches. Research Approaches Observational Research Ethnographic Research Survey Research Experimental Research.
Copyright © Allyn & Bacon 2008 Intelligent Consumer Chapter 14 This multimedia product and its contents are protected under copyright law. The following.
Business Project Nicos Rodosthenous PhD 08/10/2013 1
Experimental Psychology PSY 433 Chapter 5 Research Reports.
Observational Methods Think Aloud Cooperative evaluation Protocol analysis Automated analysis Post-task walkthroughs.
Chapter 9 evaluation techniques. Evaluation Techniques Evaluation –tests usability and functionality of system –occurs in laboratory, field and/or in.
Chapter 9 evaluation techniques. Evaluation Techniques Evaluation – tests usability and functionality of system – occurs in laboratory, field and/or in.
School of Engineering and Information and Communication Technology KIT305/607 Mobile Application Development Week 7: Usability (think-alouds) Dr. Rainer.
Overview Introduction to marketing research Research design Data collection Data analysis Reporting results.
Day 8 Usability testing.
Introduction to Marketing Research
Human Computer Interaction
Evaluation through user participation
Human Computer Interaction Lecture 15 Usability Evaluation
Experimental Psychology
Evaluation techniques
From Controlled to Natural Settings
Explaining the Methodology : steps to take and content to include
Evaluation.
HCI Evaluation Techniques
Evaluation Techniques
Experimental Evaluation
evaluation techniques
Presentation transcript:

IS 4800 Empirical Research Methods for Information Science Class Notes Feb. 29, 2012 Instructor: Prof. Carole Hafner, 446 WVH Tel: Course Web site:

2 Oral Presentation of Study Results

Presenting your research ■A research project should “tell a story” ■A brief presentation (20 min or less) is more difficult than a longer one – selectivity is critical ■The T-shaped talk and the U-shaped talk ■Rehearse for timing

4 Oral Presentation ■Main concepts and ideas ■Do not go into great detail on experimental methods – BUT enough so people understand what you did ■Focus on motivation, results, implications ■If listener wants details they can read the paper or ask questions

5 Oral Presentation Don’t do this…

6 Oral Presentation Do use figures a lot WEEK 1WEEK 4 COMPOSITE BOND TASK GOAL COMPOSITE BOND TASK GOAL

7 Oral Presentation Guide for Visuals ■Visuals should be exhibits that you talk about ■Do not put lots of text on charts ■Do not read your charts for your presentation ■Use interactivity, video, images to keep your audience awake

8 Common Questions ■How did you evaluate that? ■How did you measure that? ■How did you control for extraneous variable X? ■Why didn’t you use statistic Y? ■Isn’t that a biased sample? ■What was your control group? ■How did you do study procedure Z?

9 Tips ■Describe your sample ■Minimal demographics – number of subjects, broken down by gender ■Better: age, occupation, major, year ■Minimize text on your charts ■If you use a novel measure (e.g., new survey) you must give details on the measure ■Examples of most important actual questions asked ■Any reliability/validity/psychometrics done ■If you do interviews, include actual quotes ■Build from data to conclusions ■Practice your timing/delivery with your project team

10 Written Study Reports ■Objectives (also critiques) ■Describe what your study is about – should also tell a story, and maybe a more complex one ■Motivate your study ■Assure reader you have conducted a sound study Research Methods – often presented in small font ■Present results in an objective manner ■Discuss implications ■Discuss future work ■Enable replication

11 Typical Academic IS/CS/HCI Paper Structure Astract Introduction Motivation Related work Hypotheses Method Results Discussion Limitations Implications Future work References

12 Typical Design/Development Study ■Abstract (?) ■Executive Summary and Recommendations ■Introduction ■Motivation ■Related work ■System design ■Evaluation ■Hypotheses ■Method ■Results ■Discussion – summary, limitations ■Conclusion ■Implications ■Future work ■References

13 The Abstract ■Concise summary (one paragraph!) ■Abstract for an empirical study should include ■Information on the problem under study ■The nature of the subject sample ■A description of methods, equipment, and procedures ■A statement of the results ■A statement of the findings or conclusions drawn ■Often the last thing you write

14 The Introduction ■Part of paper giving justification for study ■Usually has the following information ■Introduction to the topic under study ■Brief review of research and theory related to the topic ■A statement of the problem to be addressed ■A statement of the purpose of the research ■A brief description of the research strategy ■A description of predictions and hypotheses ■CS/IS papers often put Related Work as a separate section after Introduction ■For each, describe how your work is different

15 Organization of the Introduction: General to Specific

16 The Method Section ■Includes information on exactly how a study was carried out ■Subsections ■Participants or subjects Describe in detail the participant or subject sample Human participants go in a Participants subsection, and animal subjects in a Subjects subsection ■Apparatus or materials Describe in detail any equipment or materials used Equipment is usually described in an Apparatus subsection and written materials in a Materials subsection

17 ■Procedure ■Describe Exactly how the study was carried out The conditions to which subjects were exposed or under which observed The behaviors measured and how they were scored When and where observations were made Debriefing procedures ■Enough detail should be included in all sections so that the study could be replicated The Method Section

18 The Results Section ■Objective, dry, boring – just the facts ■All relevant analyses are reported in the results section ■Do not present raw data ■Data should be reported in summary form ■Descriptive statistics ■Inferential statistics ■Results of descriptive and inferential statistics must be presented in narrative format ■Describe the source of any unconventional statistical tests

19 Commonly Used Statistical Citations Statistical TestFormat Analysis of varianceF (1,85) = 5.96, p <.01 Chi-squareχ 2 (3) = 11.34, p <.01 t testt (56) = 4.78, p <.01

20 Abbreviations for Statistical Notation AbbreviationMeaning dfDegrees of freedom FF ratio MArithmetic average (mean) NNumber of subjects in entire sample pp value SDStandard deviation tt statistic zResults from a z test or z score μPopulation mean (mu)  Population stddev

21 The Discussion Section ■This is where you can take some liberties with describing what the results mean ■Results are interpreted, conclusions drawn, and findings are related to previous research ■Section begins with a brief restatement of hypotheses ■Next, indicate if hypotheses were confirmed ■The rest of the section is dedicated to integrating findings with previous research ■It is fine to speculate, but speculations should not stray far from the data

22 Organization of Discussion: Specific to General

23 Example

24

25

26

27

28

29

30

31

32

33

34

35

36 Citations ■Liberally cite previous & related work. ■If you copy passages you must cite and, depending on length, format to indicate it is copied. ■Suggest using EndNote, BibTex or similar.

37 Ethical Issues ■Report all of your findings (not just the ones you like) ■Adhere to your original plan ■Report any deviations and why ■Power analysis, statistics, measures ■Do not drop subjects or data points without rigorous justification ■If your hypothesis test was not significant you cannot say anything about difference in means (example). ■If you did not do an experiment, attempting to control for extraneous variables, you cannot mention or imply causality.

Introduction to Usability Testing I. Summative evaluation: Measure/compare user performance and satisfaction Quantitative measures Statistical methods II. Formative Evaluation: Identify Usability Problems Quantitative and Qualitative measures Ethnographic methods such as interviews, focus groups

Usability Goals (Nielsen) 1. Learnability 2. Efficiency 3. Memorability 4. Error avoidance/recovery 5. User satisfaction Operationalize these goals to evaluate usability

What is a Usability Experiment? Usability testing in a controlled environment There is a test set of users They perform pre-specified tasks Data is collected (quantitative and qualitative) Take mean and/or median value of measured attributes Compare to goal or another system Contrasted with “expert review” and “field study” evaluation methodologies The growth of usability groups and usability laboratories

Subjects representative sufficient sample Variables independent variable (IV) characteristic changed to produce different conditions. e.g. interface style, number of menu items. dependent variable (DV) characteristics measured in the experiment e.g. time taken, number of errors. Experimental factors

Hypothesis prediction of outcome framed in terms of IV and DV null hypothesis: states no difference between conditions aim is to disprove this. Experimental design within groups design each subject performs experiment under each condition. transfer of learning possible less costly and less likely to suffer from user variation. between groups design each subject performs under only one condition no transfer of learning more users required variation can bias results. Experimental factors (cont.)

Summative Analysis What to measure? (and it’s relationship to a usability goal) Total task time User “think time” (dead time??) Time spent not moving toward goal Ratio of successful actions/errors Commands used/not used frequency of user expression of: confusion, frustration, satisfaction frequency of reference to manuals/help system percent of time such reference provided the needed answer

Measuring User Performance Measuring learnability Time to complete a set of tasks Learnability/efficiency trade-off Measuring efficiency Time to complete a set of tasks How to define and locate “experienced” users Measuring memorability The most difficult, since “casual” users are hard to find for experiments Memory quizzes may be misleading

Measuring User Performance (cont.) Measuring user satisfaction Likert scale (agree or disagree) Semantic differential scale Physiological measure of stress Measuring errors Classification of minor v. serious

Reliability and Validity Reliability means repeatability. Statistical significance is a measure of reliability Validity means will the results transfer into a real-life situation. It depends on matching the users, task, environment Reliability - difficult to achieve because of high variability in individual user performance

Formative Evaluation What is a Usability Problem?? Unclear - the planned method for using the system is not readily understood or remembered (info. design level) Error-prone - the design leads users to stray from the correct operation of the system (any design level) Mechanism overhead - the mechanism design creates awkward work flow patterns that slow down or distract users. Environment clash - the design of the system does not fit well with the users’ overall work processes. (any design level) Ex: incomplete transaction cannot be saved

Qualitative methods for collecting usability problems Thinking aloud studies Difficult to conduct Experimenter prompting, non-directive Alternatives: constructive interaction, coaching method, retrospective testing Output: notes on what users did and expressed: goals, confusions or misunderstandings, errors, reactions expressed Questionnaires Should be usability-tested beforehand Focus groups, interviews

user observed performing task user asked to describe what he is doing and why, what he thinks is happening etc. Advantages simplicity - requires little expertise can provide useful insight can show how system is actually use Disadvantages subjective selective act of describing may alter task performance Observational Methods - Think Aloud

variation on think aloud user collaborates in evaluation both user and evaluator can ask each other questions throughout Additional advantages less constrained and easier to use user is encouraged to criticize system clarification possible Observational Methods - Cooperative evaluation

paper and pencil cheap, limited to writing speed audio good for think aloud, diffcult to match with other protocols video accurate and realistic, needs special equipment, obtrusive computer logging automatic and unobtrusive, large amounts of data difficult to analyze user notebooks coarse and subjective, useful insights, good for longitudinal studies Mixed use in practice. Transcription of audio and video difficult and requires skill. Some automatic support tools available Observational Methods - Protocol analysis

analyst questions user on one to one basis usually based on prepared questions informal, subjective and relatively cheap Advantages can be varied to suit context issues can be explored more fully can elicit user views and identify unanticipated problems Disadvantages very subjective time consuming Query Techniques - Interviews

Set of fixed questions given to users Advantages quick and reaches large user group can be analyzed more rigorously Disadvantages less flexible less probing Query Techniques - Questionnaires

Advantages: specialist equipment available uninterrupted environment Disadvantages: lack of context difficult to observe several users cooperating Appropriate if actual system location is dangerous or impractical for to allow controlled manipulation of use. Laboratory studies: Pros and Cons

Steps in a usability experiment 1.The planning phase 1.The execution phase 1.Data collection techniques 1.Data analysis

The planning phase Who, what, where, when and how much? Who are test users, and how will they be recruited? Who are the experimenters? When, where, and how long will the test take? What equipment/software is needed? How much will the experiment cost? Prepare detailed test protocol *What test tasks? (written task sheets) *What user aids? (written manual) *What data collected? (include questionnaire) How will results be analyzed/evaluated? Pilot test protocol with a few users

Detailed Test Protocol What tasks? Criteria for completion? User aids What will users be asked to do (thinking aloud studies)? Interaction with experimenter What data will be collected? All materials to be given to users as part of the test, including detailed description of the tasks.

Execution phase Prepare environment, materials, software Introduction should include: purpose (evaluating software) voluntary and confidential explain all procedures recording question-handling invite questions During experiment give user written task description(s), one at a time only one experimenter should talk De-briefing

Execution phase: ethics of human experimentation applied to usability testing Users feel exposed using unfamiliar tools and making erros Guidelines: Re-assure that individual results not revealed Re-assure that user can stop any time Provide comfortable environment Don’t laugh or refer to users as subjects or guinea pigs Don’t volunteer help, but don’t allow user to struggle too long In de-briefing answer all questions reveal any deception thanks for helping

Execution Phase: Designing Test Tasks Tasks: Are representative Cover most important parts of UI Don’t take too long to complete Goal or result oriented (possibly with scenario) Not frivolous or humorous (unless part of product goal) First task should build confidence Last task should create a sense of accomplishment

Data collection - usability labs and equipment Pad and paper the only absolutely necessary data collection tool! Observation areas (for other experimenters, developers, customer reps, etc.) - should be shown to users Videotape (may be overrated) - users must sign a release Video display capture Portable usability labs Usability kiosks

Before you start to do any statistics: look at data save original data Choice of statistical technique depends on type of data information required Type of data discrete - finite number of values continuous - any value Analysis of data

Testing usability in the field 1. Direct observation in actual use discover new uses take notes, don’t help, chat later 2. Logging actual useobjective, not intrusive great for identifying errors which features are/are not usedprivacy concerns

Testing Usability in the Field (cont.) 3. Questionnaires and interviews with real users ask users to recall critical incidents questionnaires must be short and easy to return 4. Focus groups 6-9 users skilled moderator with pre-planned script computer conferencing?? 5 On-line direct feedback mechanisms initiated by users may signal change in user needs trust but verify 6. Bulletin boards and user groups

Advantages: natural environment context retained (though observation may alter it) longitudinal studies possible Disadvantages: distractions noise Appropriate for “beta testing” where context is crucial for longitudinal studies Field Studies: Pros and Cons