The evolution of evaluation Joseph Jofish Kaye Microsoft Research, Cambridge Cornell University, Ithaca, NY cornell.edu.

Slides:



Advertisements
Similar presentations
Evaluation in Experience-Focused HCI: Exploring the Virtual Intimate Object Future Applications Lab Viktoria Institute Gothenburg, Sweden 13 January 2006.
Advertisements

Chapter 15: Analytical evaluation
Evolution of Evaluation in HCI Joseph Jofish Kaye Microsoft Research, Cambridge Cornell University, Ithaca, NY cornell.edu HCI Seminar Series.
The evolution of evaluation Joseph Jofish Kaye Microsoft Research, Cambridge Cornell University, Ithaca, NY cornell.edu.
Critical Reading Strategies: Overview of Research Process
Map of Human Computer Interaction
Action Research Not traditional educational research often research tests theory not practical Teacher research in classrooms and/or schools/districts.
CS305: HCI in SW Development Evaluation (Return to…)
WHAT IS INTERACTION DESIGN?
©N. Hari Narayanan Computer Science & Software Engineering Auburn University 1 COMP 7620 Evaluation Chapter 9.
Methodology Overview Dr. Saul Greenberg John Kelleher.
Chapter 15: Analytical evaluation. 2 FJK User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept. Cal Poly San.
1 User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept. Cal Poly San Luis Obispo FJK 2009.
Writing a Science or Engineering Paper: It is just a story Frank Shipman Department of Computer Science Texas A&M University.
Empirical Methods in Human- Computer Interaction.
Psychological Aspects Presented by Hanish Patel. Overview  HCI (Human Computer Interaction)  Overview of HCI  Human Use of Computer Systems  Science.
Evaluation Methodologies
Design and Evaluation of Iterative Systems n For most interactive systems, the ‘design it right first’ approach is not useful. n The 3 basic steps in the.
Heuristic Evaluation Evaluating with experts. Discount Evaluation Techniques  Basis: Observing users can be time- consuming and expensive Try to predict.
Evaluating with experts
Writing Good Software Engineering Research Papers A Paper by Mary Shaw In Proceedings of the 25th International Conference on Software Engineering (ICSE),
An evaluation framework
Usability 2004 J T Burns1 Usability & Usability Engineering.
4. Interaction Design Overview 4.1. Ergonomics 4.2. Designing complex interactive systems Situated design Collaborative design: a multidisciplinary.
PHCL 328: Introduction to Drug and Poison Information
Some questions of hypermedia and CHI Josep Blat Universitat Pompeu Fabra.
4. Interaction Design Overview 4.1. Ergonomics 4.2. Designing complex interactive systems Situated design Collaborative design: a multidisciplinary.
INTRODUCTION. Concepts HCI, CHI Usability User-centered Design (UCD) An approach to design (software, Web, other) that involves the user Interaction Design.
Evolving Evaluation: from Engineers to Experience Stanford University Human-Computer Interaction Seminar 27 April 2007 Joseph ‘Jofish’ Kaye Cornell University,
©2011 1www.id-book.com Analytical evaluation Chapter 15.
On Education Gerrit C. van der Veer most work done by Anne Bowser Elizabeth Churchill Jennifer Preece.
1. Human – the end-user of a program – the others in the organization Computer – the machine the program runs on – often split between clients & servers.
ICEE 2005GLIWICE, POLAND JULY 2005 FEDERAL CENTER OF TECHNOLOGICAL EDUCATION – CEFET-RJ – BRAZIL PRODUCTION ENGINEERING DEPARTMENT CSCW: A FORMATION.
Predictive Evaluation
Towards an activity-oriented and context-aware collaborative working environments Presented by: Ince T Wangsa Supervised by:
User-Centered Development Methodology A user interface comprises “ those aspects of the system that the user comes in contact with.” ● Moran [1981]
BUILDING BLOCKS OF PSYCHOLOGY MS. CARMELITANO. What is this?
Human Computer Interaction
HCI in Software Process Material from Authors of Human Computer Interaction Alan Dix, et al.
Methodology and Explanation XX50125 Lecture 3: Usability testing Dr. Danaë Stanton Fraser.
Assessing the Frequency of Empirical Evaluation in Software Modeling Research Workshop on Experiences and Empirical Studies in Software Modelling (EESSMod)
Testing & modeling users. The aims Describe how to do user testing. Discuss the differences between user testing, usability testing and research experiments.
Evaluation of User Interface Design 4. Predictive Evaluation continued Different kinds of predictive evaluation: 1.Inspection methods 2.Usage simulations.
Introduction to Science Informatics Lecture 1. What Is Science? a dependence on external verification; an expectation of reproducible results; a focus.
Chapter 12: Introducing Evaluation. The aims To illustrate how observation, interviews and questionnaires that you encountered in Chapters 7 and 8 are.
Chapter 12: Introducing Evaluation. The aims To illustrate how observation, interviews and questionnaires that you encountered in Chapters 7 and 8 are.
Design Process … and some design inspiration. Course ReCap To make you notice interfaces, good and bad – You’ll never look at doors the same way again.
EVALUATION PROfessional network of Master’s degrees in Informatics as a Second Competence – PROMIS ( TEMPUS FR-TEMPUS-JPCR)
Cognitive Walkthrough More evaluating with experts.
Chapter 15: Analytical evaluation. Aims: Describe inspection methods. Show how heuristic evaluation can be adapted to evaluate different products. Explain.
Writing a Science or Engineering Paper: It is just a story Frank Shipman Department of Computer Science Texas A&M University.
Problem Solving. Definition Basic intellectual process that has been refined and systemized for the various challenges people face.
Introduction to Evaluation “Informal” approaches.
Fall 2002CS/PSY Predictive Evaluation (Evaluation Without Users) Gathering data about usability of a design by a specified group of users for a particular.
HCI Meeting 1 Thursday, August 26. Class Activities [1] Student questionnaire Answer the following questions: 1.When and where was the computer mouse.
User Interface Evaluation Introduction Lecture #15.
A. Strategies The general approach taken into an enquiry.
PHILOSOPHY AS A SECOND ORDER DISCIPLINE
PHILOSOPHY AS A SECOND ORDER DISCIPLINE
Human Computer Interaction (HCI)
SIE 515 Design Evaluation Lecture 7.
Methodology Overview 2 basics in user studies Lecture /slide deck produced by Saul Greenberg, University of Calgary, Canada Notice: some material in this.
Architecture Components
Introducing Evaluation
WHAT IS INTERACTION DESIGN?
Seminal Ideas in Human-Computer Interaction
Fundamentals of Human Computer Interaction (HCI)
User Interface Design and Evaluation
Evaluation.
Presentation transcript:

The evolution of evaluation Joseph Jofish Kaye Microsoft Research, Cambridge Cornell University, Ithaca, NY cornell.edu

What is evaluation? Something you do at the end of a project to show it works… … so you can publish it. A tradition in a field A way of defining a field A process that changes over time A reason papers get rejected

HCI Evaluation: Validity Methods for establishing validity vary depending on the nature of the contribution. They may involve empirical work in the laboratory or the field, the description of rationales for design decisions and approaches, applications of analytical techniques, or proof of concept system implementations CHI 2007 Website

So… How did we get to where we are today? Why did we end up with the system(s) we use today? How can our current approaches to evaluation deal with novel concepts of HCI, such as experience-focused (rather than task focused) HCI?

Experience focused HCI (a question to think about during this talk) What does it mean when this is your evaluation method?

A Brief History and plan for the talk 1.Evaluation by Engineers 2.Evaluation by Computer Scientists 3.Evaluation by Experimental Psychologists & Cognitive Scientists 4.Evaluation by HCI Professionals 5.Evaluation in CSCW 6.Evaluation for Experience

A Brief History and plan for the talk 1.Evaluation by Engineers 2.Evaluation by Computer Scientists 3.Evaluation by Experimental Psychologists & Cognitive Scientists a.Case study: Evaluation of Text Editors 4.Evaluation by HCI Professionals a)Case Study: The Damaged Merchandise Debate 5.Evaluation in CSCW 6.Evaluation for Experience

3 Questions to ask about an era Who are the users? Who are the evaluators? What are the limiting factors?

Evaluation by Engineers Users are engineers & mathematicians Evaluators are engineers The limiting factor is reliability

Evaluation by Computer Scientists Users are programmers Evaluators are programmers The speed of the machine is the limiting factor

Evaluation by Experimental Psychologists & Cognitive Scientists Users are users: the computer is a tool, not an end result Evaluators are cognitive scientists and experimental psychologists: theyre used to measuring things through experiment The limiting factor is what the human can do

Perceptual issues such as print legibility and motor issues arose in designing displays, keyboards and other input devices… [new interface developments] created opportunities for cognitive psychologists to contribute in such areas as motor learning, concept formation, semantic memory and action. In a sense, this marks the emergence of the distinct discipline of human- computer interaction. (Grudin 2006) Evaluation by Experimental Psychologists & Cognitive Scientists

Case Study: Text Editors Roberts & Moran, 1982, Their methodology for evaluating text editors had three criteria: objectivity thoroughness ease-of-use

Case Study: Text Editors objectivity implies that the methodology not be biased in favor of any particular editors conceptual structure thoroughness implies that multiple aspects of editor use be considered ease-of-use (of the method, not the editor itself) the methodology should be usable by editor designers, managers of word processing centers, or other nonpsychologists who need this kind of evaluative information but who have limited time and equipment resources

Case Study: Text Editors objectivity implies that the methodology not be biased in favor of any particular editors conceptual structure thoroughness implies that multiple aspects of editor use be considered. ease-of-use (of the method (not the editor itself), the methodology should be usable by editor designers, managers of word processing centers, or other nonpsychologists who need this kind of evaluative information but who have limited time and equipment resources.

Case Study: Text Editors Text editors are the white rats of HCI Thomas Green, 1984, in Grudin, 1990.

Case Study: Text Editors Text editors are the white rats of HCI Thomas Green, 1984, in Grudin, …which tells us more about HCI than it does about text editors.

Evaluation by HCI Professionals Usability professionals They believe in expertise (e.g. Neilsen 1984) Theyve made a decision decision to decide to focus on better results, regardless of whether they were experimentally provable or not.

Case Study: The Damaged Merchandise Debate

Damaged Merchandise Setup Early eighties: usability evaluation methods (UEMs) - heuristics (Neilsen) - cognitive walkthrough - GOMS - …

Damaged Merchandise Comparison Studies Jefferies, Miller, Wharton and Uyeda (1991) Karat, Campbell and Fiegel (1992) Neilsen (1992) Desuirve, Kondziela, and Atwood (1992) Neilsen and Phillips (1993)

Damaged Merchandise Panel Wayne D. Gray, Panel at CHI95 Discount or Disservice? Discount Usability Analysis at a Bargain Price or Simply Damaged Merchandise

Damaged Merchandise Paper Wayne D. Gray & Marilyn Salzman Special issue of HCI: Experimental Comparisons of Usability Evaluation Methods

Damaged Merchandise Response Commentary on Damaged Merchandise Karat: experiment in context Jefferies & Miller: real-world Lund & McClelland: practical John: case studies Monk: broad questions Oviatt: field-wide science MacKay: triangulate Newman: simulation & modelling

Damaged Merchandise Whats going on? Gray & Salzman, p19 There is a tradition in the human factors literature of providing advice to practitioners on issues related to, but not investigated in, an experiment. This tradition includes the clear and explicit separation of experiment- based claims from experience-based advice. Our complaint is not against experimenters who attempt to offer good advice… the advice may be understood as research findings rather than the researchers opinion.

Damaged Merchandise Whats going on? Gray & Salzman, p19 There is a tradition in the human factors literature of providing advice to practitioners on issues related to, but not investigated in, an experiment. This tradition includes the clear and explicit separation of experiment- based claims from experience-based advice. Our complaint is not against experimenters who attempt to offer good advice… the advice may be understood as research findings rather than the researchers opinion.

Damaged Merchandise Clash of Paradigms Experimental Psychologists & Cognitive Scientists (who believe in experimentation) vs. HCI Professionals (who believe in experience and expertise, even if unprovable) (and who were trying to present their work in the terms of the dominant paradigm of the field.)

Evaluation in CSCW A story Im not telling CSCW vs. HCI Not just groups, but philosophy (ideology!) Member-created, dynamic, not cognitive, modelable Follows failure of workplace studies to characterize IE Plans and Situated Actions vs. The Psychology of Human- Computer Interaction

Evaluation of Experience Focused HCI A possibly emerging sub-field: –Gaver et. al. –Isbister et. al. –Höök et. al. –Sengers et. al. –Etc. How to evaluate?

Epistemology How does a field know what it knows? How does a field know that it knows it? Science: experiment… But literature? Anthropology? Sociology? Therapy? Art? Theatre? Design?

Epistemology Formally: The aim of this work is to recognize the ways in which multiple epistemologies, not just the experimental paradigm of science, can and do inform the hybrid discipline of human-computer interaction.

Shouts To My Homies Maria Håkansson Lars Erik Holmquist Alex Taylor & MS Research Phoebe Sengers & CEmCom Cornell S&TS Department Many discussions over the last year… and this one to come.