Presentation is loading. Please wait.

Presentation is loading. Please wait.

Evaluation techniques

Similar presentations


Presentation on theme: "Evaluation techniques"— Presentation transcript:

1 Evaluation techniques

2 Overview Evaluation tests the usability, functionality and acceptability of an interactive system. Evaluation may take place: In the laboratory In the field

3 Overview: Laboratory studies
Advantages: Specialist equipment available Uninterrupted environment Disadvantages: Lack of context Difficult to observe several users cooperating Appropriate If system location is dangerous Or impractical for constrained single user systems to allow controlled manipulation of use

4 Overview: Field studies
Advantages: Natural environment Context retained (though observation may alter it) Longitudinal studies possible Disadvantages: Distractions Noise Appropriate Where context is crucial for longitudinal studies

5 Types of Evaluation Formative evaluation Summative evaluation

6 Formative evaluation Repeatedly, as development proceeds
Purpose: to support iterative refinement Nature: structured Fairly informal Average of 3 major design/test/re-design Many minor cycles to check minor changes “The earlier poor design features or errors are detected, the easier and cheaper they are to correct”

7 Summative evaluation Done once Important in field or ‘beta’ testing
After implementation, ‘beta’-release Important in field or ‘beta’ testing Purpose: quality control Product is reviewed to check for approval Own specifications Prescribed standards, e.g. Health and Safety ISO Nature: formal, often involving statistical inferences

8 Goals of Evaluation Assess extent of Assess effect of
Measure the level of System functionality System usability Assess effect of Measure the result of Interface on user Identify specific problems

9 What to evaluate? Usability specifications at all lifecycle stages
Initial designs (pre-implementation) Partial Integrated Basically all representations produced ! Prototype at various stages Final implementation (α and β-release) Documentation

10 Evaluation Approaches/Methods
Some approaches are based on expert evaluation (experts) : Analytic (Cognitive Walkthrough) methods Heuristic (Review) methods Model-based methods. Some approaches are based on empirical evaluation (involves users) : Observational methods Query (Survey) methods. Experimental methods Statistical Validation An evaluation method must be chosen carefully and must be suitable for the job.

11 Evaluation Approaches/Methods
Interface Development User Involvement Analytic Specification No users Heuristic Specification & Prototype No users, role playing only Model based Observation Simulation & Prototype Real users Query Experiment Normal full Prototype Expert cos t s Empirical

12 Cognitive Walkthrough Heuristic Evaluation Model-based evaluation
Expert Evaluation Cognitive Walkthrough Heuristic Evaluation Model-based evaluation

13 Expert Evaluation Analytic (Cognitive Walkthrough) methods
Heuristic (Review) methods (L-23) Model-based methods

14 Cognitive Walkthrough
Proposed by Polson et al. Evaluates design on how well it supports user in learning task Usually performed by expert in cognitive psychology Expert ‘walks though’ design to identify potential problems using psychological principles Forms used to guide analysis

15 Cognitive Walkthrough (ctd)
For each task walkthrough considers What impact will interaction have on user? What cognitive processes are required? What learning problems may occur? Analysis focuses on goals and knowledge: does the design lead the user to generate the correct goals?

16 Heuristic Evaluation (L-23)
Proposed by Nielsen and Molich. Usability criteria (heuristics) are identified Design examined by experts to see if these are violated Example heuristics (Nielsen 10 rules) System behaviour is predictable System behaviour is consistent Feedback is provided Heuristic evaluation `debugs' design.

17 Model based evaluation (L-35-39)
Results from the literature used to support or refute parts of design. Care needed to ensure results are transferable to new design. Cognitive models used to filter design options E.g. GOMS prediction of user performance. Design rationale can also provide useful evaluation information

18 Empirical Evaluation

19 Empirical Evaluation Observational methods Query (Survey) methods
Experimental methods

20 Observational Methods
Think Aloud Cooperative evaluation Protocol analysis Automated analysis Post-task walkthroughs

21 Think Aloud User observed performing task
User asked to describe what he is doing and why, what he thinks is happening etc. Advantages Simplicity - requires little expertise Can provide useful insight Can show how system is actually use Disadvantages Subjective Selective Act of describing may alter task performance

22 Cooperative evaluation
Variation on think aloud User collaborates in evaluation Both user and evaluator can ask each other questions throughout Additional advantages Less constrained and easier to use User is encouraged to criticize system Clarification possible

23 Interviews Questionnaires
Query Techniques Interviews Questionnaires

24 Interviews Analyst questions user on one-to-one basis usually based on prepared questions Informal, subjective and relatively cheap Advantages Can be varied to suit context Issues can be explored more fully Can elicit user views and identify unanticipated problems Disadvantages Very subjective Time consuming

25 Questionnaires Set of fixed questions given to users Advantages
Quick and reaches large user group Can be analyzed more rigorously Disadvantages Less flexible Less probing

26 Questionnaires (ctd) Need careful design Styles of question
What information is required? How are answers to be analyzed? Styles of question General Open-ended Scalar Multi-choice Ranked

27 Experimental evaluation

28 Experimental evaluation
Controlled evaluation of specific aspects of interactive behaviour Evaluator chooses hypothesis to be tested A number of experimental conditions are considered which differ only in the value of some controlled variable. Changes in behavioural measure are attributed to different conditions

29 Experimental factors Subjects Variables Hypothesis Experimental design
Who – representative, sufficient sample Variables Things to modify and measure Hypothesis What you’d like to show Experimental design How you are going to do it

30 Subject groups Larger number of subjects  more expensive
Longer time to `settle down’ … even more variation! Difficult to timetable so … often only three or four groups

31 Variables Independent variable (IV) Dependent variable (DV)
Characteristic changed to produce different conditions E.g. Interface style, number of menu items Dependent variable (DV) Characteristics measured in the experiment E.g. Time taken, number of errors.

32 Hypothesis Prediction of outcome Null hypothesis:
Framed in terms of IV and DV E.g. “error rate will increase as font size decreases” Null hypothesis: States no difference between conditions Aim is to disprove this E.g. null hyp. = “no change with font size” (Eg: logos, images)

33 Experimental design Within groups design Between groups design
Each subject performs experiment under each condition. Transfer of learning possible Less costly and less likely to suffer from user variation. Between groups design Each subject performs under only one condition No transfer of learning More users required Variation can bias results.


Download ppt "Evaluation techniques"

Similar presentations


Ads by Google