Nigel G. Ward, Anais G. Rivera, Karen Ward, David G. Novick

Slides:



Advertisements
Similar presentations
DEVELOPING A METHODOLOGY FOR MS3305 CW2 Some guidance.
Advertisements

Tuning Jenny Burr August Discussion Topics What is tuning? What is the process of tuning?
Do Bystanders and Dialog Participants Differ in Preferences for Telecommunications Channels? -- The Effects of Noise and Delay -- Nigel Ward Anais G. Rivera.
User Interface Design Main issues: What is the user interface How to design a user interface.
Task-Centered User Interface Design who are the users? what are the tasks? plagiarize! iterative design –rough descriptions, mock-ups, prototypes test.
Interaksi Manusia Komputer – Marcello Singadji. design rules Designing for maximum usability – the goal of interaction design Principles of usability.
Member FINRA/SIPCThursday, November 12, 2009 Resource Menu Changes - Report User Experience Study | Kevin Cornwall.
On the road to the creation of situation-adaptive dialogue managers Ajay Juneja Dialog Seminar.
5/10/20151 Evaluating Spoken Dialogue Systems Julia Hirschberg CS 4706.
Ch 11 Cognitive Walkthroughs and Heuristic Evaluation Yonglei Tao School of Computing and Info Systems GVSU.
Part 4: Evaluation Days 25, 27, 29, 31 Chapter 20: Why evaluate? Chapter 21: Deciding on what to evaluate: the strategy Chapter 22: Planning who, what,
Heuristic Evaluation. Sources for today’s lecture: Professor James Landay: stic-evaluation/heuristic-evaluation.ppt.
Discount Evaluation Evaluating with experts. Agenda Online collaboration tools Heuristic Evaluation Perform HE on each other’s prototypes Cognitive Walkthrough.
Empirical Methods in Human- Computer Interaction.
What can humans do when faced with ASR errors? Dan Bohus Dialogs on Dialogs Group, October 2003.
Discount Evaluation Evaluating with experts. Agenda Part 4 preview Heuristic Evaluation Perform HE on each other’s prototypes Cognitive Walkthrough Perform.
Heuristic Evaluation Evaluating with experts. Discount Evaluation Techniques  Basis: Observing users can be time- consuming and expensive Try to predict.
Nine principles of design Simple and natural dialog Speak the user’s language Minimize user’s memory load Be consistent Provide feedback Provide clearly.
Evaluating with experts
SIMS 213: User Interface Design & Development Marti Hearst Tues Feb 13, 2001.
Chapter 7 design rules.
Usability and Evaluation Dov Te’eni. Figure ‎ 7-2: Attitudes, use, performance and satisfaction AttitudesUsePerformance Satisfaction Perceived usability.
Evaluation in HCI Angela Kessell Oct. 13, Evaluation Heuristic Evaluation Measuring API Usability Methodology Matters: Doing Research in the Behavioral.
Beyond Usability: Measuring Speech Application Success Silke Witt-Ehsani, PhD VP, VUI Design Center TuVox.
Study of Human factors in Software Engineering CSC 532 Sree Harsha Pothireddy.
Interactive Dialogue Systems Professor Diane Litman Computer Science Department & Learning Research and Development Center University of Pittsburgh Pittsburgh,
Discount Evaluation Evaluating with experts. Discount Evaluation Techniques Basis: – Observing users can be time-consuming and expensive – Try to predict.
Nielsen’s Ten Usability Heuristics
PG 1 Caroline V Nelson Speech Solutions Team Technical Lead Nortel Enterprise Multimedia Professional Services August 20, 2007 Success Criteria for Speech.
User Interface Design Main issues: What is the user interface How to design a user interface ©2008 John Wiley & Sons Ltd.
SEG3120 User Interfaces Design and Implementation
Chapter 6: Thinking about requirements and describing them.
1 Natural Language Processing Lecture Notes 14 Chapter 19.
Evaluating a UI Design Expert inspection methods Cognitive Walkthrough
Heuristic Evaluation Short tutorial to heuristic evaluation
Evaluation Methods - Summary. How to chose a method? Stage of study – formative, iterative, summative Pros & cons Metrics – depends on what you want to.
User Responses to Prosodic Variation in Fragmentary Grounding Utterances in Dialog Gabriel Skantze, David House & Jens Edlund.
Introduction to Evaluation “Informal” approaches.
Fall 2002CS/PSY Predictive Evaluation (Evaluation Without Users) Gathering data about usability of a design by a specified group of users for a particular.
Objectives of session By the end of today’s session you should be able to: Define and explain pragmatics and prosody Draw links between teaching strategies.
Predicting and Adapting to Poor Speech Recognition in a Spoken Dialogue System Diane J. Litman AT&T Labs -- Research
MULTIMODAL AND NATURAL COMPUTER INTERACTION Domas Jonaitis.
Design rules.
Designing the Inputs and User Interface
SIE 515 Design Evaluation Lecture 7.
Chapter 6: Checklists, Rating Scales & Rubrics
Human-Computer Interaction
Chapter 6. Data Collection in a Wizard-of-Oz Experiment in Reinforcement Learning for Adaptive Dialogue Systems by: Rieser & Lemon. Course: Autonomous.
Cognitive Behavior Modification
SIE 515 Design Rules Lecture 5.
Software Quality Engineering
Spoken Dialogue Systems
Integrating Learning of Dialog Strategies and Semantic Parsing
Dialog Design 4 Speech & Natural Language
Creating User Interfaces
Sharing on EL Oral Communication
A NEW FACE OF THE TECHNICAL COMMUNICATOR – UX IS OUR STRENGTH – WE DESIGN TO WRITE BY CHRIS GANTA © 2016, STC INDIA CHAPTER.
Spoken Dialogue Systems
One-timer?. A new face of the technical communicator – UX is our strength – we design to write.
Leadership Chapter 7 – Path-Goal Theory Northouse, 4th edition.
Evaluation.
Nilesen 10 hueristics.
Chapter 7 design rules.
Formative Evaluation cs3724: HCI.
Chapter 7 design rules.
Chapter 7 design rules.
Designing & Conducting Formative Evaluation
Education & AI High level discussion points relating ‘techniques / tools’ to possible ‘projects’
Chapter 7 design rules.
Evaluation of a multimodal Virtual Personal Assistant Glória Branco
Presentation transcript:

Nigel G. Ward, Anais G. Rivera, Karen Ward, David G. Novick Some Usability Issues and Research Priorities in Spoken Dialog Applications Nigel G. Ward, Anais G. Rivera, Karen Ward, David G. Novick Interactive Systems Group, The University of Texas at El Paso Questions Why are many dialog systems poor in usability? are developers failing to use best practice? are researchers missing the important issues? Analysis Phase 1: Usability Event Analysis Three labelers identified 115 times when the user experienced poor system usability 62 times when the subject experienced deft operator actions and categorized them by severity and likely causes Recognition and Understanding +++ / +++ / +++* direct effects on task completion, satisfaction effects on user behavior (slow, considered) effects on designer behavior (more explicit prompts, leading to time cost) Time-Outs +++ / + / ++* delayed reaction when user wants a reprompt Endpointing and Responsiveness +++ / + / ++* delayed reaction after the user finishes speaking Generation and Synthesis +++ / + / +* conservative, invariant speaking rate, causing time-cost pragmatic force unclear Feedback + / / +* unclear indications of dialog activity/ status weak, ineffective dialog control Adaptation of Vocab., Speaking Rate… ++ / /* Prosodically-Marked Speech Acts + / / +* Detecting User Self-Talk + / / +* Other + / + / + * Research Issues Identified Analysis Phase 2: Time Analysis We compared the cummulative times for various dialog activities across the two conditions. Methods Each subject performed the same task twice, once with an automated system, and once with a human operator. We observed and categorized the differences. Advantages: reveals missed opportunities, not just clear errors usability-centered, not performance-centered empirical, not theory-based exploratory, not prematurely quantitative (c.f. Walker et al. 200, Dybkjaer & Bernsen 2000, Moller 2005 …) Dialog Activity system- human human- difference normal system-side utterances 32% 37% 29% normal user utterances 13% 41% -6% error-recovery related utterances 17% 4% 27% system-side silences 6% 20% user silences 14% 11% 16% simultaneous talk 1% 0% experimenter interventions 8% Totals 100% User Operator Spoken Dialog System compare Analysis Phase 3: Relating Usability Issues, System Capabilites, and Research Issues Clear, expressive, flexible synthesis Natural language generation Accurate, recognition complete Normally users behaving sensitivity Prosodic Domain and Corpus The domain was credit card billing support. Subjects were given three structured tasks in each condition. The automated system was based on Nuance Voice Builder. The operator matched the system’s vocabulary and formality level. 20 subjects. 53 minutes total. Concise prompts Adaptive speaking rate Timeouts * impact on time/task-completion/stress Endpointing feedback Informative Responsiveness Easy recovery error Summary have identified outstanding research issues: endpointing, responsiveness, appropriate feedback, etc. have shown how experiences with applications can determine research priorities Vocabulary adaptation Initial Observations With the system, versus with human: Task-completion was lower (58% vs 77%). Time to completion per task was higher (avg. 130 vs 40 sec.) Low stress and cognitive load Quick accomplishment High task completion Domain competence Attractiveness Usability Perceived value