CMPUT 301: Lecture 31 Out of the Glass Box Martin Jagersand Department of Computing Science University of Alberta.

Slides:



Advertisements
Similar presentations
Creating a Multimodal Design Environment Using Speech and Sketching Aaron Adler Student Oxygen Workshop September 12, 2003.
Advertisements

Chapter 10 universal design.
How To Design a User Interface For People With Disabilties
Natural Language Systems
Created By: Lauren Snyder, Juliana Gerard, Dom Williams, and Ryan Holsopple.
Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
Human Computer Interaction
Input to the Computer * Input * Keyboard * Pointing Devices
universal design by Prof. Dr. Sajjad Mohsin
1 CS 430: Information Discovery Lecture 22 Non-Textual Materials 2.
Class 6 LBSC 690 Information Technology Human Computer Interaction and Usability.
Copyright © 2006 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Technology Education Copyright © 2006 by The McGraw-Hill Companies,
ISTD 2003, Audio / Speech Interactive Systems Technical Design Seminar work: Audio / Speech Ville-Mikko Rautio Timo Salminen Vesa Hyvönen.
Universal Design CSE 491 Michigan State Fall 2007 E. Kraemer.
Non-Traditional Interfaces CSCI324, IACT403, IACT 931, MCS9324 Human Computer Interfaces.
Direct Manipulation and Vision and Touch based User Interfaces Martin Jägersand Johns Hopkins University CIPS and CISST groups (Computational Interaction.
Auditory User Interfaces
Adapted from CTAE Resources Network PROFITT Curriculum Basic Computer Skills Module 1 Hardware.
Text-To-Speech System for Marathi Miss. Deepa V. Kadam Indian Institute of Technology, Bombay.
CHAPTER 2 Input & Output Prepared by: Mrs.sara salih 1.
Digital Sound and Video Chapter 10, Exploring the Digital Domain.
Introduction to Multimedia. The beginning ( History )… 1945 : “…a device in which one stores all his books, records and communications, and which is mechanized.
Universal Design Material from Authors of Human Computer Interaction Alan Dix, et al.
Assistive Technology Russell Grayson EDUC 504 Summer 2006.
Chapter 11: Interaction Styles. Interaction Styles Introduction: Interaction styles are primarily different ways in which a user and computer system can.
Copyright John Wiley & Sons, Inc. Chapter 3 – Interactive Technologies HCI: Developing Effective Organizational Information Systems Dov Te’eni Jane.
Interaction Media & Communication, Department of Computer Science, Queen Mary University of London THE INFLUENCE.
Enlightening minds. Enriching lives. Tamil Digital Industry Badri Seshadri K.S.Nagarajan New Horizon Media.
Knowledge Base approach for spoken digit recognition Vijetha Periyavaram.
School of Computer Science & Information Technology G6DPMM - Lecture 11 Media Design II – Sound.
Multimedia is a program that combines:
Input Devices What is input? Everything we tell the computer is input.
Chapter 10 universal design.
11.10 Human Computer Interface www. ICT-Teacher.com.
CP SC 881 Spoken Language Systems. 2 of 23 Auditory User Interfaces Welcome to SLS Syllabus Introduction.
Lecture 6 User Interface Design
IAS2223: Human Computer Interaction Chapter 5: Universal Design.
I.T MEDIA MAISRUL www.roelsite.yolasite.com
CMPUT 301: Lecture 18 Usability Paradigms and Principles Lecturer: Martin Jagersand Department of Computing Science University of Alberta Notes based on.
Object Orientated Data Topic 5: Multimedia Technology.
COMPUTER PARTS AND COMPONENTS INPUT DEVICES
Multi-Sensory Systems z More than one sensory channel in interaction y e.g. sounds, text, hypertext, animation, video, gestures, vision z Used in a range.
GENESIS OF VIRTUAL REALITY  The term ‘Virtual reality’ (VR) was initially coined by Jaron Lanier, founder of VPL Research (1989)..
Human-Computer Interaction
22CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University. Lecture 10: Advanced Input.
1 CS 430: Information Discovery Lecture 22 Non-Textual Materials: Informedia.
Microsoft Assistive Technology Products Brought to you by... Jill Hartman.
E.g.: MS-DOS interface. DIR C: /W /A:D will list all the directories in the root directory of drive C in wide list format. Disadvantage is that commands.
KAMI KITT ASSISTIVE TECHNOLOGY Chapter 7 Human/ Assistive Technology Interface.
1 Human Computer Interaction Week 5 Interaction Devices and Input-Output.
HCI 입문 Graphics Korea University HCI System 2005 년 2 학기 김 창 헌.
USER INTERFACE DESIGN (UID). Introduction & Overview The interface is the way to communicate with a product Everything we interact with an interface Eg.
Human Computer Interaction. Outline  Multimodal HCI Issues  More than 1 sensory channel in interaction  e.g. Sound, text, hypertext, animation, video,
Copyright John Wiley & Sons, Inc. Chapter 3 – Interactive Technologies HCI: Developing Effective Organizational Information Systems Dov Te’eni Jane.
1 Interaction Devices CIS 375 Bruce R. Maxim UM-Dearborn.
What is Multimedia Anyway? David Millard and Paul Lewis.
Chapter 10 universal design. Universal Design “The process of designing products so that they can be used by as many people as possible in as many situations.
MULTIMODAL AND NATURAL COMPUTER INTERACTION Domas Jonaitis.
Presented By, S.Yamuna AP/CSE
11.10 Human Computer Interface
Universal design-2.
Human Computer Interaction Lecture 20 Universal Design
CEN3722 Human Computer Interaction Advanced Interfaces
CSE310 Human-Computer Interaction
Multi-Sensory Systems
Chapter 10 universal design.
Chapter 10 universal design.
Chapter 10 universal design.
universal design (web accessibility)
Human Computer Interaction Lecture 19 Universal Design
Presentation transcript:

CMPUT 301: Lecture 31 Out of the Glass Box Martin Jagersand Department of Computing Science University of Alberta

2 Overview Idea: –why only use the sense of vision in user interfaces? –increase the bandwidth of the interaction by using multiple sensory channels, instead of overloading the visual channel

3 Overview Multi-sensory systems: –use more than one sensory channel in interaction –e.g., sound, video, gestures, physical actions etc.

4 Overview Usable senses: –sight, sound, touch, taste, smell, –Haptics, proprioception and accelerations –each is important on its own –together, they provide a fuller interaction with the natural world

5 Overview Usable senses: –computers rarely offer such a rich interaction –we can use sight, sound, and sometimes touch –Flight simulators and some games uses accelerations to create a multimodal immersion experience. –we cannot (yet) use taste or smell

6 Overview Multi-modal systems: –use more than one sense in the interaction –e.g., sight and sound: a word processor that speaks the words as well as rendering them on the screen

7 Overview Multi-media systems: –use a number of different media to communicate information –e.g., a computer-based teaching system with video, animation, text, and still images

8 Speech Human speech: –natural mastery of language –instinctive, taken for granted –difficult to appreciate the complexities –potentially a useful way to extend human- computer interaction

9 Speech Structure: –phonemes (English) –40 (24 consonant and 16 vowel sounds) –basic atomic units of speech –sound slightly different depending on context …

10 Speech Structure: –allophones: –120 to 130 –all the sounds in the language –count depends on accents

11 Speech Structure: –morphemes –basic atomic units of language –part or whole words –formed into sentences using the rules of grammar

12 Speech Prosody: –variations in emphasis, stress, pauses, and pitch to impart more meaning to sentences Co-articulation: –the effect of context on the sound –transforms phonemes into allophones

13 Speech Recognition Problems: –different people speak differently (e.g., accent, stress, volume, etc.) –background noises –“ummm …” and “errr …” –speech may conflict with complex cognition

14 Speech Recognition Issues: –recognizing words is not enough –need to extract meaning –understanding a sentence requires context, such as information about the subject and the speaker

15 Speech Recognition Phonetic typewriter: –developed for Finnish (a phonetic language) –trained on one speaker, tries to generalize to others –uses neural network that clusters similar sounds together, for a character –poor performance on speakers it has not been trained on –requires a large dictionary of minor variations

16 Speech Recognition Currently: –single user, limited vocabulary systems can work satisfactorily –no general user, general vocabulary systems are commercial successful, yet Current commercial examples: –Simple telephone based UI such as Train schedule information systems

17 Speech Recognition Potential: –for users with physical disabilities –for lightweight, mobile devices –for when user’s hands are already occupied with a manual task (auto mechanic, surgeon)

18 Speech Synthesis What: –computer-generated speech –natural and familiar way of receiving information

19 Speech Synthesis Problems: –human find it difficult to adjust to monotonic, non-prosodic speech –computer needs to understand natural language and the domain –Speech is transient (hard to review or browse) –produces noise in the workplace or requires headphones (intrusive)

20 Speech Synthesis Potential: –screen readers –read a textual display to a visually impaired person –warning signals –spoken information especially for aircraft pilots whose visual and haptic channels are busy

21 Speech Synthesis Virtual newscaster (Ananova)

22 Uninterpreted Speech What: –fixed, recorded speech –e.g., played back in airport announcements –e.g., attached as voice annotation to files

23 Uninterpreted Speech Digital processing: –change playback speed without changing pitch –to quickly scan phone messages –to manually transcribe voice to text –to figure out the lyrics and chords of a song –spatialization and environmental effects

24 Non-Speech Sound What: –boings, bangs, squeaks, clicks, etc. –commonly used in user interfaces to provide warnings and alarms

25 Non-Speech Sound Why: –fewer typing mistakes with key clicks –video games harder without sound

26 Non-Speech Sound? D’oh!

27 Non-Speech Sound Dual mode displays: –information presented along two different sensory channels –e.g., sight and sound –allows for redundant presentation –user uses whichever they find easiest –allows for resolution of ambiguity in one mode through information in the other

28 Non-Speech Sound Dual mode displays: –humans can react faster to auditory than visual stimuli –sound is especially good for transient information that would otherwise clutter a visual display –sound is more language and culture independent (unlike speech)

29 Non-Speech Sound Auditory icons: –use natural sounds to represent different types of objects and actions in the user interface –e.g., breaking glass sound when deleting a file –direction and volume of sounds can indicate position and importance/size –SonicFinder –not all actions have an intuitive sound

30 Non-Speech Sound Earcons: –synthetic sounds used to convey information –structured combinations of motives (musical notes) to provide rich information

31 Non-Speech Sound Earcons:

32 Handwriting Recognition Handwriting: –text and graphic input –complex strokes and spaces –natural

33 Handwriting Recognition Problems: –variation in handwriting between users –variation from day to day and over years for a single user –variation of letters depending on nearby letters

34 Handwriting Recognition Currently: –limited success with systems trained on a few users, with separated letters –generic, multi-user, cursive text recognition systems are not accurate enough to be commercially successful Current applications e.g. pre-sorting of mail (but human has to assist with failures)

35 Handwriting Recognition Newton: –printing or cursive writing recognition –dictionary of words –contextual recognition –fine tune spacing and letter shapes –fine tune recognition speed –learn handwriting over time

36 Handwriting Recognition Newton:

37 End What did I learn today? What questions do I still have?