Stanford hci group / cs376 u Scott Klemmer · 16 November 2006 Speech & Multimod al.

Slides:



Advertisements
Similar presentations
PUI 98 Leveraging Human Capabilities in Perceptual Interfaces George G. Robertson Microsoft Research George G. Robertson Microsoft Research.
Advertisements

Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
Empirical and Data-Driven Models of Multimodality Advanced Methods for Multimodal Communication Computational Models of Multimodality Adequate.
Overview of Graphic Systems
Stanford hci group / cs376 research topics in human-computer interaction Tangible Interaction / Augmented Reality Scott Klemmer.
Stanford hci group / cs376 research topics in human-computer interaction Vision-based Interaction Scott Klemmer 17 November 2005.
A multimodal dialogue-driven interface for accessing the content of recorded meetings Agnes Lisowska ISSCO/TIM/ETI University of Geneva IM2.MDM Work done.
John Hu Nov. 9, 2004 Multimodal Interfaces Oviatt, S. Multimodal interfaces Mankoff, J., Hudson, S.E., & Abowd, G.D. Interaction techniques for ambiguity.
Stanford hci group / cs376 Research Topics in Human-Computer Interaction Design Tools 01 November 2005.
Stanford hci group / cs376 research topics in human-computer interaction Multimodal Interfaces Scott Klemmer 15 November 2005.
Stanford hci group / cs376 Design Processes and Tools Sharon Lin Bowen Li.
Single Display Groupware Ana Zanella - CPSC
CS335 Principles of Multimedia Systems Multimedia and Human Computer Interfaces Hao Jiang Computer Science Department Boston College Nov. 20, 2007.
Stanford hci group / cs376 research topics in human-computer interaction Tangible Interaction / Augmented Reality Scott Klemmer.
Designing a User Interface for People with Disabilities u u
Support for Palm Pilot Collaboration Including Handwriting Recognition.
T.Sharon-A.Frank 1 Internet Resources Discovery (IRD) Multimedia - Motivation & Definition.
Stanford hci group / cs376 research topics in human-computer interaction I/O Toolkits Scott Klemmer 29 November 2005.
Stanford hci group / cs376 research topics in human-computer interaction UI Software Tools Scott Klemmer 27 October 2005.
Data: A collection of raw facts and figures. It may consist of numbers, characters, symbols or pictures. Information: Organized and processed form of.
Lesson Objectives To understand that users with disabilities require different input and output devices To be able to identify these devices and explain.
Mobile Multimodal Applications. Dr. Roman Englert, Gregor Glass March 23 rd, 2006.
User Interface Design Tools for the Future Multimodal UI Research in the HCC James A. Landay Jason Hong, Scott Klemmer, Jimmy Lin, Mark Newman, & Anoop.
ASSISTIVE TECHNOLOGY PRESENTED BY ABDUL BARI KP. CONTENTS WHAT IS ASSISTIVE TECHNOLOGY? OUT PUT: Screen magnifier Speech to Recogonizing system Text to.
Copyright John Wiley & Sons, Inc. Chapter 3 – Interactive Technologies HCI: Developing Effective Organizational Information Systems Dov Te’eni Jane.
Towards a Unified Interaction Framework for Ubicomp User Interfaces Jason I. Hong Scott Lederer Mark W. Newman G r o u p f o r User Interface Research.
Jeanne Corcoran, OTD, OTR/L October 6 th, The mission of Envision Center for Data Perceptualization is to serve, support, and collaborate with faculty,
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
11.10 Human Computer Interface www. ICT-Teacher.com.
Break-out Group # D Research Issues in Multimodal Interaction.
Fall UI Design and Implementation1 Lecture 20: HCI Research Topics.
1 Computer Graphics Assistant Professor Dr. Sana’a Wafa Al-Sayegh 2 nd Semester ITGD3107 University of Palestine.
Human Computer Interaction © 2014 Project Lead The Way, Inc.Computer Science and Software Engineering.
Human-Computer Interaction
Computer Ports Mouse Port (Input).
Stanford hci group / cs376 u Scott Klemmer · 26 October 2006 Design Tools.
Cs3724: Introduction to HCI Dr. Chris North GTA: Purvi Saraiya.
Stanford hci group / cs147 u 27 November 2007 Ubiquitous Computing & “Natural” Interaction Scott Klemmer tas: Marcello Bastea-Forte,
People today are limited to a mouse and keyboard when using a computer There are little to no alternatives out in the market at this moment Natural human.
Computer Parts and Functions. How A Computer Works.
Stanford hci group / cs376 Research Topics in Human-Computer Interaction Design Tools Ron B. Yeh 26 October 2004.
Copyright John Wiley & Sons, Inc. Chapter 3 – Interactive Technologies HCI: Developing Effective Organizational Information Systems Dov Te’eni Jane.
1 Prof. Dr. Nizamettin AYDIN
1 Interaction Devices CIS 375 Bruce R. Maxim UM-Dearborn.
Immersive Virtual Characters for Educating Medical Communication Skills J. Hernendez, A. Stevens, D. S. Lind Department of Surgery (College of Medicine)
Unit 6 of COMP648 User Interface and Interaction Methods Dr Oscar Lin School of Computing and Information Systems Faculty of Science and Technology Athabasca.
Stanford hci group / cs376 research topics in human-computer interaction Information Visualization Scott Klemmer 03 November.
Stanford hci group / cs376 u Jeffrey Heer · 19 May 2009 Speech & Multimodal Interfaces.
W3C Multimodal Interaction Activities Deborah A. Dahl August 9, 2006.
Software Architecture for Multimodal Interactive Systems : Voice-enabled Graphical Notebook.
Stanford hci group / cs376 u Scott Klemmer · 28 November 2006 Vision- Based Interacti on.
Iterative Design and Evaluation of Event Architecture for Pen-and-Paper Interfaces HCI Group Stanford University Ron B. Yeh Andreas Paepcke Scott R. Klemmer.
COMP413: Computer Graphics Overview of Graphics Systems Chapter 1.
Papier-Mache: Toolkit Support for Tangible Input HCI Group University of California Scott R. Klemmer Jack Li James Lin DUB Group University of Washington.
Speech and multimodal Jesse Cirimele. papers “Multimodal interaction” Sharon Oviatt “Designing SpeechActs” Yankelovich et al.
Multimodal and Natural computer interaction Evelina Stanevičienė.
MULTIMODAL AND NATURAL COMPUTER INTERACTION Domas Jonaitis.
Identify internal hardware devices (e. g
Human Computer Interaction (HCI)
Human Computer Interaction (HCI)
DESIGNING WEB INTERFACE Presented By, S.Yamuna AP/CSE
Multimodal Interfaces
Vision-based Interaction
Tangible Interaction & Augmented Reality
Multimodal Human-Computer Interaction New Interaction Techniques 22. 1
Speech & Multimodal Scott Klemmer · 16 November 2006.
Map of Human Computer Interaction
Professor John Canny Spring 2003
Professor John Canny Spring 2004
Mapping GUIs to Auditory Interfaces
Presentation transcript:

stanford hci group / cs376 u Scott Klemmer · 16 November 2006 Speech & Multimod al

2 Some hci definitions  Multimodal generally refers to an interface that can accept input from two or more combined modes  Multimedia generally refers to an interface that produces output in two or more modes  The vast majority of multimodal systems have been speech + pointing (pen or mouse) input, with graphical (and sometimes voice) output

3 Canonical App: Maps  Why are maps so well-suited?  A visual artifact for computation (Hutchins)

4 What is an interface  Is it an interface if there’s no method for a user to tell if they’ve done something?  What might an example be?  Is it an interface if there’s no method for explicit user input?  example: health monitoring apps

5 Sensor Fusion  multimodal = multiple human channels  sensor fusion = multiple sensor channels  Example app: Tracking people (1 human channel)  might use: RFID + vision + keyboard activity + …  I disagree with the Oviatt paper  Speech + lips is sensor fusion, not multimodality

6 What constitutes a modality?  To some extent, it’s a matter of semantics  Is pen a different modality than a mouse?  Are two mice different modalities if one is controlling a gui, and the other controls a tablet-like ui?  Is a captured modality the same as an input modality?  How does the audio notebook fit into this?

7 Input modalities  mouse  pen: recognized or unrecognized  speech  non-speech audio  tangible object manipulation  gaze, posture, body-tracking  Each of these experiences has different implementing technologies  e.g., gaze tracking could be laser-based or vision-based

8 Output modalities  Visual displays  Raster graphics, Oscilloscope, paper printer, …  Haptics: Force Feedback  Audio  Smell  Taste

9 Dual Purpose Speech

10 Why multimodal?  Hands busy / eyes busy  Mutual disambiguation  Faster input  “More natural”

11 On Anthropomorphism  The multimodal community grew out of the AI and speech communities  Should human communication with computers be as similar as possible to human-human communication?

12 Multimodal Software Architectures  OAA, AAA, OOPS

13 Next Time… Vision-Based Interaction Computer Vision for Interactive Computer Graphics, William T. Freeman, Yasunari Miyake, Ken- ichi Tanaka, David B. Anderson, Paul A. Beardsley, Chris N. Dodge, Michal Roth, Craig D. Weissman, William S. Yerazunis, Hiroshi Kage, Kazuo Kyuma A Design Tool for Camera-based Interaction, Jerry Alan Fails and Dan R. Olsen

CS547 Tomorrow  Ben Shneiderman, University of Maryland – Science 2.0: The Design Science of Collaboration 14