Audiovisual Attentive User Interfaces Attending to the needs and actions of the user Paulina Modlitba T-121.900 Seminar on User Interfaces and Usability.

Slides:



Advertisements
Similar presentations
Some Reflections on Augmented Cognition Eric Horvitz ISAT & Microsoft Research November 2000 Some Reflections on Augmented Cognition Eric Horvitz ISAT.
Advertisements

Structured Design The Structured Design Approach (also called Layered Approach) focuses on the conceptual and physical level. As discussed earlier: Conceptual.
Design, prototyping and construction
Map of Human Computer Interaction
A.Mani Kumari 07A21F0007.  Introduction  Definition  Working  Applications  Drawbacks  Conclusion.
Clipping Lists & Change Borders: Improving Multitasking Efficiency with Peripheral Information Design Mary Czerwinski George Robertson Desney Tan Microsoft.
Input & Output Devices ASHIMA KALRA.
Input to the Computer * Input * Keyboard * Pointing Devices
Design Activities in Usability Engineering laura leventhal and julie barnes.
Human Computer Interface. HCI and Designing the User Interface The user interface is a critical part of an information system -- it is what the users.
Class 6 LBSC 690 Information Technology Human Computer Interaction and Usability.
MUltimo3-D: a Testbed for Multimodel 3-D PC Presenter: Yi Shi & Saul Rodriguez March 14, 2008.
Psychological Aspects Presented by Hanish Patel. Overview  HCI (Human Computer Interaction)  Overview of HCI  Human Use of Computer Systems  Science.
Cognetics and the Locus of Attention
Psychological Aspects
Discussion Silvia Lindtner INF 132 April 07. Fitts’ law - recap A predictive model of time to point at an object Help decide the location and size of.
Input devices, processing and output devices Hardware Senior I.
Introduction ‘Have you ever played video games before? Look at the joystick movement. When you move the joystick to the left, the plane on the TV screen.
Design, goal of design, design process in SE context, Process of design – Quality guidelines and attributes Evolution of software design process – Procedural,
 At the end of this class, students are able to  Describe definition of input devices clearly  List out the examples of input devices  Describe.
Many other computer components also must communicate with the CPU. In fact, each hardware input, output, or storage device requires these elements.
IT Introduction to Information Technology CHAPTER 05 - INPUT.
Knowledge Systems Lab JN 8/24/2015 A Method for Temporal Hand Gesture Recognition Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
INPUT DEVICES. KEYBOARD Most common input device for a computer.
   Input Devices Main Memory Backing Storage PROCESSOR
1. Human – the end-user of a program – the others in the organization Computer – the machine the program runs on – often split between clients & servers.
Spring /6.831 User Interface Design and Implementation1 Lecture 3: Visibility.
2-3 note. 2 Peripheral Devices “Peripheral devices” are hardware plugged into ports or connected to a computer wirelessly. These devices can be for input,
Mobile HCI Presented by Bradley Barnes. Mobile vs. Stationary Desktop – Stationary Users can devote all of their attention to the application. Very graphical,
Oct 30, 2006 LUONNOS Navigation techniques for construction industry product models Jukka Rönkkö, HUT/VTT
2003Lenko Grigorov, CISC 839 eyePROXY Lenko Grigorov, CISC 839 Supervisor: Roel Vertegaal Additional support by Skaburskis A and Changuk S.
Two Handed and Gaze Input Stanford and Princeton Lecture Nov 29, 1999 Shumin Zhai.
CGMB214: Introduction to Computer Graphics
CSC 480 Software Engineering Lecture 19 Nov 11, 2002.
©RavichandranUser interface Slide 1 User interface design.
Denis Gouin Valérie Lavigne Alexandre Bergeron-Guyard Innovative Interfaces and Interactions Group Intelligence and Information Section DRDC Valcartier.
EE 492 ENGINEERING PROJECT LIP TRACKING Yusuf Ziya Işık & Ashat Turlibayev Yusuf Ziya Işık & Ashat Turlibayev Advisor: Prof. Dr. Bülent Sankur Advisor:
Josh Stephens Comp Characteristics Degrees of Freedom: particular, independent way that a body moves in space Input type/Frequency of data: Discrete:
CMPUT 301: Lecture 18 Usability Paradigms and Principles Lecturer: Martin Jagersand Department of Computing Science University of Alberta Notes based on.
COMPUTER PARTS AND COMPONENTS INPUT DEVICES
GENESIS OF VIRTUAL REALITY  The term ‘Virtual reality’ (VR) was initially coined by Jaron Lanier, founder of VPL Research (1989)..
Gaze-Controlled Human-Computer Interfaces Marc Pomplun Department of Computer Science University of Massachusetts at Boston Homepage:
Human Computer Interaction © 2014 Project Lead The Way, Inc.Computer Science and Software Engineering.
卓越發展延續計畫分項三 User-Centric Interactive Media ~ 主 持 人 : 傅立成 共同主持人 : 李琳山,歐陽明,洪一平, 陳祝嵩 水美溫泉會館研討會
 Input Devices Input Devices  Examples of Input Devices Examples of Input Devices  Keyboard Keyboard  Pointing Devices Pointing Devices Mouse Joystick.
Input Devices Lecture 3 Input Devices Md. Mahbubul Alam, PhD PRESENTED BY MD. MAHBUBUL ALAM, PHD 1.
WHAT IS HARDWARE? Hardware refers to the set of components that make up the internal part of a computer. In case of the personal computers, it doesn’t.
ENTERFACE 08 Project 1 “MultiParty Communication with a Tour Guide ECA” Mid-term presentation August 19th, 2008.
KAMI KITT ASSISTIVE TECHNOLOGY Chapter 7 Human/ Assistive Technology Interface.
1 Human Computer Interaction Week 5 Interaction Devices and Input-Output.
Gaming ISV TOBII CONFIDENTIAL INFORMATION. Imagine a computer that knows where you want to point before you do  By looking at your point of gaze the.
Aiming Computing Technology at Enhancing the Quality of Life of People with ALS Some Sketches on Directions in Minimal Signaling Communication Communication.
Chapter 5:User Interface Design Concepts Of UI Interface Model Internal an External Design Evaluation Interaction Information Display Software.
Chapter 2 HAEDWAER.
Introduction to Computing Slides By ADEELA MUSTAFA.
Stanford hci group / cs376 u Jeffrey Heer · 19 May 2009 Speech & Multimodal Interfaces.
IMPROVING THE HUMAN TECHNOLOGY INTERFACE NURSING INFORMATICS CHAPTER 4 1.
Prepared byMIDHUN.T CHM S5.  Aims at creating computational machines that have perceptual and sensory ability like those of human beings.  Use camera.
Presented by TRIPTI TRIPATHI EC Final year. WHAT IS BLUE EYE TECHNOLOGY? Aim at creating computational machine that have perceptual and sensory ability.
Chapter 6 : User interface design
Lesson 4 Alternative Methods Of Input.
EYE TRACKING TECHNOLOGY
Standard Methods of Input.
Ubiquitous Computing and Augmented Realities
Lesson 4 Alternative Methods Of Input.
NBKeyboard: An Arm-based Word-gesture keyboard
Attentive User Interfaces
What is blue eyes ? aims on creating computational machines that have perceptual and sensory ability like those of human beings. interactive computer.
Map of Human Computer Interaction
Higher School of Economics , Moscow, 2016
Presentation transcript:

Audiovisual Attentive User Interfaces Attending to the needs and actions of the user Paulina Modlitba T Seminar on User Interfaces and Usability

What is an Attentive User Interface? (1/2) Negotiate the timing and volume of communication with the user Use specific input, output and turn-taking techniques to determine what task, device or person a user is attending to User’s presence, orientation, speech activity and gaze and statistically modeling attention and interaction are detected

Four characteristic components –visual attention –turn-taking techniques –modeling techniques for the attention –focus and context displays and visualisation Dürsteler (2003) What is an Attentive User Interface? (2/2)

Why are they needed? Roel Vertegaal (2003) Multiple ubiquitous computing devices lead to a growing demands on users’ attention Metaphor: modern traffic light system –Sensors –Statistical models of traffic volume –Peripheral displays (traffic lights) Disruptive effect of interruptions can be avoided

Evolution of human-machine interaction 1960s-1980s: many-one1980s-1990s: one-one 2000s-2010s: many-many1990s-2000s: one-many

Visual attention Eye-gaze tracking: detecting the user’s visual focus of attention Operate by sending an infrared light source toward the user’s eye Provides information about the context Central I/O channel in communication Limitations in existing hardware/software Biological limitations

Reasons for implementing gaze tracking Kaur et al. (2003) The gaze location is the only reliable predictor of the locus of visual attention Gaze can be used as a “natural” mode of input that avoids the need for learned hand-eye coordination Gaze selection of screen objects is expected to be significantly faster than the traditional hand- eye coordination Gaze allows for hands-free interaction

Current issues Limited size of fovea (1-3°) Subconscious eye movements Eyes are not control organs (Zhai et al., 2003) No natural analogy to current input devices, e.g. mouse Gaze is always active (Kaur et al., 2003)

Current state Eye-gaze control used as an additional input channel Provides context to the action Combined with manual input gaze tracking can improve the robustness and reliability of a system

EASE Chinese Input (1/2) Zhai et al. (2002) Supports pinyin type-writing –official Chinese phonetic alphabet based on Roman characters –Chinese characters are homophonic - each syllable corresponds to several Chinese characters –When the user types the pinyin of a character, a number of possible characters with the same pronunciation are displayed

Normally, user chooses a character by pressing a number on the keyboard With EASE user only has to press the spacebar as soon as he or she sees the wished-for character in the list The system selects the character closest to the user’s current gaze location EASE Chinese Input (2/2)

Speech recognition (1/2) Limited technology, despite extensive research and progress Crucial issues –error rate of speech recognition engines and how these errors can be reduced –the effort required to port the speech technology applications between different application domains or languages (Deng & Huang, 2004)

Three directions for enhancing the technique –improve the microphone ergonomics for enhancing the signal-to-noise ratio –equipping speech recognizers with the ability to learn and to correct errors –add semantic (meaning) and pragmatic (application context) knowledge (Deng & Huang, 2004) Speech recognition (2/2)

Multimodal interfaces Can provide more natural human-machine interaction Improves the robustness of the interaction by using redundant or complementary information Today: usually gaze/speech + manual control (e.g. mouse) Future: gaze + speech, gaze, speech

Main issue Shumin Zhai (2003) “We need to design unobtrusive, transparent and subtle turn-taking processes that coordinate attentive input with the user’s explicit input in order to contribute to the user’s goal without the burden of explicit dialogues.”

Manual and Gaze Input Cascaded (MAGIC) Pointing interaction technique that utilizes eye movement to assist the control task Zhai et al. have constructed two MAGIC pointing techniques, one liberal and one conservative (Zhai et al., 1999)

Liberal approach (1/2) The cursor is warped to every new object that the user looks at The user can then manually take control of the cursor near (or on) the target, or ignore it and search for the next target New target defined by distance (e.g. 120 pixels) from the current cursor position Issues: pro-active (cursor waits readily); overactive (gaze enough to move cursor)

Liberal approach (2/2)

Conservative approach (1/2) Warps the cursor to a target when the manual input device has been actuated Once moved, the cursor appears in motion towards the target Hence, the cursor never jumps directly to a target that the user does not intend to obtain May be slower than the liberal approach

Conservative approach (2/2)

EyeCOOK Bradbury et al. (2003) Multimodal attentive cookbook that helps unaccustomed computer users cook a meal User interacts with the eyeCOOK system by using eye- gaze and speech commands System responds visually and verbally The system replaces the object of the user’s gaze with the word “this” If the user’s gaze can not be tracked by the eyeCOOK system the user has to specify the target verbally

EyeCOOK in Page Display Mode

GAZE-2 Vertegaal et al, 2003 A new group video conferencing system that uses gaze-controlled cameras to convey eye- contact Consists of a video tunnel that makes it possible to place cameras behind the participant images on the screen system automatically directs the video cameras in this tunnel using a gaze tracker by selecting the camera closest to the user’s current focus of attention (gaze location)

GAZE-2 system structure

3D rendering The 2D video images of the participants are displayed in a 3D virtual meeting room and are automatically rotated to face the participant each user is looking at. In the picture bellow, everyone is looking at the left person, who’s image is broadcasted in a higher resolution.

Turn-taking in video conferencing Misunderstandings cause interruptions Eye contact plays an important role in turn-taking (Vertegaal, et al., 2003)

References Vertegaal, et al., 2003 Bradbury et al. (2003) Zhai et al., 1999 Dürsteler (2003) Vertegaal (2003) Kaur et al. (2003) Shumin Zhai (2003) Zhai et al. (2002) (Deng & Huang, 2004)

Things missing Are attentive user interfaces better in following the user in order to "capture his/her context" to make proactive actions for him/her, or are they better used as input devices (an approach you take). The distinction between explicit and implicit input, as presented by Horvitz (you can find a link from the seminar homepage), is thus important here and could give you benefit. Please take some real world examples of prototypes and real situations to your presentation. This makes grasping the idea better and arguing more concrete. You might consider presenting other application ideas as well as the ones already in the paper. I think you would benefit from considering in more detail, for each particular application, why attention and preferences are tracked and how they might be combined, effectively, to minimize disruption and make interaction more fluent. Binding the presentation more tightly to the "let's make interruptions go away" theme of the seminar is important here. Consequently, the presentation, it would be nice to see your analysis of "how things were" and "how things are" (now with AUIs).

Oulasvirta Attention Working memory Long-time memory Task resumptions Control Trust Stress Social interaction