Human-centered Interfaces

Slides:



Advertisements
Similar presentations
Team:. Prepared By: Menna Hamza Mohamed Mohamed Hesham Fadl Mona Abdel Mageed El-Koussy Yasmine Shaker Abdel Hameed Supervised By: Dr. Magda Fayek.
Advertisements

Perceptive Context for Pervasive Computing Trevor Darrell Vision Interface Group MIT AI Lab.
Hand Gesture for Taking Self Portrait Shaowei Chu and Jiro Tanaka University of Tsukuba Japan 12th July 15 minutes talk.
Learning and Vision for Multimodal Conversational Interfaces Trevor Darrell Vision Interface Group MIT CSAIL Lab.
Joint Eye Tracking and Head Pose Estimation for Gaze Estimation
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Sensor-based Situated, Individualized, and Personalized Interaction in Smart Environments Simone Hämmerle, Matthias Wimmer, Bernd Radig, Michael Beetz.
Paper by Shi, Qin, Suo, & Xiao Presented by Alan Kelly November 16, 2011.
Computer Vision REU Week 2 Adam Kavanaugh. Video Canny Put canny into a loop in order to process multiple frames of a video sequence Put canny into a.
Supporting Collaboration: Digital Desktops to Intelligent Rooms Mary Lou Maher Design Computing and Cognition Group Faculty of Architecture University.
Recent Developments in Human Motion Analysis
Class 6 LBSC 690 Information Technology Human Computer Interaction and Usability.
Head Tracking and Virtual Reality by Benjamin Nielsen.
1 Discussion Class 10 Informedia. 2 Discussion Classes Format: Question Ask a member of the class to answer. Provide opportunity for others to comment.
CS335 Principles of Multimedia Systems Multimedia and Human Computer Interfaces Hao Jiang Computer Science Department Boston College Nov. 20, 2007.
Stockman MSU Fall Computing Motion from Images Chapter 9 of S&S plus otherwork.
Trip Report for The IASTED International Conference on Internet and Multimedia Systems and Applications (EuroIMSA 2006) February 13-15, 2006 Innsbruck,
Computer Vision Systems for the Blind and Visually Disabled. STATS 19 SEM Talk 3. Alan Yuille. UCLA. Dept. Statistics and Psychology.
MACHINE VISION GROUP Multimodal sensing-based camera applications Miguel Bordallo 1, Jari Hannuksela 1, Olli Silvén 1 and Markku Vehviläinen 2 1 University.
On the Design, Construction and Operation of a Diffraction Rangefinder MS Thesis Presentation Gino Lopes A Thesis submitted to the Graduate Faculty of.
Capture, Create, and Share Digital Movies Basic Features of Video.
A Brief Overview of Computer Vision Jinxiang Chai.
Where, Who and Intelligent Affective Interaction ICANN, Sept. 14, Athens, Greece Aristodemos Pnevmatikakis, John Soldatos and Fotios Talantzis.
Kinect Part II Anna Loparev.
Guest Audio Guide. At the top of the virtual meeting room, you will see the Meeting Tab, click on “Manage My Settings”, select “My Connection Speed” and.
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
Prepared By: Menna Hamza Mohamed Mohamed Hesham Fadl Mona Abdel Mageed El-Koussy Yasmine Shaker Abdel Hameed Supervised By: Dr. Magda Fayek.
NM – LREC 2008 /1 N. Moreau 1, D. Mostefa 1, R. Stiefelhagen 2, S. Burger 3, K. Choukri 1 1 ELDA, 2 UKA-ISL, 3 CMU s:
An Information Fusion Approach for Multiview Feature Tracking Esra Ataer-Cansizoglu and Margrit Betke ) Image and.
Computer Science Department Pacific University Artificial Intelligence -- Computer Vision.
Adobe Connect Audio Setup and Best Practices Tutorial
GENESIS OF VIRTUAL REALITY  The term ‘Virtual reality’ (VR) was initially coined by Jaron Lanier, founder of VPL Research (1989)..
Perceptive Context for Pervasive Computing Trevor Darrell Vision Interface Group MIT AI Lab.
卓越發展延續計畫分項三 User-Centric Interactive Media ~ 主 持 人 : 傅立成 共同主持人 : 李琳山,歐陽明,洪一平, 陳祝嵩 水美溫泉會館研討會
Model of the Human  Name Stan  Emotion Happy  Command Watch me  Face Location (x,y,z) = (122, 34, 205)  Hand Locations (x,y,z) = (85, -10, 175) (x,y,z)
CAMEO: Year 1 Progress and Year 2 Goals Manuela Veloso, Takeo Kanade, Fernando de la Torre, Paul Rybski, Brett Browning, Raju Patil, Carlos Vallespi, Betsy.
Mixed Reality: A Model of Mixed Interaction Céline Coutrix, Laurence Nigay User Interface Engineering Team CLIPS-IMAG Laboratory, University of Grenoble.
Student: Ibraheem Frieslaar Supervisor: Mehrdad Ghaziasgar.
MIT Artificial Intelligence Laboratory — Research Directions The Next Generation of Robots? Rodney Brooks.
Detecting Eye Contact Using Wearable Eye-Tracking Glasses.
IEEE International Conference on Multimedia and Expo.
MIT Artificial Intelligence Laboratory — Research Directions Intelligent Perceptual Interfaces Trevor Darrell Eric Grimson.
Student Gesture Recognition System in Classroom 2.0 Chiung-Yao Fang, Min-Han Kuo, Greg-C Lee, and Sei-Wang Chen Department of Computer Science and Information.
Human-Robot Interaction
Perceptive Computing Democracy Communism Architecture The Steam Engine WheelFire Zero Domestication Iron Ships Electricity The Vacuum tube E=mc 2 The.
Lesson 4 Alternative Methods Of Input.
EYE TRACKING TECHNOLOGY
Hand Gestures Based Applications
Alternative Methods Of Input
Standard Methods of Input.
Introducing virtual REALITY
Computer vision: models, learning and inference
VIRTUAL INTELLIGENCE PROJECT NATAL (Kinect & Xbox 360)
Methods of Computer Input and Output
Jun Shimamura, Naokazu Yokoya, Haruo Takemura and Kazumasa Yamazawa
CALO VISUAL INTERFACE RESEARCH PROGRESS
Ubiquitous Computing and Augmented Realities
Rogerio Feris 1, Ramesh Raskar 2, Matthew Turk 1
Lesson 4 Alternative Methods Of Input.
Common Classification Tasks
--- Stereoscopic Vision and Range Finders
--- Stereoscopic Vision and Range Finders
Wearable Visual Information Systems (VINST)
Interactive Background Blurring
Capture, Create, and Share Digital Movies
Lesson 4 Alternative Methods Of Input.
Liyuan Li, Jerry Kah Eng Hoe, Xinguo Yu, Li Dong, and Xinqi Chu
LEAP MOTION: GESTURAL BASED 3D INTERACTIONS
CSE (c) S. Tanimoto, 2001 Image Understanding
Computer Vision Readings
Presentation transcript:

Perceptive Context for Pervasive Computing Trevor Darrell Vision Interface Group

Human-centered Interfaces Free users from desktop and wired interfaces Allow natural gesture and speech commands Give computers awareness of users Work in open and noisy environments Outdoors -- PDA next to construction site! Indoors -- crowded meeting room Vision’s role: provide perceptive context

Perceptive Context Who is there? (presence, identity) What is going on? (activity) Where are they? (individual location) Which person said that? (audiovisual grouping) What are they looking / pointing at? (pose, gaze)

Virtual views Profile silhouette: Frontal Head:

Examples: VH-generated views Faces: Gait:

Range-based stereo person tracking Range can be insensitive to fast illumination change Compare range values to known background Project into 2D overhead view Plan view Foreground Range Intensity Merge data from multiple stereo cameras.. Group into trajectories… Examine height for sitting/standing…

Audio input in noisy environments Acquire high-quality audio from untethered, moving speakers “Virtual” headset microphones for all users

Vision guided microphone array Cameras Microphones

Audio-visual synchrony detection

AVMI Applications + Audio weighting from video (detected face) Image localization from audio + image variance AVMI Audio weighting from video (detected face) Audio associated with left face Audio associated with right face

Brightness and depth motion constraints Parameter space I t I t + 1 I Z Z t Z t + 1 yt = yt-1

Track users head gaze for hands-free pointing… Closed-loop 3D tracker Track users head gaze for hands-free pointing…

Head-driven cursor Related Projects: Schiele Kjeldsen Toyama Current application for second pointer or scrolling / focus of attention…

Subject not looking at SAM “Look-to-talk” Subject not looking at SAM ASR turned off Subject looking at SAM ASR turned on