Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.

Slides:



Advertisements
Similar presentations
Some Reflections on Augmented Cognition Eric Horvitz ISAT & Microsoft Research November 2000 Some Reflections on Augmented Cognition Eric Horvitz ISAT.
Advertisements

National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory
An overview of EMMA— Extensible MultiModal Annotation Michael Johnston AT&T Labs Research 8/9/2006.
Frontiers in Interaction The Power of Multimodal Standards Deborah Dahl Principal, Conversational Technologies Chair, W3C Multimodal Interaction Working.
Sensor-based Situated, Individualized, and Personalized Interaction in Smart Environments Simone Hämmerle, Matthias Wimmer, Bernd Radig, Michael Beetz.
Chapter 11 Beyond Bag of Words. Question Answering n Providing answers instead of ranked lists of documents n Older QA systems generated answers n Current.
ICS 463, Intro to Human Computer Interaction Design: 9 “Theory”. Input and Output Dan Suthers.
Supervised by Prof. LYU, Rung Tsong Michael Department of Computer Science & Engineering The Chinese University of Hong Kong Prepared by: Chan Pik Wah,
MULTIMEDIA SYSTEMS IREK DEFEE MULTIMEDIA SYSTEMS and SERVICES SGN-5116 Irek Defée TF316,
Stanford hci group / cs376 research topics in human-computer interaction Multimodal Interfaces Scott Klemmer 15 November 2005.
Presentation Outline  Project Aims  Introduction of Digital Video Library  Introduction of Our Work  Considerations and Approach  Design and Implementation.
ISTD 2003, Audio / Speech Interactive Systems Technical Design Seminar work: Audio / Speech Ville-Mikko Rautio Timo Salminen Vesa Hyvönen.
MULTIMEDIA SIGNAL PROCESSING MMSP SGN-5016 Irek Defée Tietotalo TF 316
REQUIREMENTS FOR MULTIMEDIA PROCESSING MULTIMEDIA SYSTEMS IREK DEFEE.
Visual Information Retrieval Chapter 1 Introduction Alberto Del Bimbo Dipartimento di Sistemi e Informatica Universita di Firenze Firenze, Italy.
1. Plan : 1. Models of interaction 2. Types of interaction 3. Existing technologies 4. Advances in HCI 5. Architecture 2.
Support for Palm Pilot Collaboration Including Handwriting Recognition.
CHAPTER 2 Input & Output Prepared by: Mrs.sara salih 1.
1 / 23 Microsoft Tablet PC Technology Thomas Dodds Declan O’Gorman David Pickles Stephen Pond An overview of Microsoft Tablet PC technology and current.
Mobile Multimodal Applications. Dr. Roman Englert, Gregor Glass March 23 rd, 2006.
Chapter II The Multimedia Sysyem. What is multimedia? Multimedia means that computer information can be represented through audio, video, and animation.
ASSISTIVE TECHNOLOGY PRESENTED BY ABDUL BARI KP. CONTENTS WHAT IS ASSISTIVE TECHNOLOGY? OUT PUT: Screen magnifier Speech to Recogonizing system Text to.
1 Skip Cave Chief Scientist, Intervoice Inc. Multimodal Framework Proposal.
Enabling enactive interaction in virtualized experiences Stefano Tubaro and Augusto Sarti DEI – Politecnico di Milano, Italy.
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
Conversational Applications Workshop Introduction Jim Larson.
Multimedia Databases (MMDB)
Multimodal user interfaces: Implementation Chris Vandervelpen
ITCS 6010 SALT. Speech Application Language Tags (SALT) Speech interface markup language Extension of HTML and other markup languages Adds speech and.
ST01 - Introduction 1 Introduction Lecturer: Smilen Dimitrov Sensors Technology – MED4.
Fall UI Design and Implementation1 Lecture 20: HCI Research Topics.
Chapter 5 Input and Output: The User Connection. Input n input device - hardware device that translates raw data into a format understandable by the computer.
COMPUTER PARTS AND COMPONENTS INPUT DEVICES
Outline Grammar-based speech recognition Statistical language model-based recognition Speech Synthesis Dialog Management Natural Language Processing ©
Spoken Dialog Systems and Voice XML Lecturer: Prof. Esther Levin.
Voice User Interface
GENESIS OF VIRTUAL REALITY  The term ‘Virtual reality’ (VR) was initially coined by Jaron Lanier, founder of VPL Research (1989)..
COMPUTER MAIN PART DANILO GARCÍA GONZÁLEZ TÉCNICO EN SISTEMAS.
Human Computer Interaction © 2014 Project Lead The Way, Inc.Computer Science and Software Engineering.
Human-Computer Interaction
Hapgets, Towards Haptically- enhanced Widgets based on a User Interface Description Language Nikolaos Kaklanis, Juan González Calleros, Jean Vanderdonckt,
卓越發展延續計畫分項三 User-Centric Interactive Media ~ 主 持 人 : 傅立成 共同主持人 : 李琳山,歐陽明,洪一平, 陳祝嵩 水美溫泉會館研討會
March 20, 2006 © 2005 IBM Corporation Distributed Multimodal Synchronization Protocol (DMSP) Chris Cross IETF 65 March 20, 2006 With Contribution from.
Microsoft Assistive Technology Products Brought to you by... Jill Hartman.
Toward a Unified Scripting Language 1 Toward a Unified Scripting Language : Lessons Learned from Developing CML and AML Soft computing Laboratory Yonsei.
Win OS & Hardware. Input Getting data into the computer.
March 31, 1998NSF IDM 98, Group F1 Group F Multi-modal Issues, Systems and Applications.
March 20, 2006 © 2005 IBM Corporation Distributed Multimodal Synchronization Protocol (DMSP) Chris Cross IETF 65 March 21, 2006 With Contribution from.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Human Factors in Mobile Computing By: Ed Leland EEL
Intelligent MultiMedia Storytelling System (IMSS) - Automatic Generation of Animation From Natural Language Input By Eunice Ma Supervisor: Prof. Paul Mc.
Preparing for the 2008 Beijing Olympics : The LingTour and KNOWLISTICS projects. MAO Yuhang, DING Xiao-Qing, NI Yang, LIN Shiuan-Sung, Laurence LIKFORMAN,
1 Interaction Devices CIS 375 Bruce R. Maxim UM-Dearborn.
James A. Larson Developing & Delivering Multimodal Applications 1 EMMA Extensible MultiModal Annotation markup language Canonical structure for semantic.
CS1354 GRAPHICS AND MULTIMEDIA Handle By, S.JENILA AP/IT.
W3C Multimodal Interaction Activities Deborah A. Dahl August 9, 2006.
What is Multimedia Anyway? David Millard and Paul Lewis.
Iterative Design and Evaluation of Event Architecture for Pen-and-Paper Interfaces HCI Group Stanford University Ron B. Yeh Andreas Paepcke Scott R. Klemmer.
MULTIMODAL AND NATURAL COMPUTER INTERACTION Domas Jonaitis.
SIE 515 Universal Design Lecture 9.
Chp 4: Input and Output Devices
Presentation of Input & Output Devices
Visual Information Retrieval
Group “A” Researchers 7/20/2005
Multimodal Interfaces
What is Pattern Recognition?
Pervasive Computing Happening?
Speech & Multimodal Scott Klemmer · 16 November 2006.
Android Sensor Programming
Multimedia Systems & Interfaces
Presentation transcript:

Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day

Irek Defée Current status Web is developed for traditional data and computer I/O: text, keyboard, mouse This is simple and effective but not a natural way of human interaction with the world Humans interact via perceptual system

Irek Defée Human Perceptual System Human perceptual system has multiple senses: visual, acoustical, haptic (touch, body position, temperature) and actuators (vocal tract, muscles, motoric system) The perceptual system is intrinsically MULTIMODAL: multiple senses and actuators operate in perfectly coordinated way

Irek Defée Perceptual Information Technology Information technology is evolving towards natural MULTIMODAL human interaction:  Touch gestures revolutionized mobile devices  Intelligent speech input is available  There is more to come: new sensors, cameras and intelligence

Irek Defée Signal Processing Role Perceptual Information Technology requires sophisticated signal processing and it is hard due to: - Complex input signals - Complex information encoding - Complex databases of knowledge Highly sophisticated algorithms and huge processing power are required

Irek Defée Multimodal Web The trend towards perceptual information is noted at the W3C: Extending the Web to allow multiple modes of interaction: GUI, Speech, Vision, Pen, Gestures, Haptic interfaces,... Multimodal Interaction Activity: - Multimodal Architecture and Interfaces - EMMA - InkML - EmotionML

Irek Defée Multimodal Architecture

Irek Defée EMMA Extensible Multimodal Markup Language for Annotations - containing and annotating the interpretation of user input - transcription into words of a raw signal, for instance derived from speech, pen - interpretation is to be generated by signal interpretation processes, such as speech and ink recognition, semantic interpreters

Irek Defée Ink Markup Language data format for representing ink input and processing of handwriting, gestures, sketches, music using traces of pen Trace attributes

Irek Defée Emotion Markup Language Annotation of material involving emotionality Automatic recognition of emotions from sensors Generation of emotion-related system responses: speech, music, colors, gestures, synthetic faces Emotion vocabularies and representations: <emotion category- set=" voc/xml#big6"> <category name="surprise" confidence="0.9

Irek Defée Department of Signal Processing Signal processing has a key role as a front-end for the Multimodal Web Department is on the forefront of research in the natural information processing: - Multimedia information analysis, retrieval and databases - Audio information analysis : speech and music - Media information handling: representation and compression

Irek Defée for your attention!