Multimodal user interfaces: Implementation Chris Vandervelpen

Slides:



Advertisements
Similar presentations
1 XML Technologies You Need For Workflow Applications Roger L. Costello 2 January, 2009.
Advertisements

INTEGRATION OF VOICE SERVICES IN INTERNET APPLICATIONS By Eduardo Carrillo (lecturer), J. J Samper, J.J. Martínez-Durá Universidad Autónoma de Bucaramanga.
VRIXML: A User Interface Description for Virtual Environments Erwin Cuppens, Chris Raymaekers, Karin Coninx Expertise Centre for Digital Media (EDM) Limburgs.
M. Honkala, A. Vainio XForms & SVG MITA Seminar, HUT SVG Introduction Features Benefits Simple Example Standardization Mobile Environment Current.
H E L S I N K I U N I V E R S I T Y O F T E C H N O L O G Y G O p r o j e c t : S e r v i c e A r c h i t e c t u r e f o r t h e N o m a d i c I n t e.
Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
Empirical and Data-Driven Models of Multimodality Advanced Methods for Multimodal Communication Computational Models of Multimodality Adequate.
Which development tool is right for you? Commercial Tools John Fuentes – Principal Solutions Architect
XISL language XISL= eXtensible Interaction Sheet Language or XISL=eXtensible Interaction Scenario Language.
JAXB Java Architecture for XML Binding Andy Fanton Khalid AlBayat.
Visual Model-based Design of Multi- platform Interactive Applications HCI Group/ISTI-CNR 23 September 2002 Pisa (I)
The State of the Art in VoiceXML Chetan Sharma, MS Graduate Student School of CSIS, Pace University.
Pace VoiceXML Absentee System Paul Visokey, Ping Gallivan, Yani Mulyani, Lisa Jordan, Elaine Li, George Mathew, Qisheng Hong Presenter Name : Paul Visokey.
Introduction to SVG & Batik Presented by Shang-Ming Huang.
XML Prashant Karmarkar Brendan Nolan Alexander Roda.
Supervised by Prof. LYU, Rung Tsong Michael Department of Computer Science & Engineering The Chinese University of Hong Kong Prepared by: Chan Pik Wah,
Presentation Outline  Project Aims  Introduction of Digital Video Library  Introduction of Our Work  Considerations and Approach  Design and Implementation.
LemGen (Linguistic EMulation and Generation ENgine) CS491 Project Chris Lemcke.
Multimodal Architecture for Integrating Voice and Ink XML Formats Under the guidance of Dr. Charles Tappert By Darshan Desai, Shobhana Misra, Yani Mulyani,
Device-Independence with UIML (User Interface Markup Language) Marc Abrams uiml.org W3C Workshop on Web Device Independent Authoring.
VoiceXML Builder Arturo Ramirez ACS 494 Master’s Graduate Project May 04, 2001.
XForms: A case study Rajiv Shivane & Pavitar Singh.
CPS120: Introduction to Computer Science The World Wide Web Nell Dale John Lewis.
Helsinki University Of Technology X-Smiles Telecommunications Software and Multimedia Laboratory (TML) XForms – the Future of XML Input
Some Thoughts on HPC in Natural Language Engineering Steven Bird University of Melbourne & University of Pennsylvania.
1 The Architectural Design of FRUIT: A Family of Retargetable User Interface Tools Yi Liu, H. Conrad Cunningham and Hui Xiong Computer & Information Science.
Conversational Applications Workshop Introduction Jim Larson.
Helsinki University Of Technology X-Smiles Telecommunications Software and Multimedia Laboratory (TML) XForms in X-Smiles Mikko Honkala Helsinki University.
July 13, 2006 © 2006 IBM Corporation Distributed Multimodal Synchronization Protocol (DMSP) Chris Cross IETF 66 July 13, 2006 With Contribution from Gerald.
Jozef Stefan Institute Program Generators and Control System Software Development Klemen Žagar Anže.
Mihir Daptardar Software Engineering 577b Center for Systems and Software Engineering (CSSE) Viterbi School of Engineering 1.
ITCS 6010 SALT. Speech Application Language Tags (SALT) Speech interface markup language Extension of HTML and other markup languages Adds speech and.
Funded by: European Commission – 6th Framework Project Reference: IST WP 2: Learning Web-service Domain Ontologies Miha Grčar Jožef Stefan.
Robbie Schäfer (Paderborn University) UIML TC Discussion on DISL.
Spoken dialog for e-learning supported by domain ontologies Dario Bianchi, Monica Mordonini and Agostino Poggi Dipartimento di Ingegneria dell’Informazione.
1Copyright © PIPEBEACH AB All rights reserved. Scott McGlashan zW3C Voice Browser Dialog Requirements and Specifications zTranscoding WML into VoiceXML.
CUIML:A language for generating multimodal human-computer interface Christian Sandor Thomas Reicher 報告人 try.
The Voice-Enabled Web: VoiceXML and Related Standards for Telephone Access to Web Applications 14 Feb Christophe Strobbe K.U.Leuven - ESAT-SCD-DocArch.
Outline Grammar-based speech recognition Statistical language model-based recognition Speech Synthesis Dialog Management Natural Language Processing ©
Spoken Dialog Systems and Voice XML Lecturer: Prof. Esther Levin.
Voice User Interface
Design Options for Multimodal Web Applications Adrian Stanciulescu and Jean Vanderdonckt {stanciulescu, UCL/IAG/BCHI.
IBM - CVUT Student Research Projects Google maps with voice Martin Absolon Ivo Čermák
March 20, 2006 © 2005 IBM Corporation Distributed Multimodal Synchronization Protocol (DMSP) Chris Cross IETF 65 March 20, 2006 With Contribution from.
Microsoft Assistive Technology Products Brought to you by... Jill Hartman.
1 Representing New Voice Services and Their Features Ken Turner University of Stirling 11th June 2003.
A comprehensive framework for multimodal meaning representation Ashwani Kumar Laurent Romary Laboratoire Loria, Vandoeuvre Lès Nancy.
Mir Farooq Ali Computer Science, Virginia Tech May 9, 2003 Building Multi-platform User Interfaces using UIML.
Generative Approaches for Application Tailoring of Mobile Devices Victoria M. Davis, Dr. Jeff Gray (UAB) and Dr. Joel Jones (UA) Portions of this research.
SkyNET Visualization Team Demo and Architecture Overview.
Justin Tilton uPortal Framework and User Interface Demonstration.
Strategies for Virtual Enterprises using XForms and the Semantic Web Albert Rainer, Jürgen Dorn & Peter Hrastnik.
Device-Independence with UIML (User Interface Markup Language) Marc Abrams uiml.org W3C Workshop on Web Device Independent Authoring.
A Transformational Approach for Multimodal Web User Interfaces based on UsiXML Adrian Stanciulescu, Quentin Limbourg, Jean Vanderdonckt, Benjamin Michotte.
XFormsDB An XForms - Based Framework for Simplifying Web Application Development Markku Laine Master’s Thesis Presentation March 9, 2010.
Cooperative Computing & Communication Laboratory A Survey on Transformation Tools for Model-Based User Interface Development Robbie Schäfer – Paderborn.
March 20, 2006 © 2005 IBM Corporation Distributed Multimodal Synchronization Protocol (DMSP) Chris Cross IETF 65 March 21, 2006 With Contribution from.
1 Galatea: Open-Source Software for Developing Anthropomorphic Spoken Dialog Agents S. Kawamoto, et al. October 27, 2004.
Presentation Title 1 1/27/2016 Lucent Technologies - Proprietary Voice Interface On Wireless Applications Protocol A PDA Implementation Sherif Abdou Qiru.
CARE properties Chris Vandervelpen
SEESCOASEESCOA SEESCOA Meeting Activities of LUC 9 May 2003.
James A. Larson Developing & Delivering Multimodal Applications 1 EMMA Extensible MultiModal Annotation markup language Canonical structure for semantic.
 XML derives its strength from a variety of supporting technologies.  Structure and data types: When using XML to exchange data among clients, partners,
W3C Multimodal Interaction Activities Deborah A. Dahl August 9, 2006.
Software Architecture for Multimodal Interactive Systems : Voice-enabled Graphical Notebook.
Software Group 7-December-2005 | Cross © 2005 IBM Corporation Distributed Multimodal Synchronization Protocol (DMSP) Chris Cross, Multimodal Browser Architect.
Metadata Editing by Schema Hussein Suleman University of Cape Town August 2003.
Presented By Sharmin Sirajudeen S7 CS Reg No :
(No need of Desktop computer)
Dr Tappert Shreenath Laxman and Darshan Desai
Presentation transcript:

Multimodal user interfaces: Implementation Chris Vandervelpen

Overview Introduction VoiceXml X+V From models to X + V Demo: ACCESS Netfront Conclusions Questions

Introduction Focus on speech/direct manipulation on mobile device How can we deploy a multi modal UI –Build our own framework using speech synthesizer/recognizers that interpret the designed models (reinventing the wheel) –Build software that generates standardized markup from the models (use existing technologies)  start point

VoiceXml Markup language for speech only interfaces Telephone interfaces Using grammars for speech recognition –Java Speech Grammar Format (JSGF) –Nuance Grammar Specification Language (NGSL) Speech output –Synthesis –Prerecorded audio

VoiceXml <![CDATA[ #JSGF V1.0; grammar cities; = brussels | antwerp | amsterdam; ]]> What departure city do you like?? For example, brussels, antwerp or amsterdam Your departure city is ………

VoiceXml Mixed-initiative forms –Single user input for several fields –Supports more natural language For example –I want to fly from “brussels” to “amsterdam” –Filling in departure_city and destination_city fields

X + V –XHtml: visual channel –VoiceXml snippets: speech channel Synchronization between modalities using Xml Events Multimodal browsers supporting X+V –ACCESS Netfront multimodal browser (PocketPC) –Opera al/x+v/12/ al/x+v/12/

X + V <input id=“to” name=“to” size=“20” ev:event=“inputfocus” ev:handler=“#voice_city_to” />

X + V <![CDATA[ #JSGF V1.0; grammar cities; = brussels | antwerp | amsterdam; ]]> What departure city do you like?? For example, brussels, antwerp or amsterdam …….

X + V Also usable with XForms VoiceXml snippets and XForms influence same XForms instance model  synchronization

Models to X + V

Annotate UI description for speech [Shao2003: Transcoding HTML to VoiceXML Using Annotations] Extend this approach to UIML and X + V –Identify particular information structures Text areas Menu/List structures Top-level visual region –Define their representation in XHTML and VoiceXml –Generate the synchronization XML eventing code

Model to X + V Define a generic UIML widget vocabulary mapping for both GUI and speech [Plomp2002] TextEntry – (VoiceXml) – (XHtml) –System.Windows.Forms.TextBox Collection – (VoiceXml) – (XHtml) –System.Windows.Forms.Panel

Access Netfront multimodal browser PocketPC Ordering pizza Ordering Chinese Demo

Conclusions X + V –built-in modality synchronization –alternative to own multimodal implementation –declarative –transformation from UIML possible

Questions?