Referring to Objects with Spoken and Haptic Modalities Frédéric LANDRAGIN Nadia BELLALEM & Laurent ROMARY LORIA Laboratory Nancy, FRANCE.

Slides:



Advertisements
Similar presentations
Reference Resolution And Cognitive Grammar Susanne Salmon-Alt Laurent Romary Loria - Nancy, France ICCS-01 San Sebastian, May 2001.
Advertisements

Web Services Architecture An interoperability architecture for the World Wide Service Network.
Ch:8 Design Concepts S.W Design should have following quality attribute: Functionality Usability Reliability Performance Supportability (extensibility,
Dr Yanna Vogiazou, Goldsmiths College M-Learning Symposium, 9th February 2007, WLE Centre Institute of Education Identifying and designing socio-emotional.
RDF Schemata (with apologies to the W3C, the plural is not ‘schemas’) CSCI 7818 – Web Technologies 14 November 2001 Van Lepthien.
An overview of EMMA— Extensible MultiModal Annotation Michael Johnston AT&T Labs Research 8/9/2006.
ARCH-05 Application Prophecy UML 101 Peter Varhol Principal Product Manager.
Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
1 Introduction to XML. XML eXtensible implies that users define tag content Markup implies it is a coded document Language implies it is a metalanguage.
Understand Web Services
Design Activities in Usability Engineering laura leventhal and julie barnes.
John Hu Nov. 9, 2004 Multimodal Interfaces Oviatt, S. Multimodal interfaces Mankoff, J., Hudson, S.E., & Abowd, G.D. Interaction techniques for ambiguity.
Handhelds and Collaborative Command and Control Brad Myers Human Computer Interaction Institute Carnegie Mellon University February 23, 2001.
Information Retrieval Concerned with the: Representation of Storage of Organization of, and Access to Information items.
Marakas: Decision Support Systems, 2nd Edition © 2003, Prentice-Hall Chapter Chapter 1: Introduction to Decision Support Systems Decision Support.
ReQuest (Validating Semantic Searches) Norman Piedade de Noronha 16 th July, 2004.
Visual Information Retrieval Chapter 1 Introduction Alberto Del Bimbo Dipartimento di Sistemi e Informatica Universita di Firenze Firenze, Italy.
Supporting Information Needs by Ostensive Definition in an Adaptive Information Space Iain Campbell 1995 Gretchen Schwarz.
Web Service Architecture Part I- Overview and Models (based on W3C Working Group Note Frank.
The Software Product Life Cycle. Views of the Software Product Life Cycle  Management  Software engineering  Engineering design  Architectural design.
DFKI Approach to Dialogue Management Norbert Reithinger, Elsa Pecourt, Markus Löckelt
MULTIMODAL METALANGUAGE wordpress.com/about-the- glossary/
Domain-Specific Software Engineering Alex Adamec.
Smart Learning Services Based on Smart Cloud Computing
Towards User Interface Derivation from Business Processes: A Model-Driven Approach for Organizational Engineering Kênia Sousa, Hildeberto Mendonça, Jean.
GMD German National Research Center for Information Technology Innovation through Research Jörg M. Haake Applying Collaborative Open Hypermedia.
1 Programming Concepts Module Code : CMV6107 Class Contact Hours: 45 hours (Lecture 15 hours) (Laboratory/Tutorial 30 hours) Module Value: 1 Textbook:
XP Tutorial 7New Perspectives on Creating Web Pages with HTML, XHTML, and XML 1 Working with Cascading Style Sheets Creating a Style for Online Scrapbooks.
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
Spring /6.831 User Interface Design and Implementation1 Lecture 3: Visibility.
Recognition of meeting actions using information obtained from different modalities Natasa Jovanovic TKI University of Twente.
Working group on multimodal meaning representation Dagstuhl workshop, Oct
1 ICAS’2008 – Gosier, March 16-21, 2008 A Transformational Approach for Pattern-based Design of User Interfaces Costin Pribeanu Jean Vanderdonckt National.
Affective Interfaces Present and Future Challenges Introductory statement by Antonio Camurri (Univ of Genoa) Marc Leman (Univ of Gent) MEGA IST Multisensory.
Experiments on Building Language Resources for Multi-Modal Dialogue Systems Goals identification of a methodology for adapting linguistic resources for.
Break-out Group # D Research Issues in Multimodal Interaction.
APML, a Markup Language for Believable Behavior Generation Soft computing Laboratory Yonsei University October 25, 2004.
Multimodal Information Access Using Speech and Gestures Norbert Reithinger
Multi-Modal Dialogue in Human-Robot Conversation Information States in a Multi- modal Dialogue System for Human-Robot Conversation (Lemon et al. 2001)
Copyright 2002 Prentice-Hall, Inc. Modern Systems Analysis and Design Third Edition Jeffrey A. Hoffer Joey F. George Joseph S. Valacich Chapter 20 Object-Oriented.
What is a Business Analyst? A Business Analyst is someone who works as a liaison among stakeholders in order to elicit, analyze, communicate and validate.
Towards multimodal meaning representation Harry Bunt & Laurent Romary LREC Workshop on standards for language resources Las Palmas, May 2002.
Key Centre of Design Computing and Cognition – University of Sydney Concept Formation in a Design Optimization Tool Wei Peng and John S. Gero 7, July,
Vocabularies for Description of Accessibility Issues in MMUI Željko Obrenović, Raphaël Troncy, Lynda Hardman Semantic Media Interfaces, CWI, Amsterdam.
An Ontological Framework for Web Service Processes By Claus Pahl and Ronan Barrett.
NLP ? Natural Language is one of fundamental aspects of human behaviors. One of the final aim of human-computer communication. Provide easy interaction.
A Context Model based on Ontological Languages: a Proposal for Information Visualization School of Informatics Castilla-La Mancha University Ramón Hervás.
A comprehensive framework for multimodal meaning representation Ashwani Kumar Laurent Romary Laboratoire Loria, Vandoeuvre Lès Nancy.
CMPS 435 F08 These slides are designed to accompany Web Engineering: A Practitioner’s Approach (McGraw-Hill 2008) by Roger Pressman and David Lowe, copyright.
PLCS DEXs Trine Hansen DNV 20 April Content OASIS / PLCS Organization PLCS DEXs DEX architecture Process – define and verify capabilities Way forward.
TMF - Terminological Markup Framework Laurent Romary Laboratoire LORIA (CNRS, INRIA, Universités de Nancy) ISO meeting London, 14 August 2000.
ICT EMMSAD’05 13/ Assessing Business Process Modeling Languages Using a Generic Quality Framework Anna Gunhild Nysetvold* John Krogstie *, § IDI,
Working with Ontologies Introduction to DOGMA and related research.
1 1 Spatialized Haptic Rendering: Providing Impact Position Information in 6DOF Haptic Simulations Using Vibrations 9/12/2008 Jean Sreng, Anatole Lécuyer,
Design and Evaluation of an Ambient Display to Support Time Management during Meetings Valentina Occhialini, Harm van Essen, Berry Eggen Intelligent Lighting.
Exam 2 Review Software Engineering CS 561. Outline Requirements Development UML Class Diagrams Design Patterns Users, Usability, and User Interfaces Software.
1 Galatea: Open-Source Software for Developing Anthropomorphic Spoken Dialog Agents S. Kawamoto, et al. October 27, 2004.
MPEG-7 Audio Overview Ichiro Fujinaga MUMT 611 McGill University.
Presentation Title 1 1/27/2016 Lucent Technologies - Proprietary Voice Interface On Wireless Applications Protocol A PDA Implementation Sherif Abdou Qiru.
Multi-Modal Dialogue in Personal Navigation Systems Arthur Chan.
SemAF – Basics: Semantic annotation framework Harry Bunt Tilburg University isa -6 Joint ISO - ACL/SIGSEM workshop Oxford, January 2011 TC 37/SC.
Understanding Naturally Conveyed Explanations of Device Behavior Michael Oltmans and Randall Davis MIT Artificial Intelligence Lab.
Supporting the design of interactive systems a perspective on supporting people’s work Hans de Graaff 27 april 2000.
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
NCP meeting Jan 27-28, 2003, Brussels Colette Maloney Interfaces, Knowledge and Content technologies, Applications & Information Market DG INFSO Multimodal.
Visual Information Retrieval
Web Development A Visual-Spatial Approach
Chapter 20 Object-Oriented Analysis and Design
Multimodal Human-Computer Interaction New Interaction Techniques 22. 1
Jewitt, C. (2014). The Routledge Handbook of Multimodal Analysis
Presentation transcript:

Referring to Objects with Spoken and Haptic Modalities Frédéric LANDRAGIN Nadia BELLALEM & Laurent ROMARY LORIA Laboratory Nancy, FRANCE

Overview Context: Conception phase of the EU IST/MIAMM project (Multidimensional Information Access using Multiple Modalities - with DFKI, TNO, Sony, Canon) –Study of the design factors for a future haptic PDA like device Background: Interpretation of natural language and spontaneous gestures. –A model of contextual interpretation of multimodal referring expressions in visual contexts. Objective: To show the possible extension of the model to an interaction mode including tactile and kinesthetic feedback (henceforth “Haptic”)

MIAMM - wheel mode

MIAMM - axis mode

MIAMM - galaxy mode

The use of perceptual grouping Reference domains and visual contexts « these three objects »  {,, }

The use of perceptual grouping Reference domains and visual contexts « these three objects »  {,, } « the triangle »  { } « the two circles »  {, } The use of salience

Multimodal fusion architecture reference domain(s) history task module visual module dialogue manager request domains request domains request domains referential expression referent history language module under- specification user model referential gesture

Haptics and deixis Haptic gestures can take three classical functions of gesture in man-machine interaction: –semiotic function: ‘select this object’ –ergotic function: ‘reduce the size of this object’ –epistemic function: ‘save the compliance of this object’ How can the system identify the function(s)? –linguistic clues (referential expression, predicate) –task indications (possibilities linked to a type of objects, “affordances”) Deictic role: to make the object salient, whatever the function, in order to focus the addressee’s attention on it.

Haptics and perceptual grouping Interest: formalism for the focalization on a subset of objects Grouping factors: –objects which have similar haptic (in particular, tactile) properties (shape, consistency, texture) –objects that have been browsed by the user (the elements of such a group are ordered) –objects that are stuck together, parts of a same object...

Haptics and perceptual domains Can visual and tactile perceptions work together? –simultaneous visual and tactile perception implies the same world of objects (and synchronized feedbacks) –a referring expression can be interpreted in visual context or in tactile context How can the system identify the nature of perception? –for immediate references, the visual context gives the reference domain and haptics gives the starting point in it –for ieterating references, each type of context can provide the reference domain (so both hypotheses must be tested)

Haptics and dialogue history Interpretations that require an ordering within the reference domain: ‘the first one’, ‘the next one’, ‘the last one’ –in visual perception, guiding lines can be helpful (if none, an order can always be built with the reading direction) –in haptic perception, the only criterion can be the manipulation order Some referring expressions that do not need an order may be interpreted in the haptic manipulation history –‘the big one’ (in the domain of browsed objects) –‘them’ (the most pressured objects)

Architecture for speech-haptic referring reference domain(s) history task module visuo- tactile module dialogue manager request domains request domains request domains referent history language module under- specification user model haptic gesture referential expression

Summary What does not change from deictic to haptic: –the status of speech and gesture in the architecture –the repartition of information among speech and gesture –the need of reference domains –the use of salience and the use of orderings in domains –the algorithms for the exploitation of all these notions What does change: –some unchanged notions can have more than one cause –objects must be browsed to be grouped in a haptic domain –one aspect of the architecture: the visual perception module becomes the visuo-tactile perception module

Application - MMIL design MMIL - Multimodal Interface Language –Unique communication language within the MIAMM architecture –XML based representation embedded in a web service protocol (SOAP) –Oriented towards multimodal content representation (cf. Joint ACL/SIGSEM and TC37/SC4 initiative) Example: VisHapTac notification to Dialogue Manager

Overall structure e0e0 set 1 s1s1 s2s2 set 2 s 25 … description s 2-1 s 2-2 s 2-3 inFocus inSelection Visual haptic state Participant setting Sub-divisions

MMIL format for VisHapTac - 1 HGState galaxy <tempSpan startPoint=“ T14:12:06” endPoint=“ T14:12:13”/> …

MMIL format for VisHapTac - 2 … Let it be set inFocus Lady Madonna … inSelection Revolution 9 …

Future work Within the dialogue manager module, domains may be confronted, using a relevance criterion The way the linguistic contraints of the referring expression apply in the different domains may be such a criterion. Validation in the MIAMM architecture The transition from deictic to haptic may not be an additional cost for the development of a dialogue system, both from the architecture point of view and the dialogue management point of view.