ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents.

Slides:



Advertisements
Similar presentations
National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory
Advertisements

MDI 2010, Oslo, Norway Behavioural Interoperability to Support Model-Driven Systems Integration Alek Radjenovic, Richard Paige The University of York,
Expressive Gestures for NAO NAO TechDay, 13/06/2012, Paris Le Quoc Anh - Catherine Pelachaud CNRS, LTCI, Telecom-ParisTech, France.
Enabling Access to Sound Archives through Integration, Enrichment and Retrieval WP1. Project Management.
Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
University of Minho School of Engineering Centre ALGORITMI Uma Escola a Reinventar o Futuro – Semana da Escola de Engenharia - 24 a 27 de Outubro de 2011.
Empirical and Data-Driven Models of Multimodality Advanced Methods for Multimodal Communication Computational Models of Multimodality Adequate.
Persuasive Listener in a Conversation Elisabetta Bevacqua, Chris Peters, Catherine Pelachaud (IUT de Montreuil - Paris 8)
ENTERFACE’08 Multimodal high-level data integration Project 2 1.
23-May-151 Multiparty Communication with a Tour Guide ECA Aleksandra Čereković HOTLab group Department of telecommunications Faculty of electrical engineering.
I1-[OntoSpace] Ontologies for Spatial Communication John Bateman, Kerstin Fischer, Reinhard Moratz Scott Farrar, Thora Tenbrink.
Media Coordination in SmartKom Norbert Reithinger Dagstuhl Seminar “Coordination and Fusion in Multimodal Interaction” Deutsches Forschungszentrum für.
NEW TIES year 2 review NEW TIES = New and Emergent World models Through Individual, Evolutionary and Social learning.
Supervised by Prof. LYU, Rung Tsong Michael Department of Computer Science & Engineering The Chinese University of Hong Kong Prepared by: Chan Pik Wah,
Program Flow Charting How to tackle the beginning stage a program design.
1 IUT de Montreuil Université Paris 8 Emotion in Interaction: Embodied Conversational Agents Catherine Pelachaud.
Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Cognitive Robots © 2014, SNU CSE Biointelligence Lab.,
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
Conversational Applications Workshop Introduction Jim Larson.
Recognition of meeting actions using information obtained from different modalities Natasa Jovanovic TKI University of Twente.
Markup of Multimodal Emotion-Sensitive Corpora Berardina Nadja de Carolis, Univ. Bari Marc Schröder, DFKI.
ITCS 6010 SALT. Speech Application Language Tags (SALT) Speech interface markup language Extension of HTML and other markup languages Adds speech and.
Parser-Driven Games Tool programming © Allan C. Milne Abertay University v
MediaHub: An Intelligent MultiMedia Distributed Platform Hub Glenn Campbell, Tom Lunney & Paul Mc Kevitt School of Computing and Intelligent Systems Faculty.
Chapter 7. BEAT: the Behavior Expression Animation Toolkit
APML, a Markup Language for Believable Behavior Generation Soft computing Laboratory Yonsei University October 25, 2004.
OBJECT ORIENTED SYSTEM ANALYSIS AND DESIGN. COURSE OUTLINE The world of the Information Systems Analyst Approaches to System Development The Analyst as.
Outline Grammar-based speech recognition Statistical language model-based recognition Speech Synthesis Dialog Management Natural Language Processing ©
By Ian Jackman Davit Stepanyan.  User executed untested code.  The order in which statements were meant to be executed are different than the order.
Oct 14, 2001OOPSLA’01- DSVL1 Experiences with Visual Programming Languages for End-Users and Specific Domains Philip T. Cox Trevor J. Smedley Dalhousie.
Spoken Dialog Systems and Voice XML Lecturer: Prof. Esther Levin.
1 ISA&D7‏/8‏/ ISA&D7‏/8‏/2013 Methodologies of the SDLC Traditional Approach to SDLC Object-Oriented Approach to SDLC CASE Tools.
1 Introduction to Software Engineering Lecture 1.
Lecture 15 – Social ‘Robots’. Lecture outline This week Selecting interfaces for robots. Personal robotics Chatbots AIML.
Backchannels Through Gaze as Indicators of Persuasive Success E. Bevacqua, M. Mancini, C. Peters, C. Pelachaud University of Paris 8 Isabella Poggi Università.
卓越發展延續計畫分項三 User-Centric Interactive Media ~ 主 持 人 : 傅立成 共同主持人 : 李琳山,歐陽明,洪一平, 陳祝嵩 水美溫泉會館研討會
Scenarios for a Learning GRID Online Educa Nov 30 – Dec 2, 2005, Berlin, Germany Nicola Capuano, Agathe Merceron, PierLuigi Ritrovato
Project funded by the Future and Emerging Technologies arm of the IST Programme FET-Open scheme Project funded by the Future and Emerging Technologies.
ENTERFACE 08 Project 1 “MultiParty Communication with a Tour Guide ECA” Mid-term presentation August 19th, 2008.
DARPA ITO/MARS Project Update Vanderbilt University A Software Architecture and Tools for Autonomous Robots that Learn on Mission K. Kawamura, M. Wilkes,
Fall 2004EE 3563 Digital Systems Design EE 3563 VHSIC Hardware Description Language  Required Reading: –These Slides –VHDL Tutorial  Very High Speed.
Using RouteGraphs as an Appropriate Data Structure for Navigational Tasks SFB/IQN-Kolloquium Christian Mandel, A1-[RoboMap] Overview Goal scenario.
ECE450 - Software Engineering II1 ECE450 – Software Engineering II Today: Introduction to Software Architecture.
1 Workshop « Multimodal Corpora » Jean-Claude MARTIN Patrizia PAGGIO Peter KÜEHNLEIN Rainer STIEFELHAGEN Fabio PIANESI.
Feedback Elisabetta Bevacqua, Dirk Heylen,, Catherine Pelachaud, Isabella Poggi, Marc Schröder.
4 November 2000Bridging the Gap Workshop 1 Control of avatar gestures Francesca Barrientos Computer Science Division UC Berkeley.
Requirements of an ITS/Simulation Interoperability Standard (I/SIS) Presented by:Dick Stottler Stottler Henke Associates, Inc.
Abstract A Structured Approach for Modular Design: A Plug and Play Middleware for Sensory Modules, Actuation Platforms, Task Descriptions and Implementations.
Chapter 1. Cognitive Systems Introduction in Cognitive Systems, Christensen et al. Course: Robots Learning from Humans Park, Sae-Rom Lee, Woo-Jin Statistical.
ARCHITECTURES AND STANDARDS FOR IVAS AT THE SOCIAL COGNITIVE SYSTEMS GROUP H. van Welbergen, K. Bergmann, H. Buschmeier, S. Kahl, I. de Kok, A. Sadeghipour,
Thinking behind the environment for Making Construals (MCE)
DFKI GmbH, , R. Karger Perspectives for the Indo German Scientific and Technological Cooperation in the Field of Language Technology Reinhard.
1 Galatea: Open-Source Software for Developing Anthropomorphic Spoken Dialog Agents S. Kawamoto, et al. October 27, 2004.
ELEE 4303 Digital II Introduction to Verilog. ELEE 4303 Digital II Learning Objectives Get familiar with background of HDLs Basic concepts of Verilog.
ENTERFACE 08 Project #1 “ MultiParty Communication with a Tour Guide ECA” Final presentation August 29th, 2008.
SEESCOASEESCOA SEESCOA Meeting Activities of LUC 9 May 2003.
Dan Bohus Researcher Microsoft Research in collaboration with: Eric Horvitz, ASI Zicheng Liu, CCS Cha Zhang, CCS George Chrysanthakopoulos, Robotics Tim.
Current research in Intelligence Agents Victor Govindaswamy.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents mid-term presentation.
Conversational role assignment problem in multi-party dialogues Natasa Jovanovic Dennis Reidsma Rutger Rienks TKI group University of Twente.
W3C Multimodal Interaction Activities Deborah A. Dahl August 9, 2006.
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
The typical recent textbook listening task (Field, 1998) Pre-listening (for context and motivation) Extensive listening  questions to establish the situation;
Communication and Interpersonal Skills By Adel Ali 18/09/14371Communication Skills, Adel Ali.
NCP meeting Jan 27-28, 2003, Brussels Colette Maloney Interfaces, Knowledge and Content technologies, Applications & Information Market DG INFSO Multimodal.
Overview of Year 1 Progress Angelo Cangelosi & ITALK team
Hardware and system development:
Multimodal Human-Computer Interaction New Interaction Techniques 22. 1
PASSI (Process for Agent Societies Specification and Implementation)
Presentation transcript:

eNTERFACE’08 Multimodal Communication with Robots and Virtual Agents

Overview Context: Exploitation of multi-modal signals for the development of an active robot/agent listener Storytelling experience : –Speakers told a story of an animated cartoon they had just seen 1- See the cartoon 2- Tell the story to a robot or an agent

Overview Active listening : –During natural interaction, speakers see if the statements have been correctly understood (or at least heard). –Robots/agents should also have active listening skills… Characterization of multi-modal signals as inputs of the feedback model: –Speech analysis : prosody, keywords recognition, pauses –Partner analysis : face traking, smile detection Robot/agent feedbacks (outputs): –Lexical non-verbal behaviors Dialog management: –Feedback model: exploitation of both inputs and outputs signals Evaluation: –Storytelling experiences are usually evaluated by annotation

Organization: Workpackages: WP1: Speech feature extraction and analysis WP2: Partner analysis: face tracking and analysis WP3: Robot and Agent Behavior Analysis WP4: Dialog management for feedback behaviors WP5: Evaluation and Annotation WP6: Deliverables, reports.

Speech Analysis Automatic detection of prominence during the interaction Computational attention algorithms:

eNTERFACE’08 Multimodal Communication with Robots and Virtual Agents Speech analysis for prominence detection

Computational attention algorithms

Have more recently been tested for audio event detection M. MANCAS, L. COUVREUR, B. GOSSELIN, B. MACQ, 2007, "Computational Attention for Event Detection", Proceedings of ICVS Workshop on Computational Attention & Applications (WCAA-2007), Bielefeld, Germany, Mar In this project, we intend to test it for the automatic detection of salient speech events, for triggering avatar/robot feedback – Underlying hypothesis: listener is a child, with limited language knowledge  test the bottom-up approach, as opposed to the more language- driven top-down approach: A Top-down Auditory Attention Model For Learning Task Dependent Influences On Prominence Detection In Speech, Ozlem Kalinli and Shrikanth Narayanan, ICASSP’08, Computational attention algorithms

Partner analysis Analysis of human behaviour (non-verbal interaction). Development of a component able to detect the face and key features of feedback analysis: shaking head, smiling… Methodology: Face detection: Viola & Jones face detection, Head shaking: frequency analysis of interest points Smile detection: Combining colorimetric and geometric approaches

Robot and Agent Behavior Analysis Integratation of existing tools to produce an ECA/robot able to display expressive backchannels. The ECA architecture follows the SAIBA framwork. It is composed of several modules connected to each other via a Representation Language. The language FML (Functional Markup Language) connects the module 'intent planning' to 'behavior planning' and BML (Behavior) connects 'behavior planning to 'behavior realiser'. Modules are connected via psyclone, a white board architecture. Tasks: -define the capabilities the ECA/robot ought to have -create BML (Behavior Markup Language) entries for the lexicon -integrate modules that will endow ECA with such expressive capabilities. -work out carefully the synchronization scheme between modules, in particular between modules of Speaker and of Listener

Dialog Management Development of a feedback model with the respect of the input signals (common) and the output capabilities (behavior) Methodology: Representation of input data: –EMMA: Extensible MultiModal Annotation markup language –Definition of task-oriented representation Dialog management: –State Chart XML (SCXML): State Machine Notation for Control Abstraction –Interpretation of the speaker’s conversation

Evaluation and Annotation Investigate the impact of the feedback provided by the robot and the virtual agent on the user. A single model of feedback will be defined but implemented differently on the robot and the agent since they have different communication capabilities. The system will be partly simulated (WOZ). If time allows, a functional version of the system will be evaluated. Tasks: Evaluation protocol: scenario, variables … System implementation: WOZ Data collection: recordings Data analysis: coding schemes, analysis of annotation, computation of evaluation metrics

Thank for your attention…