AsapRealizer 2.0: The Next Steps in Fluent Behavior Realization for ECAs Herwin van Welbergen, Ramin Yaghoubzadeh, Stefan Kopp Social Cognitive Systems.

Slides:



Advertisements
Similar presentations
Testing Relational Database
Advertisements

Non-Classroom Teacher Evaluation Guidelines. The single most influential component of an effective school is the individual teachers within that school.
Expressive Gestures for NAO NAO TechDay, 13/06/2012, Paris Le Quoc Anh - Catherine Pelachaud CNRS, LTCI, Telecom-ParisTech, France.
Simulation Verification of Different Constraints in System Level Design in SystemC Piyush Ranjan Satapathy CS220 Class Project Presentation.
Herwin van Welbergen, Yuyu Xu, Marcus Thiebaux, Wei-Wen Feng, Jingqiao Fu, Dennis Reidsma, Ari Shapiro.
Social Interaction Functions Making Conversations Work.
Let’s shake hands! On the coordination of gestures of humanoids Zsofi Ruttkay Herwin van Welbergen Balázs Varga.
Intelligent Agents Russell and Norvig: 2
CA meeting Athens 2007 Raymond Cuijpers Ellen de Bruijn Hein van Schie Roger Newman-Norlund Majken Hulstijn Jurjen Bosga Ruud Meulenbroek Harold Bekkering.
Virtual Reality Design Virtual reality systems are designed to produce in the participant the cognitive effects of feeling immersed in the environment.
Herwin van Welbergen Dennis Reidsma Stefan Kopp.  Beyond turn taking interaction ◦ Continuous perception and behavior generation  Interpersonal coordination.
Systems Engineering in a System of Systems Context
On the parameterization of clapping Herwin van Welbergen Zsófia Ruttkay Human Media Interaction, University of Twente.
Towards a Reactive Virtual Trainer Zsófia Ruttkay, Job Zwiers, Herwin van Welbergen, Dennis Reidsma HMI, Dept. of CS, University of Twente Amsterdam, The.
Temporal Specification Chris Patel Vinay Viswanathan.
Quality of Service in IN-home digital networks Alina Albu 23 October 2003.
Lecture 3: Shared Workspace and Design Coordination Dr. Xiangyu WANG.
Principles of High Quality Assessment
Sunee Holland University of South Australia School of Computer and Information Science Supervisor: Dr G Stewart Von Itzstein.
Motor Control Theories
SDC PUBLICATIONS © 2012 Chapter 16 Assembly Modeling Learning Objectives:  Understand the Assembly Modeling Methodology  Create Parts in the Assembly.
What is it? A mobile robotics system controls a manned or partially manned vehicle-car, submarine, space vehicle | Website for Students.
Architectural Design.
Introduction to Jadex programming Reza Saeedi
UML Collaboration Diagram. Recap System Sequence Diagrams (SSD) UML for SSD Examples.
Modeling Driver Behavior in a Cognitive Architecture
SOFTWARE ENGINEERING BIT-8 APRIL, 16,2008 Introduction to UML.
-Nikhil Bhatia 28 th October What is RUP? Central Elements of RUP Project Lifecycle Phases Six Engineering Disciplines Three Supporting Disciplines.
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
2APL A Practical Agent Programming Language March 6, 2007 Cathy Yen.
Chapter7 Symbolic Communication and Language. Chapter Outline  Language and Verbal Communication  Nonverbal Communication  Social Structure and Communications.
Matthew Moccaro Chapter 10 – Deployment and Mobility PART II.
Recognition of meeting actions using information obtained from different modalities Natasa Jovanovic TKI University of Twente.
Expressive Emotional ECA ✔ Catherine Pelachaud ✔ Christopher Peters ✔ Maurizio Mancini.
Chapter 7. BEAT: the Behavior Expression Animation Toolkit
APML, a Markup Language for Believable Behavior Generation Soft computing Laboratory Yonsei University October 25, 2004.
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
What are the main differences and commonalities between the IS and DA systems? How information is transferred between tasks: (i) IS it may be often achieved.
What is a Business Analyst? A Business Analyst is someone who works as a liaison among stakeholders in order to elicit, analyze, communicate and validate.
Copyright 2002 Prentice-Hall, Inc. Chapter 2 Object-Oriented Analysis and Design Modern Systems Analysis and Design Third Edition Jeffrey A. Hoffer Joey.
(Business) Process Centric Exchanges
Greta MPEG-4 compliant Script based behaviour generator system: Script based behaviour generator system: input - BML or APML input - BML or APML output.
Black Box Testing Techniques Chapter 7. Black Box Testing Techniques Prepared by: Kris C. Calpotura, CoE, MSME, MIT  Introduction Introduction  Equivalence.
Chapter 10 Analysis and Design Discipline. 2 Purpose The purpose is to translate the requirements into a specification that describes how to implement.
EEL 5937 Agent models. EEL 5937 Multi Agent Systems Lecture 4, Jan 16, 2003 Lotzi Bölöni.
ENTERFACE 08 Project 1 “MultiParty Communication with a Tour Guide ECA” Mid-term presentation August 19th, 2008.
Chapter 4 프로세스 모델 Process Models
EMOTIONAL INTELLIGENCE. 2 Emotional Intelligence at Work.
Toward a Unified Scripting Language 1 Toward a Unified Scripting Language : Lessons Learned from Developing CML and AML Soft computing Laboratory Yonsei.
Outline The role of information What is information? Different types of information Controlling information.
Communication Though Nonverbal Behavior. Def.- bodily actions and vocal qualities that typically accompany a verbal message. They are usually interpreted.
Instructional Objective  Define an agent  Define an Intelligent agent  Define a Rational agent  Discuss different types of environment  Explain classes.
1 Hypermedia Design Models & Methodologies Dr Gary Wills IAM Research Group © University of Southampton.
Animated Speech Therapist for Individuals with Parkinson Disease Supported by the Coleman Institute for Cognitive Disabilities J. Yan, L. Ramig and R.
How (Not) to Give a Good Talk Steffen Koch, Daniel Maurer, Michael Stoll, Sebastian Volz, Andrés Bruhn with contributions by Michael Raschke and others.
ARCHITECTURES AND STANDARDS FOR IVAS AT THE SOCIAL COGNITIVE SYSTEMS GROUP H. van Welbergen, K. Bergmann, H. Buschmeier, S. Kahl, I. de Kok, A. Sadeghipour,
Knowledge Networks or “The Rise of the Machines”.
Process Asad Ur Rehman Chief Technology Officer Feditec Enterprise.
How to Become an Effective Speaker and Writer
Multimodal Plan Representation for Adaptable BML Scheduling Dennis Reidsma, Herwin van Welbergen, Job Zwiers.
1 Software Testing & Quality Assurance Lecture 15 Created by: Paulo Alencar Modified by: Frank Xu.
Cognitive Architectures and General Intelligent Systems Pay Langley 2006 Presentation : Suwang Jang.
Chapter 5 Motor Programs 5 Motor Programs C H A P T E R.
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
Systems Architectures System Integration & Architecture.
CHAPTER
Lifelong learning... Professional training should be appropriate in continuous and systematic process (T. Guskey, 2004). Three aspects of education reform:
Object-Oriented Analysis
Situation Awareness through Agent Based
Motor Control Theories
Presentation transcript:

AsapRealizer 2.0: The Next Steps in Fluent Behavior Realization for ECAs Herwin van Welbergen, Ramin Yaghoubzadeh, Stefan Kopp Social Cognitive Systems Group

Behavior realization for embodied conversational agents (ECAs)  Speech, gesture, gaze, facial expression, …  Some challenges:  Multimodal synchrony E.g. phonological synchrony rule (McNeill 1995) Important for meaning (Habets et al. 2011)  Fluent behavior realization Incrementally constructed, low latency Interruptible Adaptability

BML

The bigger picture: Asap (Kopp et al., 2013)  Standardization of components  SAIBA  BML

Fluent behavior in ECAs  Traditional turn-based interaction 1. Fully analyze user contribution 2. Plan your own contribution 3. Execute your own contribution ballistically  Modern ECA-systems steer away from this paradigm

Fluent interaction, related work

Fluent interaction, AsapRealizer

AsapRealizer  Collaboration of HMI, University of Twente and Social Cognitive Systems Group, Bielefeld University  Built on >10 years of research on fluent behavior realization  Basis:  ACE (Kopp and Wachsmuth 2004): incremental utterance construction  Elckerlyc (van Welbergen et al., 2010): interactional coordination  => AsapRealizer 1.0 (van Welbergen et al., 2012) Specifications for fluent behavior realization in BML and BMLA Implementation mainly in gesture  AsapRealizer 2.0  Adds the implementation of fluent behavior realization capabilities for Speech, gaze, facial expression

Requirements for fluent behavior realization  Incremental plan construction  Graceful interruption  Adaptations of ongoing behavior  Top-down  Bottom-up  Adjustments to the changing environment

Comparison with other realizers: internal design AsapRealizer Other realizers

Incremental plan construction  Reduce latency, make the ECA more reactive  Start behavior early  Fold part of the plan construction in realization of previous increments  Has biological basis (e.g. Levelt 1989, McNeill 1995)  The increments connect smoothly, suprasegmental properties  Rhythm, sentence intonation in speech  Retraction skipping in gesture  But not always: whether or not smooth connections occur has communicative meaning  E.g. marking information boundaries (Kendon 1980)

Non-incremental specification of gesture-sequences in SmartBody (Xu et al.,2014)

Incremental BML specifications

Incrementality: suprasegmental effects  For gesture: preparation and retraction phases need to be constructed on the fly  Subject to change during realization start position of the preparation the hand position at the start/end of the stroke posture ground state at the end of the gesture  Similarly for gaze  Subject to change during realization position of the gaze target the position of e.g. the head and eyes at the start the gaze ground state at the end of the gaze implemented in novel gaze system based off (Grillon and Thalmann, 2009) with some improvements for biological realism and adaptivity

Incremental generation demo Non-incremental Naïve incremental AsapRealizer 2.0 (Using Inprotk_iSS: Baumann & Schlangen 2012) startup delay: 1400 ms startup delay: 480 msstartup delay: 50 ms IVA 2014 is an interdisciplinary annual conference and the main leading scientific forum for presenting research on modelling, developing and evaluating intelligent virtual agents.

Incremental generation demo AsapRealizer 2.0 Nonincremental

Graceful interruption: specification

Graceful interruption  More than simply stopping speech/gesture/gaze  Speech can be interrupted instantly or at word or phoneme boundaries  Gesture/gaze/face  Interruption is implemented by returning to a previously specified ‘ground state’  Ground state my be changed using BML 1.0 behaviors (gazeShift, postureShift)

Top-down adaptation: specification

Top-down adaptation: realization support  Requires flexible execution engines  Facial animation (Paul 2010): allow on-the- fly intensity changes of morphs, AUs, MPEG4, emotion  Body animation: Procedural animation that allows arbitrary parameters and mathematical formulas (van Welbergen et al., 2010) Parameter values may be changed on the fly Captures animation specified in other procedural animation systems (e.g. Ahn et al., 2012) as a subset Supports MURML (Kopp and Wachsmuth, 2004)  TTS: inproTK_iss, supports loudness, pitch, tempo

Top-down adaptation

Alignment to anticipated events: specification

Alignment to anticipated external events: Anticipators  An anticipator manages a TimePeg on the PegBoard  Uses perceptions of the world/interlocutor to continuously update the timing  Extrapolate perceptions in prediction of the timing of future events End of interlocutor turn, next beat in music, timing of user movement events in fitness exercises

Discussion  AsapRealizer’s behavior realization capabilities are (mostly) beyond the state of the art  What’s currently missing: model how a virtual human steered by AsapRealizer changes the environment  AsapRealizer is eminently suitable for behavior realization in very dynamic contexts  Fluent behavior realization is just one aspect of a successful behavior realizer  Other realizers may have an edge over AsapRealizer in other aspects, e.g. behavior surface realism This may make them more suitable in contexts where less flexible behavior suffices But: fluent behavior contributes to communicative realism (Blascovic 2002)/interactional adequacy (Bauman & Schlangen 2013)

Thanks for your attention Questions?

Eastereggs

Ongoing work  AsapRealizer allows the realization of many aspects of fluent behavior  It opens up many exciting possibilities for behavior and intent planning  Fluent turn-taking Using interruption, keep challenged turn, let yourself be (gracefully) interrupted Each can be realized in widely varying ways (speaking loud, with high pitch, fast,...) Which to select for a desired friendliness, dominance, effectivity,...

Adaptation  BML blocks specify constraints on timing and shape  Generally underspecified, realizers have some realization freedom  AsapRealizer allows on the fly changes to shape and timing of behavior  Maintaining the time constraints specified in BML

Constraint representation

Comparison with other realizers: specification of a gesture sequence... Is there anything that, besides what he wants anything you want to work on SmartBody (Xu et al. 2014) Is there anything that, besides what he wants AsapRealizer...

Preplanning/activation  Preplan a BML block for activation at a later time (or never)  Allows behavior realization to start instantly  For example in contexts where only a few responses are valid  Preplan while listening  Activate the ‘proper’ response near instantly once you have the turn

Employing fillers (uhm)  Keep the turn without having a plan at hand (yet)  In some contexts preferred over waiting with speaking until all information is available (Bauman and Schlangen 2013)  But may also communicate unintended information (hesitation, uncertainty)  Give the behavior planner control over whether or not they should occur

Employing fillers (example) The car goes around the corner and turns right.

The bigger picture: Asap  Standardization of components  SAIBA  BML

BML

Realizing BML the traditional way

AsapRealizer

Constraint representation

BML Feedback  The Behavior Planner makes changes in the ongoing behavior plan  Requires information on its progress from the Realizer  Decide whether some message can be considered as ‘arrived’ at the user and act accordingly  And on (predicted) timing  Decide whether a virtual human should interrupt itself or continue speaking after something urgent comes up, based (among other things) on a prediction of how long it will take to deliver the current content  AsapRealizer implements BML 1.0 feedback  Progress feedback: informs of timing of delivered behavior  Prediction feedback: informs of predicted timing events of ongoing behavior

Ongoing work: IU representation of the behavior plan  BML blocks incrementally construct and adapt a behavior plan  BML Feedback provides updates on the progress of the plan  Can we express the plan in an IU structure in the Behavior Planner?  Challenging to express Multimodal synchrony/time-constraints Graceful interruption

Ongoing work: IU representation of the behavior plan

The BML scheduling process  Scheduling the multimodal behavior plan

The behavior plan  BML  Specifies behaviors and the constraints between them  The multimodal behavior plan  Specified on the basis of, e.g., communicative intent  Incrementally constructed from BML blocks  The motor plan  Executes the behavior plan on the embodiment of the virtual human, using sound, joint rotations, FAPs, …  Making the motor plan flexible  Achieve interpersonal coordination  Allow on the fly adjustments of ongoing behavior  But do not violate the specified constraints

Flexible plan representation  The motor plan  Describes the low level execution of motor behavior on the embodiment of the virtual human: joint rotations, MPEG4 FAP movement, sound, …etc  To be flexible, the plan representation must  maintain information about the relation between elements in the plan and the original BML expressions from which they originated  be capable of expressing all the constraints  be capable of being modified…  …in such a way that the (remaining) constraints automatically stay satisfied

AsapRealizer’s plan representation  Central: the PegBoard  Each sync point is assigned to a TimePeg on this board  Sync points that are connected by an ‘at’ constraint share the same TimePeg  TimePegs can be moved, changing the time of the associated sync(s)  TimePegs provide local timing (irt the start of the BML block)  Each TimePeg is connected to a BMLBlockPeg  BMLBlockPegs provide global timing

AsapRealizer’s plan representation  OffsetPeg: links to another TimePeg and moves with it, retaining a time offset  All ‘at’ constraints can be expressed in Pegs, OffsetPegs en Blockpegs  After en before constraints require a specific pair of Pegs Not implemented yet (but: trivial to design and implement)

Example: solving BML to TimePegs As you can see on this painting,... <gesture id="point1" start="walk1:relax" type="POINT" target="painting1" stroke="speech1:s1+0.5"/>

Managing block timing

Architecture  Separated parsing, scheduling, execution  Central: the PegBoard  Engines handle unimodal motor plans  Sync points of PlanUnits are connected by TimePegs  Anticipators may make time adjustments through TimePegs

Architecture  The scheduler communicates with a set of engines to set up the multimodal behavior plan  It knows for each BML behavior what Engine handles it  Engines have a standardized interface with functions to: Add a BML behavior to the motor plan Remove a BML behavior from the motor plan Resolve unknown time constraints on a BML behavior given its known time constraints Check which BML behavior in the motor plan are currently invalid  Note that communication with Engines is in terms of BML behaviors

Managing adjustments of the plan  Interruption of a behavior  Will automatically shorten the block’s end  Satisfies block append constraint  Time adjustment of a sync point  Through its associated TimePeg  Will also move syncs that are connected to it with ‘at’ constraints  Might invalidate other constraints Each Engine has functionality to check for this Solution: drop behavior, drop constraint,.. And communicate back to the behavior planner

ECA ‘ground’ state  Gaze: default gaze target  Face: default/current ‘neutral’ facial expression  Posture: current rest posture  Locomotion: position  Can be set through BML with e.g. gazeShift, faceShift, postureShift  Gaze/face/gesture behavior is layered on top of these

Comparison with other realizers EMBOTGretaSMBACEElck.AsapR 1.0 AsapR 2.0 Incrementality2/101/102/104/10 5/109/10 Graceful interruption 0/4 1/40/41/42/44/4 Top down adaptation 2/50/52/5 4/5 5/5 Bottom up adaptation 0/3 2/30/31/32/3 Interacting with a changing world 0/7 6/72/73/7 Total4/291/2911/2910/2 9 12/2915/2923/29

Incremental generation: implementation  Incremental realization entails more than simply concatenating audio files and animations  Implement suprasegmental effects  Inprotk_iSS (Baumann & Schlangen 2012)  Automatic gesture retraction skipping