Towards a Reactive Virtual Trainer Zsófia Ruttkay, Job Zwiers, Herwin van Welbergen, Dennis Reidsma HMI, Dept. of CS, University of Twente Amsterdam, The.

Slides:



Advertisements
Similar presentations
National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory
Advertisements

Information Society Technologies Information Society Technologies Workprogramme 2001 DG Information Society European Commission HEALTH Action Line I.1.1.
Expressive Gestures for NAO NAO TechDay, 13/06/2012, Paris Le Quoc Anh - Catherine Pelachaud CNRS, LTCI, Telecom-ParisTech, France.
Expressive Tangible Acoustic Interfaces Antonio Camurri, Corrado Canepa, and Gualtiero Volpe InfoMus Lab, DIST-University of Genova, Viale Causa 13, Genova,
Software Process Models
Let’s shake hands! On the coordination of gestures of humanoids Zsofi Ruttkay Herwin van Welbergen Balázs Varga.
Language and Cognition Colombo, June 2011 Day 8 Aphasia: disorders of comprehension.
Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
Virtual Reality Design Virtual reality systems are designed to produce in the participant the cognitive effects of feeling immersed in the environment.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents.
Herwin van Welbergen Dennis Reidsma Stefan Kopp.  Beyond turn taking interaction ◦ Continuous perception and behavior generation  Interpersonal coordination.
 INTRODUCTION  STEPS OF GESTURE RECOGNITION  TRACKING TECHNOLOGIES  SPEECH WITH GESTURE  APPLICATIONS.
On the parameterization of clapping Herwin van Welbergen Zsófia Ruttkay Human Media Interaction, University of Twente.
Understand the football simulation source code. Understand the football simulation source code. Learn all the technical specifications of the system components.
MUltimo3-D: a Testbed for Multimodel 3-D PC Presenter: Yi Shi & Saul Rodriguez March 14, 2008.
Industrial Ontologies Group Oleksiy Khriyenko, Vagan Terziyan INDIN´04: 24th – 26th June, 2004, Berlin, Germany OntoSmartResource: An Industrial Resource.
ISTD 2003, Thoughts and Emotions Interactive Systems Technical Design Seminar work: Thoughts & Emotions Saija Gronroos Mika Rautanen Juha Sunnari.
Wednesday, 24 June rd UKIBNET Workshop1 Distributing Cognition in the design of ubiquitous computers Chris Baber Pervasive Computing Group The University.
1 IUT de Montreuil Université Paris 8 Emotion in Interaction: Embodied Conversational Agents Catherine Pelachaud.
Non-invasive Techniques for Human Fatigue Monitoring Qiang Ji Dept. of Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute
Building the Design Studio of the Future Aaron Adler Jacob Eisenstein Michael Oltmans Lisa Guttentag Randall Davis October 23, 2004.
31 st October, 2012 CSE-435 Tashwin Kaur Khurana.
HAND GESTURE BASED HUMAN COMPUTER INTERACTION. Hand Gesture Based Applications –Computer Interface A 2D/3D input device (Hand Tracking) Translation of.
Electronic Visualization Laboratory University of Illinois at Chicago Interaction between Real and Virtual Humans: Playing Checkers R. Torre, S. Balcisoy.
The Entertainment Industry Meets Simulation: The Challenge of the Holodeck Bill Swartout Institute for Creative Technologies University of Southern California.
What is Software Engineering? the application of a systematic, disciplined, quantifiable approach to the development, operation, and maintenance of software”
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
Software Development Stephenson College. Classic Life Cycle.
Playful Stimulation against Parkinson’s Disease -
Mobile HCI Presented by Bradley Barnes. Mobile vs. Stationary Desktop – Stationary Users can devote all of their attention to the application. Very graphical,
Author: James Allen, Nathanael Chambers, etc. By: Rex, Linger, Xiaoyi Nov. 23, 2009.
Expressive Emotional ECA ✔ Catherine Pelachaud ✔ Christopher Peters ✔ Maurizio Mancini.
ITCS 6010 SALT. Speech Application Language Tags (SALT) Speech interface markup language Extension of HTML and other markup languages Adds speech and.
4/12/2007dhartman, CS A Survey of Socially Interactive Robots Terrance Fong, Illah Nourbakhsh, Kerstin Dautenhahn Presentation by Dan Hartmann.
Chapter 7. BEAT: the Behavior Expression Animation Toolkit
AsapRealizer 2.0: The Next Steps in Fluent Behavior Realization for ECAs Herwin van Welbergen, Ramin Yaghoubzadeh, Stefan Kopp Social Cognitive Systems.
Affective Interfaces Present and Future Challenges Introductory statement by Antonio Camurri (Univ of Genoa) Marc Leman (Univ of Gent) MEGA IST Multisensory.
Michael Lawo Using Wearable Computing Technology to Empower the Mobile Worker TNC 2009 Malaga Michael Lawo, Otthein Herzog, Peter.
© 2007 Tom Beckman Features:  Are autonomous software entities that act as a user’s assistant to perform discrete tasks, simplifying or completely automating.
APML, a Markup Language for Believable Behavior Generation Soft computing Laboratory Yonsei University October 25, 2004.
Interactive Spaces Huantian Cao Department of Computer Science The University of Georgia.
Full-body motion analysis for animating expressive, socially-attuned agents Elisabetta Bevacqua Paris8 Ginevra Castellano DIST Maurizio Mancini Paris8.
Human Interaction Development Using the Countess Quanta Robot Brad Pitney Yin Shi.
© Siemens AG, 2002 s CP RS Agenda The Role of IT for Accident-free Driving Interaction with driver’s physical condition Interaction with the roadside environment.
卓越發展延續計畫分項三 User-Centric Interactive Media ~ 主 持 人 : 傅立成 共同主持人 : 李琳山,歐陽明,洪一平, 陳祝嵩 水美溫泉會館研討會
ENTERFACE 08 Project 1 “MultiParty Communication with a Tour Guide ECA” Mid-term presentation August 19th, 2008.
A Common Ground for Virtual Humans: Using an Ontology in a Natural Language Oriented Virtual Human Architecture Arno Hartholt (ICT), Thomas Russ (ISI),
User-System Interaction: from gesture to action Prof. dr. Matthias Rauterberg IPO - Center for User-System Interaction TU/e Eindhoven University of Technology.
Toward a Unified Scripting Language 1 Toward a Unified Scripting Language : Lessons Learned from Developing CML and AML Soft computing Laboratory Yonsei.
Österreichisches Forschnungsinstitut für Artificial Intelligence Representational Lego for ECAs Brigitte Krenn.
4 November 2000Bridging the Gap Workshop 1 Control of avatar gestures Francesca Barrientos Computer Science Division UC Berkeley.
Human Figure Animation. Interacting Modules The ones identified –Speech, face, emotion Plus others: –Perception –Physiological states.
Animated Speech Therapist for Individuals with Parkinson Disease Supported by the Coleman Institute for Cognitive Disabilities J. Yan, L. Ramig and R.
CPSC 171 Introduction to Computer Science System Software and Virtual Machines.
ARCHITECTURES AND STANDARDS FOR IVAS AT THE SOCIAL COGNITIVE SYSTEMS GROUP H. van Welbergen, K. Bergmann, H. Buschmeier, S. Kahl, I. de Kok, A. Sadeghipour,
Aiming Computing Technology at Enhancing the Quality of Life of People with ALS Some Sketches on Directions in Minimal Signaling Communication Communication.
1 Galatea: Open-Source Software for Developing Anthropomorphic Spoken Dialog Agents S. Kawamoto, et al. October 27, 2004.
Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais.
Intelligent MultiMedia Storytelling System (IMSS) - Automatic Generation of Animation From Natural Language Input By Eunice Ma Supervisor: Prof. Paul Mc.
Multimodal Plan Representation for Adaptable BML Scheduling Dennis Reidsma, Herwin van Welbergen, Job Zwiers.
Immersive Virtual Characters for Educating Medical Communication Skills J. Hernendez, A. Stevens, D. S. Lind Department of Surgery (College of Medicine)
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
What is Multimedia Anyway? David Millard and Paul Lewis.
Multimodal and Natural computer interaction Evelina Stanevičienė.
MULTIMODAL AND NATURAL COMPUTER INTERACTION Domas Jonaitis.
Perceptive Computing Democracy Communism Architecture The Steam Engine WheelFire Zero Domestication Iron Ships Electricity The Vacuum tube E=mc 2 The.
Modeling Expressivity in ECAs
Towards lifelike Computer Interfaces that learn
Lindgren Robb, Tscholl Michael, Wang Shuai, Johnson Emily Presented By
Presentation transcript:

Towards a Reactive Virtual Trainer Zsófia Ruttkay, Job Zwiers, Herwin van Welbergen, Dennis Reidsma HMI, Dept. of CS, University of Twente Amsterdam, The Netherlands

page 2 Overview n RVT usage n Related work n RVT technological challenges –Architecture –Integration of reactive and proactive actions –Multi-modal sync n A close look at clapping - demos

page 3 RVT usage n RVT = IVA with expert and psychological knowledge of a real physiotherapist, to be used e. g. to: –prevent RSI for computer workers –preserve/restore weight and physical condition as (personal) trainer –act as physiotherapist to cure illnesses affecting motion n RVT is medium and emphatic consultant n Relevance for society –ageing population, unhealthy life-style, –human experts: low number, expensive, at certain locations n RVT usage context –PC camera in normal setting (homes, offices) –‘instructed’ by authorized person (may be the user, as well as developer) –can be adapted/extended

page 4 Related work Trainer calibrationMedium/consultantInputFeedbackMotiondemo/correctionExercise revisionAuthoring J. Davis, A. Bobick: Virtual PAT, MIT, moviesplit2 cam.assess ment --script S-P.Chao et al: Tai Chi synthesizer, m----nl script W. IJsselsteijn et al (Philips): Fun and Sports: Enhancing the Home Fitness Experience, Proc. of ICEC cheart- rate assess ment --? Sony’s EyeToy: Kinetic ‘game’, m/c1 cam.gener al, well- place d d-By User from pre- set choice/ty pes T. Bickmore: Laura & FitTrack1cdata to be typed assess ment --closed?

page 5 Own related work – Virtual Rap Dancer

page 6 Own related work – Virtual Conductor

page 7 RVT technological challenges n Vision-based perception, may be extended with biosignals n Reactive on exercise performance, physical state, overall performance n Smalltalk, exercise correction, plan revision n VRT body and motion parameters adaptable/calibrated n Authoring by human n Extensible by expert (new exercises) n Motion with music, speech or clapping (also as input for tempo) n Playground for multi-modal output generation n “Exercise motion intelligence”: timing, concatenation, idle poses, …

page 8 RVT architecture Calibration of user Multi-sensor integration Authoring scenario Exercise sce- nario revision Optical motion tracking Motion interpretation Motion specification Biosensing module(s) Acoustic beat tracking VT Monitoring the user Multi-modal feedback Motion demonstratio n Presentation of feedback of VT Planning action of VT Human expert User Interfaces

page 9 Multi-modal sync n Exercises are executed using several modalities –Body movement –Speech –Music –Sound (clap, foot tap) n Challenges –Synchronization –Monitoring user => real time (re)planning Exaggeration to point out details Speed up / slow down Feedback/correction …

page 10 Synchronization: related work n Classic approach in speech/gesture synchronization: –Speech leads, gesture follows n MURML (Kopp et al.) –No leading modality –Planning in sequential chunks containing one piece of speech and one aligned gesture –Co-articulation at the border of chunks n BML (Kopp, Krenn, Marsella, Marshall, Pelachaud, Pirker, Thórisson, Vilhjalmsson) –No leading modality –Synchronized alignment points in behavior phases –For now, aimed mainly at speech/gesture synchronization –In development

page 11 Synchronization: own previous work n Virtual Dancer Synchronization between music (beats) and dance animation Dance move selection by user interaction n Virtual Presenter Synchronization between speech, gesture, posture and sheet display Leading modality can change over time n GESTYLE markup language with par/seq and wait constructs

page 12 Close look at clapping stroke (hold) retraction (hold)

page 13 Clapping Exercise

page 14 Close look at clapping n Start with a simple clap exercise and see what we run into n The clap exercise: –Clap for the tempo of the beat of a metronome (later: of music) –When the palms touch, a clap sound is heard –Count while clapping, using speech synthesis Possible alignment at: word start/end, phonological peak start/center/end For now, we pick the center of the phonological peak, but we do generate the other alignment points for easy adaptation

page 15 Two examples for multi-modal sync n Specification in BML T n Planning in real-time – under/overspecification!

page 16 What if we speed up the tempo? n The clapping animation should be faster n Possibilities: –Lower amplitude? –Linear speedup? –Speedup of stroke? –Speedup of retraction? –A combination of above?

page 17 What if we slow down the metronome? n Slower clapping? (movies here) –Linear slowdown? –Slowdown of stroke? –Slowdown of retraction? –Hold at end of retraction (hands open)? –Hold after stroke (clap)? –A combination of above? n Back to idle position?

page 18 Open issues on planning n What do real humans do? n Do the semantics of a motion (clap) change if we change its amplitude or velocity profile? E.g. emotions, individual features n Smooth tempo changes n Automatic concatenation and inserted idle poses n Appropriate high-level parameters –Related (e.g. amplitude/speed)? –Different of parameters for communicative gestures (e.g. by Pelachaud)? n Amplitude and motion path specification n Is our synchronization system capable to re-plan in real time?