Collection and Analysis of Multimodal Interaction in Direction Giving Dialogues Seikei University Takeo TsukamotoYumi Muroya Masashi Okamoto Yukiko Nakano.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Demonstrators: Mudasir Nazir(08-CS-41).  I am highly addicted to this field.  Working with W3C in research program(building CSS for creating web site.
Fast Algorithms For Hierarchical Range Histogram Constructions
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents.
Automatically Annotating and Integrating Spatial Datasets Chieng-Chien Chen, Snehal Thakkar, Crail Knoblock, Cyrus Shahabi Department of Computer Science.
Effective Communication
Human interaction is not constructed as a single channel – it is multimodal. Speech and gestures correlate to convey meaning. Moreover, human interaction.
SOMM: Self Organizing Markov Map for Gesture Recognition Pattern Recognition 2010 Spring Seung-Hyun Lee G. Caridakis et al., Pattern Recognition, Vol.
I1-[OntoSpace] Ontologies for Spatial Communication John Bateman, Kerstin Fischer, Reinhard Moratz Scott Farrar, Thora Tenbrink.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
MUltimo3-D: a Testbed for Multimodel 3-D PC Presenter: Yi Shi & Saul Rodriguez March 14, 2008.
Gaze Awareness for Videoconferencing: A Software Approach Nicolas Werro.
Data Analysis, Interpretation, and Reporting
Tracking Video Objects in Cluttered Background
Communicating with Avatar Bodies Francesca Barrientos Computer Science UC Berkeley 8 July 1999 HCC Research Retreat.
Text Mining: Finding Nuggets in Mountains of Textual Data Jochen Dijrre, Peter Gerstl, Roland Seiffert Presented by Huimin Ye.
Text Mining: Finding Nuggets in Mountains of Textual Data Jochen Dijrre, Peter Gerstl, Roland Seiffert Presented by Drew DeHaas.
1 Contextual Interview Shahnewaz A. Jolly CPSC 681: Research Methods in Human Computer Interaction Instructor: Dr. Saul Greenberg Date: November 4, 2009.
Science Inquiry Minds-on Hands-on.
Building the Design Studio of the Future Aaron Adler Jacob Eisenstein Michael Oltmans Lisa Guttentag Randall Davis October 23, 2004.
Acoustic and Linguistic Characterization of Spontaneous Speech Masanobu Nakamura, Koji Iwano, and Sadaoki Furui Department of Computer Science Tokyo Institute.
Automated malware classification based on network behavior
Human-Computer Interaction IS 588 Spring 2007 Week 4 Dr. Dania Bilal Dr. Lorraine Normore.
«Tag-based Social Interest Discovery» Proceedings of the 17th International World Wide Web Conference (WWW2008) Xin Li, Lei Guo, Yihong Zhao Yahoo! Inc.,
Storytelling Performance Skills. Voice Mechanics Speaks with an appropriate volume for the audience to hear. Employs clear enunciation. Uses non- monotonous,
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Human Gesture Recognition Using Kinect Camera Presented by Carolina Vettorazzo and Diego Santo Orasa Patsadu, Chakarida Nukoolkit and Bunthit Watanapa.
Chapter 7. BEAT: the Behavior Expression Animation Toolkit
APML, a Markup Language for Believable Behavior Generation Soft computing Laboratory Yonsei University October 25, 2004.
Department of Computer Science and Engineering, CUHK 1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal Video Conference Archives Indexing System.
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
Multi-modal Reference Resolution in Situated Dialogue by Integrating Linguistic and Extra-Linguistic Clues Ryu Iida Masaaki Yasuhara Takenobu Tokunaga.
Shape-based Similarity Query for Trajectory of Mobile Object NTT Communication Science Laboratories, NTT Corporation, JAPAN. Yutaka Yanagisawa Jun-ichi.
Boundary Assertion in Behavior-Based Robotics Stephen Cohorn - Dept. of Math, Physics & Engineering, Tarleton State University Mentor: Dr. Mircea Agapie.
1 MPML and SCREAM: Scripting the Bodies and Minds of Life-Like Characters Soft computing Laboratory Yonsei University October 27, 2004.
Conversation Analysis Introduction to Conversation Analysis 2e Anthony J. Liddicoat, March 2011.
Communication Additional Notes. Communication Achievements 7% of all communication is accomplished Verbally. 55% of all communication is achieved through.
Towards a Method For Evaluating Naturalness in Conversational Dialog Systems Victor Hung, Miguel Elvir, Avelino Gonzalez & Ronald DeMara Intelligent Systems.
C. Lawrence Zitnick Microsoft Research, Redmond Devi Parikh Virginia Tech Bringing Semantics Into Focus Using Visual.
Eye Tracking In Evaluating The Effectiveness OF Ads Guide : Dr. Andrew T. Duchowski.
Enhancement in Online Laboratory Learning Principal Investigators: Drs. Richard Chiou and Yongjin Kwon Research Assistants: Sweety Das Agarwal and Yueh-Ting.
4 November 2000Bridging the Gap Workshop 1 Control of avatar gestures Francesca Barrientos Computer Science Division UC Berkeley.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Chapter 4 Marketing Research and Information Systems Dr. Franck VIGNERON.
Copyright © Allyn & Bacon 2008 Intelligent Consumer Chapter 14 This multimedia product and its contents are protected under copyright law. The following.
Animated Speech Therapist for Individuals with Parkinson Disease Supported by the Coleman Institute for Cognitive Disabilities J. Yan, L. Ramig and R.
AN INTELLIGENT ASSISTANT FOR NAVIGATION OF VISUALLY IMPAIRED PEOPLE N.G. Bourbakis*# and D. Kavraki # #AIIS Inc., Vestal, NY, *WSU,
Animating Idle Gaze Humanoid Agents in Social Game Environments Angelo Cafaro Raffaele Gaito
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
Driver Attention for Information Display on Variable Message Signs with Graphics and Texts Chien-Jung Lai, Chi-Ying Wang National Chin-Yi University of.
Student: Ibraheem Frieslaar Supervisor: Mehrdad Ghaziasgar.
Electronic visualization laboratory, university of illinois at chicago Towards Lifelike Interfaces That Learn Jason Leigh, Andrew Johnson, Luc Renambot,
R ITSUMEIKAN 13 th International Conference on Multimodal Interaction (ICMI 2011) Alicante, Spain, Nov. 16th, 2011 COMmunication Software Lab. Making Virtual.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Conversational role assignment problem in multi-party dialogues Natasa Jovanovic Dennis Reidsma Rutger Rienks TKI group University of Twente.
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
Presented By Meet Shah. Goal  Automatically predicting the respondent’s reactions (accept or reject) to offers during face to face negotiation by analyzing.
An AV Control Method Using Natural Language Understanding Author : M. Matsuda, T. Nonaka, and T. Hase Date : Speaker : Sian-Lin Hong IEEE Transactions.
V k equals the vector difference between the object and the block across the first and last frames in the image sequence or more formally: Toward Learning.
King Faisal University جامعة الملك فيصل Deanship of E-Learning and Distance Education عمادة التعلم الإلكتروني والتعليم عن بعد [ ] 1 جامعة الملك فيصل عمادة.
Chapter 3 Choosing Information & Communications Technologies that Fit the Research Design Janet Salmons, PhD.
NBKeyboard: An Arm-based Word-gesture keyboard
Social Knowledge Mining
Identifying Confusion from Eye-Tracking Data
Pilar Orero, Spain Yoshikazu SEKI, Japan 2018
Communicating with Avatar Bodies
Dialogue State Tracking & Dialogue Corpus Survey
The Scientific Method.
THE ASSISTIVE SYSTEM SHIFALI KUMAR BISHWO GURUNG JAMES CHOU
Random Neural Network Texture Model
Presentation transcript:

Collection and Analysis of Multimodal Interaction in Direction Giving Dialogues Seikei University Takeo TsukamotoYumi Muroya Masashi Okamoto Yukiko Nakano Japan Towards an Automatic Gesture Selection Mechanism for Metaverse Avatars

Overview Introduction Research Goal and Questions Approach Data Collection Experiment Analysis Conclusion and Future Work

Introduction Online 3D virtual worlds based on Metaverse application is growing steadily in popularity Ex. : Second Life (SL) ⇒ The communication method is limited to: Online chat with speech balloons Manual gesture generation Hello

Introduction(Cont.) Human face-to-face communication is largely dependent on non-verbal behaviors Ex. direction giving dialogues Many spatial gestures are used in order to illustrate directions and physical relationships of buildings and landmarks How can we implement natural non-verbal behaviors into Metaverse application ?

Research Goal and Questions Goal Establish natural communication between avatars in Metaverse based on human face-to- face communication Research Questions Automation : gesture selection How to automatically generate proper gestures? Comprehensibility : gesture display How to intelligibly display gestures to interlocutor?

Previous work An automatic gesture selection mechanism for Japanese chat texts in Second Life [Tsukamoto,2010] /2 you keep going straight in this road, then you will be able to find a house having a round window on your left./

Proxemics Proxemics is important to implement comprehensible gestures in Metaverse Previous work doesn’t consider proxemics ⇒ There are some cases when avatar’s gesture becomes unintelligible to the others

Approach Conduct an experiment to collect human gestures in direction giving dialogues Collect participant’s verbal and non-verbal data Analyze the relationship between gestures and proxemics

Data Collection Experiment Direction Giver (DG) Know the way to any place on campus of Seikei Univ. Direction Receiver (DR) Know nothing about the campus of Seikei Univ. Experimental Procedure The DR asks a way to a specific building The DG explains how to get to the building DG DR

Experimental Instruction Direction Receiver Instructed to completely understand the way to the goal through a conversation with the DG Direction Giver I nstructed to confirm that the DR understood the direction correctly after the explanation was finished

Experimental Materials Each pair recorded a conversation for each goal place

Experimental Equipments Right armAbdominal HeadShoulder Headset microphone Equipments Experimental Video

Collected Data Video Data Transcription of Utterances Motion Capture Data

Analysis Investigated DG’s gesture distribution with respect to proxemics Analyzed 30 dialogues collected from 10 pairs Analysis was focused on the movements of DG’s right arm during gesturing

Automatic Gesture Annotation Extracted features  Movement of position(x, y, z)  Rotation(x, y, z)  Relative position of the right arm to shoulder(x, y, z)  Distance between right arm and shoulder Binary judge  Gesturing / Not gesturing It is very time consuming to manually annotate nonverbal behaviors Automatically annotated the gesture occurrence More than 77% of the gestures are right arm gestures Built a decision tree that identified right arm gestures Weka J48 was used for the decision tree learning

Automatic Gesture Annotation (Cont.) As the result of 10-fold cross validation, the accuracy is 97.5% Accurate enough for automatic annotation Example of automatic annotation

Gesture Display Space Defined as the overlap among the DG’s front area, the DR’s front area, and the DR’s front field of vision DGDG DR Direction Receiver Direction Giver Gesture Display Space Center Distance of DG from the center Distance of DR from the center DR’s body Direction vector DG’s body Direction vector DR’s front field of Vision

CategoryConditions Normal(12/30) 450mm ≦ Distance Both-center ≦ 950mm Close_to_DG(4/30) Distance DG-center ≦ 450mm 450mm ≦ Distance DR-center ≦ 950mm Close_to_DR(8/30) Distance DR-center ≦ 450mm 450mm ≦ Distance DG-center ≦ 950mm Close_to_Both(2/30) Distance Both-center ≦ 450mm Far_from_Both(4/30) 950mm ≦ Distance DG-center or 950mm ≦ Distance DR-center Define 450mm to 950mm as the standard distance from the center of the gesture display space Human arm length is 60cm to 80cm, by adding 15cm margin Categories of Proxemics

Analysis : Relationship between Proxemics and Gesture Distribution Analyze the distribution of gestures by plotting the DG’s right arm position Normal Close_to_DGClose_to_DRClose_to_Both Similar Wider Smaller

Analysis : Relationship between Proxemics and Gesture Distribution(Cont.) Close_to_Both < Normal = Close_to_DG < Close_to_Both

Applying the Proxemics Model Create avatar gestures based on our proxemics model To test whether the findings are applicable Close_to_DG Close_to_DR

Conclusion Conducted an experiment to collect human gestures in direction giving dialogues Investigated the relationship between the proxemics and the gesture distribution Proposed five types of proxemics characterized by the distance from the gesture display space Found that the gesture distribution range was different depending on the proxemics of the participants

Future Work Establish a computational model of determining gesture direction Examine the effectiveness of the model whether the users perceive the avatar’s gestures being appropriate and informative

Thank you for your attention

Related work [Breitfuss, 2008] Built a system that automatically adds gestural behavior and eye gaze Based on linguistic and contextual information of input text [Tepper, 2004] Proposed a method for generating novel iconic gestures Used spatial information about locations and shape of landmarks to represent concept of words From a set of parameters, iconic gestures are generated without relying on a lexicon of gesture shapes [Bergmann, 2009] Represented individual variation of gesture shapes using Bayesian network Built an extensive corpus of multimodal behaviors in direction-giving and landmark description task