SceneMaker Intelligent Multimodal Visualisation of Natural Language Scripts Eva Hanser Dipl.-Des. (FH), M.Sc. Prof. Paul Mc Kevitt, Dr. Tom Lunney, Dr.

Slides:



Advertisements
Similar presentations
Towards an Ontology for Describing Emotions 1 st World Summit of the Knowledge Society WSKS08 Juan Miguel López 1, Rosa Gil 1, Roberto García 1, Idoia.
Advertisements

GMD German National Research Center for Information Technology Darmstadt University of Technology Perspectives and Priorities for Digital Libraries Research.
National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory
Design, prototyping and construction
Extraction and Visualisation of Emotion from News Articles Eva Hanser, Paul Mc Kevitt School of Computing & Intelligent Systems Faculty of Computing &
TeleMorph & TeleTuras: Bandwidth determined Mobile MultiModal Presentation Student: Anthony J. Solon Supervisors: Prof. Paul Mc Kevitt Kevin Curran School.
A Stepwise Modeling Approach for Individual Media Semantics Annett Mitschick, Klaus Meißner TU Dresden, Department of Computer Science, Multimedia Technology.
HOMER: A Creative Story Generation System Student: Dimitrios N. Konstantinou Supervisor: Prof. Paul Mc Kevitt School of Computing and Intelligent Systems.
Designing Multimedia with Fuzzy Logic Enrique Diaz de Leon * Rene V. Mayorga ** Paul D. Guild *** * ITESM, Guadalajara Campus, Mexico ** Faculty of Engineering,
Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents.
Empirical and Data-Driven Models of Multimodality Advanced Methods for Multimodal Communication Computational Models of Multimodality Adequate.
MediaHub: An Intelligent Multimedia Distributed Hub Student: Glenn Campbell Supervisors: Dr. Tom Lunney Prof. Paul Mc Kevitt School of Computing and Intelligent.
PGNET, Liverpool JMU, June 2005 MediaHub: An Intelligent MultiMedia Distributed Platform Hub Glenn Campbell, Tom Lunney, Paul Mc Kevitt School of Computing.
RRL: A Rich Representation Language for the Description of Agent Behaviour in NECA Paul Piwek, ITRI, Brighton Brigitte Krenn, OFAI, Vienna Marc Schröder,
Presented by: Thabet Kacem Spring Outline Contributions Introduction Proposed Approach Related Work Reconception of ADLs XTEAM Tool Chain Discussion.
VISIT: Virtual Intelligent System for Informing Tourists Kevin Meehan Intelligent Systems Research Centre Supervisors: Dr. Kevin Curran, Dr. Tom Lunney,
Topics Dr. Damian Schofield Director of Human Computer Interaction.
Aug 24, Fall 2005ITCS4010/50101 Design Teams Team Structure Interdisciplinary Teams.
1 Affective Learning with an EEG Approach Xiaowei Li School of Information Science and Engineering, Lanzhou University, Lanzhou, China
Presentation Outline  Project Aims  Introduction of Digital Video Library  Introduction of Our Work  Considerations and Approach  Design and Implementation.
IT 342 : Fundamentals of Multimedia Introduction & Multimedia Authoring.
Sunee Holland University of South Australia School of Computer and Information Science Supervisor: Dr G Stewart Von Itzstein.
Integration Of CG & Live-Action For Cinematic Visual Effects by Amarnath Director, Octopus Media School.
Smart Learning Services Based on Smart Cloud Computing
CMPD 434 MULTIMEDIA AUTHORING
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Dr. Sana’a Wafa Al-Sayegh
Hfghfg 360-PlayLearn: Gamification and Game-Based Learning for Virtual Learning Environments on interactive television Gerard Downes1, Paul Mc Kevitt1,
CONFUCIUS: An Intelligent MultiMedia Storytelling Interpretation and Presentation System Minhua Eunice Ma Supervisor: Prof. Paul Mc Kevitt School of Computing.
AmbiLearn: an ambient intelligent multimodal learning environment for children Jennifer Hyndman Supervisors: Dr. Tom Lunney, Prof. Paul Mc Kevitt Intelligent.
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
Animating Virtual Humans in Intelligent Multimedia Storytelling Minhua Eunice Ma and Paul Mc Kevitt School of Computing and Intelligent Systems Faculty.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Steps Toward an AGI Roadmap Włodek Duch ( Google: W. Duch) AGI, Memphis, 1-2 March 2007 Roadmaps: A Ten Year Roadmap to Machines with Common Sense (Push.
A MBI L EARN Ambient Intelligent Multimodal Learning Environment for Children Confirmation Viva June 2009 Jennifer Hyndman, B.Sc. (Hons.) Supervisors:
Building character animation for intelligent storytelling with the H-Anim standard Minhua Eunice Ma and Paul Mc Kevitt School of Computing and Intelligent.
University of Dublin Trinity College Localisation and Personalisation: Dynamic Retrieval & Adaptation of Multi-lingual Multimedia Content Prof Vincent.
MediaHub: An Intelligent MultiMedia Distributed Platform Hub Glenn Campbell, Tom Lunney & Paul Mc Kevitt School of Computing and Intelligent Systems Faculty.
APML, a Markup Language for Believable Behavior Generation Soft computing Laboratory Yonsei University October 25, 2004.
Entertainment Writing “Critical reporting”. “Critical Reporting” Critical reporting refers to the coverage of drama, music, art, and literature by print.
PAUL ALEXANDRU CHIRITA STEFANIA COSTACHE SIEGFRIED HANDSCHUH WOLFGANG NEJDL 1* L3S RESEARCH CENTER 2* NATIONAL UNIVERSITY OF IRELAND PROCEEDINGS OF THE.
SceneMaker: Multimodal Visualisation of Natural Language Film Scripts Dr. Minhua Eunice Ma School of Computing & Intelligent Systems Faculty of Computing.
CHAPTER TWO THE MAKING OF MULTIMEDIA: AND MULTIMEDIA DEVELOPMENT TEAM
TeleMorph & TeleTuras: Bandwidth determined Mobile MultiModal Presentation Student: Anthony J. Solon Supervisors: Prof. Paul Mc Kevitt Kevin Curran School.
Sign Language corpora for analysis, processing and evaluation A. Braffort, L. Bolot, E. Chételat-Pelé, A. Choisier, M. Delorme, M. Filhol, J. Segouat,
CONFUCIUS: an Intelligent MultiMedia storytelling interpretation & presentation system Minhua Eunice Ma Supervisor: Prof. Paul Mc Kevitt School of Computing.
SceneMaker: Automatic Visualisation of Screenplays School of Computing & Intelligent Systems Faculty of Computing & Engineering University of Ulster, Magee,
1 MPML and SCREAM: Scripting the Bodies and Minds of Life-Like Characters Soft computing Laboratory Yonsei University October 27, 2004.
SoFI – Streaming Music using Song Form Intelligence By Jonathan Doherty Supervisors: Dr. Kevin Curran, Prof. Paul Mc Kevitt School of Computing and Intelligent.
1 1. Representing and Parameterizing Agent Behaviors Jan Allbeck and Norm Badler 연세대학교 컴퓨터과학과 로봇 공학 특강 학기 유 지 오.
A MBI L EARN Ambient Intelligent Multimodal Learning Environment for Children 100 day review December 2008 Jennifer Hyndman Supervisors: Dr. Tom Lunney,
INTRODUCTION GORT is a virtual 3D modeling environment for computer programmers. Its main area of focus is to aid in the education of programmers learning.
Volgograd State Technical University Applied Computational Linguistic Society Undergraduate and post-graduate scientific researches under the direction.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Intelligent MultiMedia Storytelling System (IMSS) - Automatic Generation of Animation From Natural Language Input By Eunice Ma Supervisor: Prof. Paul Mc.
PGNET, Liverpool JMU, June 2005 MediaHub: An Intelligent MultiMedia Distributed Platform Hub Glenn Campbell, Tom Lunney, Paul Mc Kevitt School of Computing.
IMSTD:Intelligent Multimedia System for teaching Databases By : NAZLIA OMAR Supervisors: Prof. Paul Mc Kevitt Dr. Paul Hanna School of Computing and Mathematical.
SEESCOASEESCOA SEESCOA Meeting Activities of LUC 9 May 2003.
SceneMaker: Automatic Visualisation of Screenplays Eva Hanser Prof. Paul Mc Kevitt Dr. Tom Lunney Dr. Joan Condell School of Computing & Intelligent Systems.
VIRTUAL REALITY (VR) INTRODUCTION AND BASIC APPLICATIONS الواقع الافتراضي : مقدمة وتطبيقات Dr. Naji Shukri Alzaza Assist. Prof. of Mobile technology Dean.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents mid-term presentation.
Credit:  Multimedia has been used in many aspects in our lives, for example in the field of business, entertainment.
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
A New Approach to Decision-making within an Intelligent MultiMedia Distributed Platform Hub Glenn Campbell, Tom Lunney, Aiden Mc Caughey, Paul Mc Kevitt.
Using Commonsense Reasoning to Improve Voice Recognition.
Under Guidance of Mr. A. S. Jalal Associate Professor Dept. of Computer Engineering and Applications GLA University, Mathura Presented by Dev Drume Agrawal.
Visual Information Retrieval
Eleanor Mulholland, Paul Mc Kevitt & Tom Lunney
Presentation transcript:

SceneMaker Intelligent Multimodal Visualisation of Natural Language Scripts Eva Hanser Dipl.-Des. (FH), M.Sc. Prof. Paul Mc Kevitt, Dr. Tom Lunney, Dr. Joan Condell School of Computing & Intelligent Systems Faculty of Computing & Engineering University of Ulster, Magee {p.mckevitt, tf.lunney,

PRESENTATION OUTLINE Aims & Objectives Literature Review Project Proposal Relation to Other Work Conclusion and Future Work

AIMES & OBJECTIVES

AIMS Automatically generate affective virtual scenes from screenplays/play scripts Realistic visualisation of emotional aspects Enhance believability of virtual actors and scene presentation Multimodal representation with 3D animation, speech, audio and cinematography : AIMES & OBJECTIVES Input: Screen- play SceneMaker System Output: Animation

OBJECTIVES Emotions and semantic information from context Cognitive reasoning rules combined with commonsense and affective knowledge bases Automatic genre recognition from text Editing 3D content on mobile devices Design, implementation and evaluation of SceneMaker : AIMES & OBJECTIVES

LITERATURE REVIEW

Text layout analysis Semantic information on location, timing, props, actors, actions and manners, dialogue Parsing formal structure of screenplays (Choujaa and Dulay 2008) SEMANTIC TEXT PROCESSING : LITERATURE REVIEW INT. M.I.T. HALLWAY -- NIGHT Lambeau and Tom come around a corner. His P.O.V. reveals a figure in silhouette blazing through the proof on the chalkboard. There is a mop and a bucket beside him. As Lambeau draws closer, reveal that the figure is Will, in his janitor's uniform. There is a look of intense concentration in his eyes. LAMBEAU Excuse me! WILL Oh, I'm sorry. LAMBEAU What're you doing? WILL (walking away) I'm sorry. Screenplay Extract from ‘Good Will Hunting (1997)’

Emotion models: Basic emotions (Ekman and Rosenberg 1997) Pleasure-Dominance-Arousal (Mehrabian 1997) OCC – appraisal rules (Ortony et al. 1988) Personality models: OCEAN (McCrae and John 1992) Belief-Desire-Intention (Bratman 1987) Emotion recognition from text: Keyword spotting, lexical affinity, statistical models, fuzzy logic rules, machine learning, common knowledge, cognitive model COMPUTATION OF EMOTION AND PERSONALITY

Scripting Notation for visual appearance of animated characters Various XML-based annotation languages: EMMA (EMMA 2003) BEAT (Cassel et al. 2001) MPML & MPML3D (Dohrn and Brügmann 2007) AffectML (Gebhard 2005) VISUAL AND EMOTIONAL SCRIPTING : LITERATURE REVIEW

Automatic physical transformation and synchronisation of 3D model Manner influences intensity, scale, force, fluency and timing of an action Multimodal annotated affective video or motion captured data (Gunes and Piccardi 2006) MODELLING AFFECTIVE BEHAVIOUR : LITERATURE REVIEW AEOPSWORLD (Okada et al. 1999) Personality&Emotion Engine (Su et al. 2007) Greta (Pelachaud 2005)

SONAS – Environment visualisation (Kelleher et al. 2001) WordsEye – Scene composition (Coyne and Sproat 2001) ScriptViz – Screenplay visualisation (Liu and Leung 2006) CONFUCIUS – Action, speech & scene animation (Ma 2006) CAMEO – Cinematic and genre visualisation (Shim and Kang 2008) VISUALISING 3D SCENES WordsEye CONFUCIUSScriptVizCAMEO : LITERATURE REVIEW

Emotional speech synthesis (Schröder 2001) - Prosody rules Music recommendation systems - Categorisation of rhythm, chords, tempo, melody, loudness and tonality - Sad or happy music and genre membership (Cano et al. 2005) - Associations between emotions and music (Kuo et al. 2005) AUDIO GENERATION : LITERATURE REVIEW

Strategies for large 3D graphics data and knowledge bases on small, limited devices: Distribution of system architecture between server and mobile device SmartKom Mobile (Wahlster 2006) Multimodal Dialogue (Turunen et al. 2005) Rendering on mobile devices M3G (Java API), OpenGL ES, VRML, MPEG-4, H-Anim MULTIMODAL MOBILE INTERFACES : LITERATURE REVIEW

PROJECT PROPOSAL

Context consideration through natural language processing, common knowledge and reasoning methods Fine grained emotion distinction with OCC Extract genre and moods from screenplays Influence on Visualisation Enhance naturalism and believability Text-to-animation software prototype, SceneMaker KEY OBJECTIVES : PROJECT PROPOSAL

ARCHITECTURE OF SCENEMAKER Architecture of SceneMaker : PROJECT PROPOSAL

Language processing module of CONFUCIUS Part-of-speech tagger, Functional Dependency Grammars, WordNet, LCS database, temporal relations, visual semantic ontology Extensions : Context and emotion reasoning : ConceptNet, Open Mind Common Sense (OMCS), Opinmind, WordNet-Affect Text pre-processing : Layout analysis tool with layout rules Genre-recognition tool with keyword co-occurrence, term frequency and dialogue/scene length SOFTWARE AND TOOLS : PROJECT PROPOSAL

Visualisation module of CONFUCIUS H-Anim 3D models, VRML, media allocation, animation scheduling Extensions: Cinematic settings (EML), Affective animation models Media module of CONFUCIUS Speech Synthesis FreeTTS Extension: Automatic music selection User Interface for mobile and desktop VRML player, script writing tool, 3D editing : PROJECT PROPOSAL SOFTWARE AND TOOLS CONT.

Evaluating 4 aspects of SceneMaker: EVALUATION OF SCENEMAKER AspectEvaluation Correctness of screenplay analysis & visual interpretation Hand-animating scenes Effectiveness of output scenes Existing feature film scenes Suitability for genre typeScenes of unknown scripts categorised by viewers Functionality of interfaceTesting with drama students and directors : PROJECT PROPOSAL

RELATION TO OTHER WORK

Context reasoning to influence emotions requires common knowledge bases and context memory Text layout analysis to access semantic information Visualisation from sentence, scene or full script Automatic genre specification Automatic development of personality, social status, narrative role and emotions POTENTIAL CONTRIBUTIONS RELATION TO OTHER WORK

PROJECT PLAN : PROJECT PROPOSAL

0. Prelims, Abstract 1. Introduction 2. Literature Review Work in area of natural language processing and intelligent multimodal visualisation 3. Theoretical contribution Automatic, intelligent, multimodal and affective text-to- animation generation 4. Description of software prototype Implementation of the text-to-animation system, SceneMaker 5. Evaluation of SceneMaker Test results and evaluation of text-to-animation generation with SceneMaker 6. Conclusion Summary, Relation to other work, Future Work Appendices, References : PROJECT PROPOSAL THESIS OUTLINE

Eva Hanser and Paul Mc Kevitt. (2009). "NewsViz: Extraction and Visualisation of Emotions from News Articles". In: ISEA 2009,International Symposium on Electronic Art. Belfast, Northern Ireland, 23 Aug – 1 Sept PUBLICATIONS : PROJECT PROPOSAL

Automatic visualisation of affective expression of screenplays/play scripts Heightened expressivity, naturalness & artistic quality Assist directors, actors, drama students, script writers Focus on semantic interpretation, computational processing of emotions, reflecting affective behaviour and expressive multi-media scene composition Future work: Implementation of SceneMaker CONCLUSION AND FUTURE WORK

Thank you. QUESTIONS OR COMMENTS ?