Authoring Directed Gaze for Full-Body Motion Capture

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Managerial Decision Modeling with Spreadsheets
Detail to attention: Exploiting Visual Tasks for Selective Rendering Kirsten Cater 1, Alan Chalmers 1 and Greg Ward 2 1 University of Bristol, UK 2 Anyhere.
Real-Time Human Pose Recognition in Parts from Single Depth Images Presented by: Mohammad A. Gowayyed.
Chapter 6: Visual Attention. Scanning a Scene Visual scanning – looking from place to place –Fixation –Saccadic eye movement Overt attention involves.
Character Setup Character Setup is the process of creating handles and controls for anything that a character animator will need to adjust in order to.
Introduction to Data-driven Animation Jinxiang Chai Computer Science and Engineering Texas A&M University.
Cognitive modelling (Cognitive Science MSc.) Fintan Costello
Probabilistic video stabilization using Kalman filtering and mosaicking.
Computer-Based Animation. ● To animate something – to bring it to life ● Animation covers all changes that have visual effects – Positon (motion dynamic)
Computer Animation CS 445/645 Fall Let’s talk about computer animation Must generate 30 frames per second of animation (24 fps for film) Issues.
Algirdas Beinaravičius Gediminas Mazrimas.  Introduction  Motion capture and motion data  Used techniques  Animating human body  Problems  Conclusion.
Input: Original intensity image. Target intensity image (i.e. a value sketch). Using Value Images to Adjust Intensity in 3D Renderings and Photographs.
Eyes Alive Sooha Park - Lee Jeremy B. Badler - Norman I. Badler University of Pennsylvania - The Smith-Kettlewell Eye Research Institute Presentation Prepared.
M ULTIFRAME P OINT C ORRESPONDENCE By Naseem Mahajna & Muhammad Zoabi.
Manipulating Attention in Computer Games Matthias Bernhard, Le Zhang, Michael Wimmer Institute of Computer Graphics and Algorithms Vienna University of.
Expressive Emotional ECA ✔ Catherine Pelachaud ✔ Christopher Peters ✔ Maurizio Mancini.
1 Nonparametric Statistical Techniques Chapter 17.
The geometry of the system consisting of the hyperbolic mirror and the CCD camera is shown to the right. The points on the mirror surface can be expressed.
Learning Sequence Motifs Using Expectation Maximization (EM) and Gibbs Sampling BMI/CS 776 Mark Craven
Enhanced Regulatory Sequence Prediction Using Gapped k-mer Features 王荣 14S
Learning video saliency from human gaze using candidate selection CVPR2013 Poster.
Describe how reaching and grasping abilities develop in the first year of life.
Humanoid دکتر سعید شیری قیداری Amirkabir University of Technology Computer Engineering & Information Technology Department.
Feedforward Eye-Tracking for Training Histological Visual Searches Andrew T. Duchowski COMPUTER SCIENCE, CLEMSON UNIVERSITY Abstract.
1 Nonparametric Statistical Techniques Chapter 18.
Computer Animation CS 446 September 5, 2001.
Hidden Markov Models BMI/CS 576
EYE TRACKING TECHNOLOGY
Introduction   Many 3-D pronunciation tutors with both internal and external articulator movements have been implemented and applied to computer-aided.
Introduction   Many 3-D pronunciation tutors with both internal and external articulator movements have been implemented and applied to computer-aided.
Background Method Results Objectives Results Discussion References
3.1 Clustering Finding a good clustering of the points is a fundamental issue in computing a representative simplicial complex. Mapper does not place any.
Reconstruction For Rendering distribution Effect
Chapter 5 Force and Motion
Predetermined Motion Time Systems
Contribution of spatial and temporal integration in heading perception
Chapter 10 Verification and Validation of Simulation Models
Sparse Rig Parameter Optimization for Character Animation
Categorizing sex and identity from the biological motion of faces
Tremor Detection Using Motion Filtering and SVM Bilge Soran, Jenq-Neng Hwang, Linda Shapiro, ICPR, /16/2018.
Chapter 5 Force and Motion
Hidden Markov Models Part 2: Algorithms
Bowei Tang, Tianyu Chen, and Christopher Atkeson
A User Attention Based Visible Watermarking Scheme
Identifying Confusion from Eye-Tracking Data
Before-After Studies Part I
Volume 28, Issue 7, Pages e5 (April 2018)
Perceptual Echoes at 10 Hz in the Human Brain
Volume 21, Issue 19, Pages (October 2011)
Easy Generation of Facial Animation Using Motion Graphs
Cost Estimation I've got Bad News and Bad News!.
COMP60621 Fundamentals of Parallel and Distributed Systems
Colin J. Palmer, Colin W.G. Clifford  Current Biology 
WELCOME.
Synthesis of Motion from Simple Animations
Inactivation of Medial Frontal Cortex Changes Risk Preference
Computer Animation Displaying animation sequences raster animation
Neural Mechanisms of Visual Motion Perception in Primates
Computer Graphics Lecture 15.
Fast Sequences of Non-spatial State Representations in Humans
Albert V. van den Berg, Jaap A. Beintema  Neuron 
COMP60611 Fundamentals of Parallel and Distributed Systems
Volume 20, Issue 7, Pages (April 2010)
Volume 28, Issue 7, Pages e5 (April 2018)
Cengizhan Can Phoebe de Nooijer
Authoring Directed Gaze for Full-Body Motion Capture
Data-Driven Approach to Synthesizing Facial Animation Using Motion Capture Ioannis Fermanis Liu Zhaopeng
Random Neural Network Texture Model
Operational Results of LHC Collimator Alignment using Machine Learning
Presentation transcript:

Authoring Directed Gaze for Full-Body Motion Capture Roald Melssen Matej Milinkovic Tomislav Pejsa Daniel Rakita Bilge Mutlu Michael Gleicher Matej

Table of Contents Introduction Related work Implementation Evaluation Limitations & Future Work Discussion Matej

Introduction Directed Gaze - Movement of line of sight Shows focus, personality and intent Involves movement of the eyes, head and torso Usually done by hand The Paper’s Goal - automatically adding editable directed gaze to a captured motion. Matej

Related Work Gaze Synthesis: - focus on Rapid eyeball movements (Deng et al. [2005];Lee at al.[2002]), or synthesize coordinated movements of the eyes, head and body toward targets (Lence et al. [2010];Heck [2007]) Peters et al. [2010] and Pejsa et al. [2015] introduce procedural models inspired by neurophysiological observations These methods are extended Designed to provide biologically plausible movements across the parametric range Matej

Related Work Gaze Inference: - Describes methods that analytically determine when and where the character should look Research on gaze control by Henderson [2003], has shown that people’s gaze is influenced by two mechanics: The spontaneous bottom-up attention and the deliberate top-down attention The paper focuses on bottom-up attention which determines the gaze using a contrast, orientation and motion[Peters and O’Sullivan 2003; Peters and Itti 2008] In other work gaze targets are determined from spatial and kinematic properties of scene entities, such as proximity, velocity and orientation [Cafaro et al. 2009; Grillon and Thalmann 2009; Kokkinara et al. 2011] Matej

Implementation Gaze Inference Gaze Synthesis Gaze Editing Determining gaze instances from mocap data, which consist of a duration and look-at target Gaze Synthesis Adjusting motion to gaze at target Based on inverse kinematics Gaze Editing Allow user control over generated gaze instances Feeds back into step 2 after edits are done Roald

Implementation: Gaze Inference A gaze instance is a tuple: is the gaze shift start frame is the fixation start frame is the fixation end frame is the gaze target (3D vector) is the head alignment parameter is the torso alignment parameter where... Roald

Implementation: Gaze Inference Input: motion capture data Output: a list of gaze instances That means, construct a list of tuples of Algorithm: Gaze Instance Inference Determine timing: Gaze Target Inference Determine target: Computing Alignment Parameters Determine head and torso alignment: G1 G2 G3 G4 Roald

Implementation: Gaze Instance Inference Need to calculate timing parameters: Idea: analyze joint angular velocities When someone gazes at a point, the body accelerates and decelerates The maximum acceleration is attained at the midpoint of the gaze event Angular acceleration of joint j at frame f is Normalized to 0...1 range: Probability of a significant gaze event occuring at frame f is J is the set of all joints is a weight for joint j which determines how important it is Higher weight for joints closer to the eye Roald

Implementation: Gaze Instance Inference Find frame that maximizes This frame likely contains a significant gaze event (a gaze start or gaze end) Thus, the interval between two significant gaze events is classified as a gaze instance G1 G2 G3 G4 Roald significant gaze events

Implementation: Gaze Target Inference We now know the timing of the gaze instance, defined by Now we want to find the look-at target Three heuristics: Character likely looks at point along movement direction of the head Character likely looks at important objects Character likely looks at objects just before picking them up or touching them Idea: build a 2D probability distribution of the eye’s view and pick as the target with the highest probability = directional term, = importance term, = hand-contact term Roald

Implementation: Gaze Target Inference

Implementation: Computing Alignment Parameters Only parameters left to compute: Idea: Project end rotation on the arc from the start rotation to the rotation that would fully align the head to the look-at target t and then determine arc ratio is the torso rotation at the gaze start is the torso rotation at the gaze end is the rotation that would fully align the torso with the target is the rotation that corresponds to no alignment at all Projecting on the arc between and gives us Finally, we calculate

Evaluation Authoring Effort is compared by measuring two metrics: Time taken and Number of keys set The average was 25 minutes and 86 keys for each minute of animation The computation time per minute of animation was 1.5 minutes Second comparison: effort required to edit gaze animation in a scene that already contained eye movements The experienced animator used MotionBuilder and a novice animator used their tool, the novice animator took ⅓ of the time and number of operations compared to the experienced animator Matej

Evaluation Animation Quality: ( 1 ) No gaze, ( 2 ) Recorded gaze (using eye tracker), ( 3 ) Hand-authored gaze and ( 4 ) synthesized gaze (Approach in the paper) Hypothesis: ( 1 ) synthesized gaze would be preferred over no gaze; ( 2 ) synthesized gaze would be preferred over recorded gaze; and ( 3 ) synthesized gaze would be seen as non-inferior to hand-authored gaze. Design: three separate studies, each consisted of five task trials and followed a within-participants design where participants were asked to choose between videos with each gaze, presented in a randomized order Matej

Evaluation Stimuli: included 5 x 4 video clips (five scenes with each condition), 9 to 14 seconds in length Scenes included ChatWithFriends, MakeSandwich, StackBoxes, StealGem and WalkCones Measures: ( 1 ) animator competence ( 2 )realism ( 3 ) communicative clarity Matej

Evaluation Results: The data from first study support hypothesis 1 The data from second study showed that synthesized gaze was not significantly preferred over recorded gaze The third study showed that synthesized gaze was not seen as non-inferior to hand-authored gaze The scene type had a significant effect on participants choice of recorded gaze Matej

Evaluation Results: The study shows that adding eye animation using the approach from the paper leads to improvement in the perceived quality of the animation over having no eye animation at all There is a small loss in quality compared to expert-crafted animations https://vimeo.com/191892437 (3:00) Matej

Limitations & Future Work Focus on modeling only directed gaze Improve on the gaze’s accuracy and sensitivity The current implementation of the system is in Unity Integration with commonly used 3D-animation software such as MotionBuilder or Maya Matej

Questions?

Discussion Is this technique actually useful for animators, or is it one of those techniques where researchers think it is useful, but in practice it’s not? We think it’s mostly useful for novice animators or short movies (kids shows) Roald