Facial Tracking and Animation Todd Belote Bryan Harris David Brown Brad Busse.

Slides:



Advertisements
Similar presentations
Windows Movie Maker Introduction to Video Editing Mindy McAdams.
Advertisements

TouchDevelop Chapter 5-7 Chapter 5 Audio Chapter 6 Camera, Graphics and Video Chapter 7 Sensors Mengfei Ren.
KINECT REHABILITATION
Designing Facial Animation For Speaking Persian Language Hadi Rahimzadeh June 2005.
Generation of Virtual Image from Multiple View Point Image Database Haruki Kawanaka, Nobuaki Sado and Yuji Iwahori Nagoya Institute of Technology, Japan.
Move With Me S.W Graduation Project An Najah National University Engineering Faculty Computer Engineering Department Supervisor : Dr. Raed Al-Qadi Ghada.
Virtual Dart: An Augmented Reality Game on Mobile Device Supervisor: Professor Michael R. Lyu Prepared by: Lai Chung Sum Siu Ho Tung.
EE442—Multimedia Networking Jane Dong California State University, Los Angeles.
1 L45 Multimedia: Applets and Applications. 2 OBJECTIVES  How to get and display images.  To create animations from sequences of images.  To create.
Video Object Tracking and Replacement for Post TV Production LYU0303 Final Year Project Spring 2004.
Multi video camera calibration and synchronization.
Top Level System Block Diagram BSS Block Diagram Abstract In today's expanding business environment, conference call technology has become an integral.
May 10, 2004Facial Tracking and Animation Todd Belote Bryan Harris David Brown Brad Busse.
System Design and Analysis
Tracking Migratory Birds Around Large Structures Presented by: Arik Brooks and Nicholas Patrick Advisors: Dr. Huggins, Dr. Schertz, and Dr. Stewart Senior.
Detecting Image Region Duplication Using SIFT Features March 16, ICASSP 2010 Dallas, TX Xunyu Pan and Siwei Lyu Computer Science Department University.
Computing motion between images
Electrical & Computer Engineering, ECE Faculty Advisor Wayne Burleson Team Members Chinedu Okongwu Andrew Maxwell Awais Kazi Collaborators W. Richards.
Department of Computer Science and Engineering, CUHK 1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal Video Conference Archives Indexing System.
ART: Augmented Reality Table for Interactive Trading Card Game Albert H.T. Lam, Kevin C. H. Chow, Edward H. H. Yau and Michael R. Lyu Department of Computer.
CSE 291 Final Project: Adaptive Multi-Spectral Differencing Andrew Cosand UCSD CVRR.
Hand Movement Recognition By: Tokman Niv Levenbroun Guy Instructor: Todtfeld Ari.
1 Angel: Interactive Computer Graphics 4E © Addison-Wesley 2005 Models and Architectures Ed Angel Professor of Computer Science, Electrical and Computer.
1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal Video Conference Archives Indexing System Supervisor: Prof Michael Lyu Presented by: Lewis Ng,
Behavior Analysis Midterm Report Lipov Irina Ravid Dan Kotek Tommer.
Facial Tracking and Animation Project Proposal Computer System Design Spring 2004 Todd BeloteDavid Brown Brad BusseBryan Harris.
Real-Time Face Detection and Tracking Using Multiple Cameras RIT Computer Engineering Senior Design Project John RuppertJustin HnatowJared Holsopple This.
Database Construction for Speech to Lip-readable Animation Conversion Gyorgy Takacs, Attila Tihanyi, Tamas Bardi, Gergo Feldhoffer, Balint Srancsik Peter.
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
DEMONSTRATION FOR SIGMA DATA ACQUISITION MODULES Tempatron Ltd Data Measurements Division Darwin Close Reading RG2 0TB UK T : +44 (0) F :
The Project AH Computing. Functional Requirements  What the product must do!  Examples attractive welcome screen all options available as clickable.
Sana Naghipour, Saba Naghipour Mentor: Phani Chavali Advisers: Ed Richter, Prof. Arye Nehorai.
Video Data Topic 4: Multimedia Technology. What is Video? A video is just a collection of bit-mapped images that when played quickly one after another.
SENG521 (Fall SENG 521 Software Reliability & Testing Software Reliability Tools (Part 8a) Department of Electrical & Computer.
Introduction to Systems Analysis and Design Trisha Cummings.
The Camera Mouse: Visual Tracking of Body Features to Provide Computer Access for People With Severe Disabilities.
Abstract Some Examples The Eye tracker project is a research initiative to enable people, who are suffering from Amyotrophic Lateral Sclerosis (ALS), to.
REAL-TIME SOFTWARE SYSTEMS DEVELOPMENT Instructor: Dr. Hany H. Ammar Dept. of Computer Science and Electrical Engineering, WVU.
CSCI-235 Micro-Computers in Science Hardware Part II.
Acceleration Based Pedometer
1 Lecture 19: Motion Capture. 2 Techniques Morphing Motion Capture.
Prepared By: Menna Hamza Mohamed Mohamed Hesham Fadl Mona Abdel Mageed El-Koussy Yasmine Shaker Abdel Hameed Supervised By: Dr. Magda Fayek.
 Refers to sampling the gray/color level in the picture at MXN (M number of rows and N number of columns )array of points.  Once points are sampled,
3. Multimedia Systems Technology
3D SLAM for Omni-directional Camera
Department of Computer Science and Engineering, CUHK 1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal Video Conference Archives Indexing System.
Reconstructing 3D mesh from video image sequences supervisor : Mgr. Martin Samuelčik by Martin Bujňák specifications Master thesis
Advanced Computer Technology II FTV and 3DV KyungHee Univ. Master Course Kim Kyung Yong 10/10/2015.
CHAPTER TEN AUTHORING.
Department of Computer Science and Engineering, CUHK 1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal VideoConference Archives Indexing System.
Towards real-time camera based logos detection Mathieu Delalandre Laboratory of Computer Science, RFAI group, Tours city, France Osaka Prefecture Partnership.
REAL-TIME SOFTWARE SYSTEMS DEVELOPMENT Instructor: Dr. Hany H. Ammar Dept. of Computer Science and Electrical Engineering, WVU.
MIRALab Where Research means Creativity SVG Open 2005 University of Geneva 1 Converting 3D Facial Animation with Gouraud shaded SVG A method.
Tracking CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
Eurecom, 6 Feb 2007http://biobimo.eurecom.fr Project BioBiMo 1.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Introduction to Interactive Media Interactive Media Tools: Authoring Applications.
CSCI-100 Introduction to Computing Hardware Part II.
Yield Cleaning Software and Techniques OFPE Meeting
By Naveen kumar Badam. Contents INTRODUCTION ARCHITECTURE OF THE PROPOSED MODEL MODULES INVOLVED IN THE MODEL FUTURE WORKS CONCLUSION.
By: David Gelbendorf, Hila Ben-Moshe Supervisor : Alon Zvirin
Motion Detection and Processing Performance Analysis Thomas Eggers, Mark Rosenberg Department of Electrical and Systems Engineering Abstract Histograms.
April / 2010 UFOCapture 1 UFOCaptureV2 Time-Shift-Motion-Detect-Video-Recording software for complete records of un-expected events.
UFOCaptureV2 Time-Shift-Motion-Detect-Video-Recording software for complete records of un-expected events April / 2010 UFOCapture.
Digital image self-adaptive acquisition in medical x-ray imaging
The Graphics Rendering Pipeline
Categorizing sex and identity from the biological motion of faces
SyDEVS Library Building7m Tutorial
Learning complex visual concepts
Presentation transcript:

Facial Tracking and Animation Todd Belote Bryan Harris David Brown Brad Busse

Problem Background Speech driven facial animation Correlate captured facial movements to audio patterns –Capture facial movements –Analyze corresponding audio

Goals and Objectives Develop an inexpensive, robust real-time system to track facial motion and process corresponding audio. The system must: –Cost around $1000 –Run on a personal computer –Allow for long periods of data acquisition –Handle head movements –Recover from point occlusion –Output only necessary information

System Description DATA ACQUISITION FAP GENERATION POINT INITIALIZATION POINT TRACKING AUDIO PROCESSING Top level system organization –Illustrates data flow –Functional block division

Division of Work Subsystem Leads: –Data Acquisition – Todd Belote –Point Initialization – David Brown –Facial Tracking – Brad Busse & Brian Harris –FAP Generation – Brian Harris

Data Acquisition CAMERA MICROPHONE AVI MOVIE FILE EH EH = EVENT HANDLER AUDIO PROCESSING VIDEO PROCESSING FRAMEGRABBER SW TIMER CAPTURECARDCAPTURECARD DEBUG MODE Total System

Data Acquisition AVI MOVIE FILE VIDEO PROCESSING FRAMEGRABBER SW TIMER PHASE 1 BMP FILE Camera Emulation –Parses AVI movie File –Sends video frame data to Video Processing –Standalone

Data Acquisition PHASE 2 CAMERA AVI MOVIE FILE EHVIDEO PROCESSING FRAMEGRABBER SW TIMER CAPTURECARDCAPTURECARD Begin Hardware Interface –Capture and Record Camera data to AVI File

Data Acquisition PHASE 3 CAMERA AVI MOVIE FILE EHVIDEO PROCESSING FRAMEGRABBER SW TIMER CAPTURECARDCAPTURECARD Hardware to Processing –Real Capture Data to Processing –Mode Switch implemented (Emulator / Hardware)

Data Acquisition PHASE 4 CAMERA MICROPHONE AVI MOVIE FILE EH AUDIO PROCESSING VIDEO PROCESSING FRAMEGRABBER SW TIMER CAPTURECARDCAPTURECARD WAV FILE Final Implementation –Audio Capture to Processing

Point Initialization Given: Grayscale bitmap of initial frame Retrieve: Point locations and identification Identify PointsFind Points DATA AQUISITION POINT TRACKING RGB Points BOOL::DONE

Point Initialization Design Constraints –Comparison of noise to points –Point motion within one frame Process –Find point which meets minimum point criteria –Find center of point –Identify all points

Point Initialization

Point Tracking Given: a frame of visual input and the initial positions of all the points Return: a list of displacements for use in FAP generation Point TransformPoint Discovery DATA AQUISITION FAP GENERATION RGB Relative Point Location POINT INITIALIZATION Initial Point Location

Point Discovery Given a frame of visual data and the last known data point positions: –Finds new data points by searching the area around the last seen position of each old data point –Updates locations of facial parameters when possible (i.e. not missing or in conflict)

Design: Point Transform Phase 1: Facial Orientation Correction Approach: Criminisi et al. Maps any arbitrary quadrilateral onto any other This can account for all six degrees of freedom as well as perspective distortion, greatly simplifying the computation required to reorient the face When using an orientation square that encompasses most of the face, this algorithm can be made as accurate as necessary

Point Transform: Demo

Design: Point Tracking Phase 2: Data Point to Facial Parameter Conversion The rectified data points are then compared with their last known positions This will determine the displacement of the facial parameters they represent, or reassign them should the points be lost or in conflict

FAP Generation Convert pixels to centimeters Normalize coordinates Output File FAP Generator Point Tracking FAP File Resolved Point Locations FAP Points

Validation / Test – Data Acquisiton Phase 1 –SW TIMER: Verify periodicity of Timer via calls to high performance clock –Test at 1000 ms –Test at 500 ms –Test at 100 ms –Test at 50 ms –Test at 33 ms –Test at 25 ms System Validated with accuracy within 10% at 33ms –FRAME GRABBER PARSE Frames from existing AVI Movie File and Save each frame as a BMP file –Verify the number of frames corresponds to the length in the AVI Header –Determine that the Frames are the correct size –Repeat on multiple file formats to insure robustness –PHASE 1 SYSTEM TEST Display data passed to VIDEO PROCCESING as an on screen bitmap at the rates listed above for the SW TIMER Testing. –Perform similar timing testing that was performed for the SW TIMER

Validation / Test – Data Acquisiton Phase 2 –Record Test Video’s of Multiple Lengths 3 seconds 30 seconds 3 minutes –Play Test Video in Windows Media Player Determine if coloring/ video appears correct –Parse Header Information to insure proper Values Compression = BI_RGB SIZE = 320x240 Rate = 30 fps –Perform FRAMEGRABBER testing with the Test AVI files. Phase 3 –Perform PHASE 1 system test with interface set to data from file. –Run system from camera and display VIDEO PROCESSING DATA on screen as bitmap. Determine if video appears correct Run system for variable times to insure stability (with MOVIE Record Turned Off) –3 seconds –30 seconds –3 minutes –30 minutes –Test error cases Invalid file name, during from file acquisition Camera not present, data from camera

Validation / Test – Data Acquisition Phase 4 –Capture audio test files, using clock calls to verify the length of capture = length of WAV file. 3 sec 30 sec 3 minutes –Play audio test files in Windows Media Player to determine length and audio quality –Data Acquisition System Timing Test Run the system on hardware capture mode Output the Audio and Video frame timestamps as they are delivered to processing Output the corresponding time in which they are delivered Analyze the data to check for synchrony, periodicity, and evidence of time shift.

Validation/Test - Initialization Test location and identification of points on many faces Failure to complete task may imply failure and may imply design constraint –Distance from camera –Initial face orientation

Validation/Test - Tracking Test a number of different faces in a number of different poses at the limits of our specified allowances If the system accomplishes the following –Correctly extracts data points from raw visual data –Reorients the face to extract the correct displacements for every available data point The system will have passed validation

Validation/Test – FAP Generation Use FAE Engine to observe synchronization between audio and facial movements –Perform specific facial motions and validate output Eg. Move chin down, move eyebrows up, smile This test will also be used to validate entire system

Environmental and Health Considerations All hardware is off the shelf No harm from infrared light No harm from other products –Eg. Reflective markers

Social, Political and Ethical Considerations Provide low cost audiovisual capture –Increase research in field by removing cost barrier –Further advances Eg. Phone for the deaf No Ethical Issues No Political affects

Economics and Sustainability No economies of scale due to narrow scope IBM PupilCAM is hard to locate and therefore sustainability with current hardware is issue –Other cameras could provide the same function