Volker Blanz Thomas Vetter. OUTLINE Introduction Related work Database Morphable 3D Face Model Matching a morphable model to images/3D scans Building.

Slides:



Advertisements
Similar presentations
Face Recognition Sumitha Balasuriya.
Advertisements

Active Appearance Models
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Sami Romdhani Volker Blanz Thomas Vetter University of Freiburg
Face Alignment with Part-Based Modeling
Automatic Feature Extraction for Multi-view 3D Face Recognition
Face Recognition in Hyperspectral Images Z. Pan, G. Healey, M. Prasad and B. Tromberg University of California Published at IEEE Trans. on PAMI Vol 25,
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Lapped Textures Emil Praun and Adam Finkelstien (Princeton University) Huges Hoppe (Microsoft Research) SIGGRAPH 2000 Presented by Anteneh.
Visual Recognition Tutorial
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #20.
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
EE 7730 Image Segmentation.
Exchanging Faces in Images SIGGRAPH ’04 Blanz V., Scherbaum K., Vetter T., Seidel HP. Speaker: Alvin Date: 21 July 2004.
LYU0603 A Generic Real-Time Facial Expression Modelling System Supervisor: Prof. Michael R. Lyu Group Member: Cheung Ka Shun ( ) Wong Chi Kin ( )
Motion Analysis (contd.) Slides are from RPI Registration Class.
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Real-time Combined 2D+3D Active Appearance Models Jing Xiao, Simon Baker,Iain Matthew, and Takeo Kanade CVPR 2004 Presented by Pat Chan 23/11/2004.
Face Recognition Based on 3D Shape Estimation
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
The Planar-Reflective Symmetry Transform Princeton University.
Learning the Appearance of Faces: A Unifying Approach for the Analysis and Synthesis of Images. Thomas Vetter Germany University of Freiburg
PhD Thesis. Biometrics Science studying measurements and statistics of biological data Most relevant application: id. recognition 2.
Statistical Color Models (SCM) Kyungnam Kim. Contents Introduction Trivariate Gaussian model Chromaticity models –Fixed planar chromaticity models –Zhu.
Radial Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Prediction of Non-Linear Aging Trajectories of Faces
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Gwangju Institute of Science and Technology Intelligent Design and Graphics Laboratory Multi-scale tensor voting for feature extraction from unstructured.
Technology and Historical Overview. Introduction to 3d Computer Graphics  3D computer graphics is the science, study, and method of projecting a mathematical.
Machine Vision for Robots
Multimodal Interaction Dr. Mike Spann
Automatic Registration of Color Images to 3D Geometry Computer Graphics International 2009 Yunzhen Li and Kok-Lim Low School of Computing National University.
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
BACKGROUND LEARNING AND LETTER DETECTION USING TEXTURE WITH PRINCIPAL COMPONENT ANALYSIS (PCA) CIS 601 PROJECT SUMIT BASU FALL 2004.
CSE554AlignmentSlide 1 CSE 554 Lecture 5: Alignment Fall 2011.
Digital Face Replacement in Photographs CSC2530F Project Presentation By: Shahzad Malik January 28, 2003.
Integral University EC-024 Digital Image Processing.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Generalized Hough Transform
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Forward-Scan Sonar Tomographic Reconstruction PHD Filter Multiple Target Tracking Bayesian Multiple Target Tracking in Forward Scan Sonar.
Face Recognition: An Introduction
: Chapter 11: Three Dimensional Image Processing 1 Montri Karnjanadecha ac.th/~montri Image.
Categorization by Learning and Combing Object Parts B. Heisele, T. Serre, M. Pontil, T. Vetter, T. Poggio. Presented by Manish Jethwa.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Review on Graphics Basics. Outline Polygon rendering pipeline Affine transformations Projective transformations Lighting and shading From vertices to.
Multimodal Interaction Dr. Mike Spann
Data-driven Methods: Faces : Computational Photography Alexei Efros, CMU, Fall 2012 Portrait of Piotr Gibas © Joaquin Rosales Gomez.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
CS 691B Computational Photography Instructor: Gianfranco Doretto Data Driven Methods: Faces.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Statistical Models of Appearance for Computer Vision 主講人:虞台文.
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
Level Set Segmentation ~ 9.37 Ki-Chang Kwak.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
Deformation Modeling for Robust 3D Face Matching Xioguang Lu and Anil K. Jain Dept. of Computer Science & Engineering Michigan State University.
3D Graphics Rendering PPT By Ricardo Veguilla.
Recognition: Face Recognition
Dynamical Statistical Shape Priors for Level Set Based Tracking
Latent Variables, Mixture Models and EM
Computer Vision Lecture 4: Color
Presentation transcript:

Volker Blanz Thomas Vetter

OUTLINE Introduction Related work Database Morphable 3D Face Model Matching a morphable model to images/3D scans Building a morphable model Results Future work

Introduction Computer aided modeling of human faces still requires a great deal of expertise and manual control to avoid unrealistic, non-face-like results. Most limitations of automated techniques for face synthesis face animation or for general changes in the appearance of an individual face can be described as the problem of finding corresponding feature locations in different faces.

Introduction In this paper, we present a parametric face modeling technique that assists in both problems. First, arbitrary human faces can be created simultaneously controlling the likelihood of the generated faces. Second, the system is able to compute correspondence between new faces. We developed an algorithm that adjusts the model parameters automatically for an optimal reconstruction of the target, requiring only a minimum of manual initialization.

Introduction

Related work The correspondence problem between different three- dimensional face data has been addressed previously by Lee et al.[20]. First, we compute the correspondence in high resolution, considering shape and texture data simultaneously. Second, instead of using a physical tissue model to constrain the range of allowed mesh deformations, we use the statistics of our example faces to keep deformations plausible. Third, we do not rely on routines that are specifically designed to detect the features exclusively found in faces, e.g., eyes, nose.

Related work Finding the correspondence Optical flow [35] T. Vetter and V. Blanz. “Estimating coloured 3d face models from single images: An example based approach.” Bootstrapping [36] T. Vetter, M. J. Jones, and T. Poggio. “A bootstrapping algorithm for learning linear models of object classes.”

Database Laser scans of 200 heads of young adults (100 male and 100 female) were used. The laser scans provide head structure data in a cylindrical representation, with radii r(h, φ) of surface points sampled at 512 equally-spaced angles φ, and at 512 equally spaced vertical steps h. The RGB-color values R(h, φ),G(h, φ),B(h, φ), were recorded in the same spatial resolution and were stored in a texture map with 8 bit per channel. The resultant faces were represented by approximately 70,000 vertices and the same number of color values.

Morphable 3D face model We represent the geometry of a face with a shape- vector S=(X 1,Y 1,Z 1,X 2,Y 2,…,X n,Y n,Z n ) T,that contains the X,Y,Z-coordinates of its n vertices. We represent the texture of a face by a texture- vector T =(R 1,G 1,B 1,R 2,G 2,…,R n,G n,B n ) T, that contains the R,G,B color values of the n corresponding vertices. Since we assume all faces in full correspondence, new shapes S model and new textures T model can be expressed as a linear combination of the shapes and textures of the m exemplar faces.

Morphable 3D face model We fit a multivariate normal distribution to our data set of 200 faces, based on the averages of shape and texture and the covariance matrices C s and C t. A common technique for data compression known as Principal Component Analysis (PCA) performs a basis transformation to an orthogonal coordinate system formed by the eigenvectors s i and t i of the covariance matrices.

Morphable 3D face model Segmented morphable model: The morphable model described in equation (1), has m-1 degrees of freedom for texture and shape. The expressiveness of the model can be increased by dividing faces into independent sub-regions that are morphed independently, for example into eyes, nose, mouth and a surrounding region. A complete 3D face is generated by blending them at the borders

Morphable 3D face model Facial attributes Some facial attributes can easily be related to biophysical measurements, such as the width of the mouth, facial femininity or being more or less bony can hardly be described by numbers. Unlike facial expressions, attributes that are invariant for each individual are more difficult to isolate. The following method allows us to model facial attributes such as gender, fullness of faces, darkness of eyebrows, double chins, and hooked versus concave noses

Morphable 3D face model Based on a set of faces (S i,T i ) with manually assigned labels μ i describing the markedness of the attribute, we compute weighted sums Multiples of (ΔS, ΔT) can now be added to or subtracted from any individual face.

Matching a morphable model to images Coefficients of the 3D model are optimized along with a set of rendering parameters such that they produce an image as close as possible to the input image. It starts with the average head and with rendering parameters roughly estimated by the user.

Matching a morphable model to images Model Parameters: Facial shape and texture are defined by coefficients α j and β j,j = 1,2,…,m-1. Rendering parameters contain camera position (azimuth and elevation), object scale, image plane rotation and translation, intensity i r,amb,i g,amb,i b,amb of ambient light, and intensity i r,dir,i g,dir,i b,dir of directed light

Matching a morphable model to images From parameters, colored images are rendered using perspective projection and the Phong illumination model. The reconstructed image is supposed to be closest to the input image in terms of Euclidean distance

In terms of Bayes decision theory, the problem is to find the set of parameters with maximum posterior probability, given an image I input. While and rendering parameters completely determine the predicted image I model, the observed image I input may vary due to noise. For Gaussian noise with a standard deviation, the likelihood to observe I input is Maximum posterior probability is then achieved by minimizing the cost function Matching a morphable model to images

The optimization algorithm described below uses an estimate of E based on a random selection of surface points. In the center of triangle k, texture (R k,G k,B k ) T and 3D location (X k,Y k,Z k ) T are averages of the values at the corners. Perspective projection maps these points to image locations (p x,k,p y,k ) T. According to Phong illumination, the color components I r,model,I g,model and I b,model take the form

Matching a morphable model to images For high resolution 3D meshes, variations in I model across each triangle are small, so E I may be approximated by where ak is the image area covered by triangle k. In gradient descent, contributions from different triangles of the mesh would be redundant. In each iteration, we therefore select a random subset of 40 triangles k and replace E I by

Parameters are updated depending on analytical derivatives of the cost function E, using and similarly for and, with suitable factors. Coarse-to-Fine: In order to avoid local minima, the algorithm follows a coarse-to-fine strategy in several respects: a) The first set of iterations is performed on a down- sampled version of the input image with a low resolution morphable model. Matching a morphable model to images

b) We start by optimizing only the first coefficients and controlling the first principal components, along with all parameters. In subsequent iterations, more and more principal components are added. c) Starting with a relatively large, which puts a strong weight on prior probability, is later reduced to obtain maximum matching quality. d) In the last iterations, the face model is broken down into segments. Multiple Images Illumination-Corrected Texture Extraction Matching a morphable model to images

Matching a morphable model to 3D scans Analogous to images, where perspective projection, the laser scans provide a two-dimensional cylindrical parameterization of the surface by means of a mapping =>. Hence, the matching algorithm from the previous section now determines and minimizing

Building a morphable model In this section, we describe how to build the morphable model from a set of unregistered 3D prototypes, and to add a new face to the existing morphable model, increasing its dimensionality. Section 4.1 finds the best match of a given face only within the range of the morphable model, it cannot add new dimensions to the vector space of faces. We use an optic flow algorithm that computes correspondence between two faces without the need of a morphable model.

3D Correspondence using Optic Flow Initially designed to find corresponding points in grey- level images, a gradient-based optic flow algorithm is modified to establish correspondence between a pair of 3D scans, taking into account color and radius values simultaneously. The algorithm computes a flow field that minimizes differences of that weights variations in texture and shape equally. Building a morphable model

Results We tested the expressive power of our morphable model by automatically reconstructing 3D faces from photographs of arbitrary Caucasian faces of middle age that were not in the database. The whole matching procedure was performed in 10 5 iterations. On an SGI R10000 processor,computation time was 50 minutes.

Results If previously unseen backgrounds become visible, we fill the holes with neighboring background pixels (Fig. 6). The face can be combined with other 3D graphic objects, such as glasses or hats, and then be rendered in front of the background, computing cast shadows or new illumination conditions (Fig. 7).

Future work Issues of implementation: We plan to speed up our matching algorithm by implementing a simplified Newton- method for minimizing the cost function (Equation 5). Extending the database: While the current database is sufficient to model Caucasian faces of middle age, we would like to extend it to children, to elderly people as well as to other races. Extending the face model: Our current morphable model is restricted to the face area, because a sufficient 3D model of hair cannot be obtained with our laser scanner.