IRCDL'07 - Padova, January 2007 3D Face Recognition using iso-Geodesic Surfaces Stefano Berretti, Alberto Del Bimbo, Pietro Pala Dipartimento di Sistemi.

Slides:



Advertisements
Similar presentations
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Advertisements

November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
ECE738 Advanced Image Processing Face Recognition by Elastic Bunch Graph Matching IEEE Trans. PAMI, July 1997.
Face Description with Local Binary Patterns:
Face Alignment with Part-Based Modeling
3D Shape Histograms for Similarity Search and Classification in Spatial Databases. Mihael Ankerst,Gabi Kastenmuller, Hans-Peter-Kriegel,Thomas Seidl Univ.
Fırat Onur Alsaran, Neşe Alyüz, Melike Esma İlter, Mürsel Taşgın, Ahmet Burak Yoldemir CmpE 58Z Term Project.
Automatic 3D Face Recognition System 邱彥霖 龔士傑 Biometric Authentication.
Automatic Feature Extraction for Multi-view 3D Face Recognition
Contactless and Pose Invariant Biometric Identification Using Hand Surface Vivek Kanhangad, Ajay Kumar, Senior Member, IEEE, and David Zhang, Fellow, IEEE.
Semi-Supervised Hierarchical Models for 3D Human Pose Reconstruction Atul Kanaujia, CBIM, Rutgers Cristian Sminchisescu, TTI-C Dimitris Metaxas,CBIM, Rutgers.
Amir Hosein Omidvarnia Spring 2007 Principles of 3D Face Recognition.
3D Face Modelling and Recognition Survey Timur Aksoy Biometrics Course Fall 2011 Sabanci University.
Image Indexing and Retrieval using Moment Invariants Imran Ahmad School of Computer Science University of Windsor – Canada.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Reverse Engineering Niloy J. Mitra.
Integrating Color And Spatial Information for CBIR NTUT CSIE D.W. Lin
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
A Study of Approaches for Object Recognition
CS CS 175 – Week 2 Processing Point Clouds Registration.
Face Recognition Based on 3D Shape Estimation
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Comparison and Combination of Ear and Face Images in Appearance-Based Biometrics IEEE Trans on PAMI, VOL. 25, NO.9, 2003 Kyong Chang, Kevin W. Bowyer,
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Oral Defense by Sunny Tang 15 Aug 2003
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
3D Models and Matching representations for 3D object models
SVD(Singular Value Decomposition) and Its Applications
Introduction --Classification Shape ContourRegion Structural Syntactic Graph Tree Model-driven Data-driven Perimeter Compactness Eccentricity.
Gender and 3D Facial Symmetry: What’s the Relationship ? Xia BAIQIANG (University Lille1/LIFL) Boulbaba Ben Amor (TELECOM Lille1/LIFL) Hassen Drira (TELECOM.
CSE 185 Introduction to Computer Vision
Computer vision.
CSE554AlignmentSlide 1 CSE 554 Lecture 8: Alignment Fall 2014.
CSE554Laplacian DeformationSlide 1 CSE 554 Lecture 8: Laplacian Deformation Fall 2012.
V. Space Curves Types of curves Explicit Implicit Parametric.
CSE554AlignmentSlide 1 CSE 554 Lecture 5: Alignment Fall 2011.
MESA LAB Multi-view image stitching Guimei Zhang MESA LAB MESA (Mechatronics, Embedded Systems and Automation) LAB School of Engineering, University of.
Expression-invariant Face Recognition using Geodesic Distance Isometries Kerry Widder A Review of ‘Robust expression-invariant face recognition from partially.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
Intelligent Vision Systems ENT 496 Object Shape Identification and Representation Hema C.R. Lecture 7.
A 3D Model Alignment and Retrieval System Ding-Yun Chen and Ming Ouhyoung.
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
Introduction --Classification Shape ContourRegion Structural Syntactic Graph Tree Model-driven Data-driven Perimeter Compactness Eccentricity.
Wenqi Zhu 3D Reconstruction From Multiple Views Based on Scale-Invariant Feature Transform.
CSE554AlignmentSlide 1 CSE 554 Lecture 8: Alignment Fall 2013.
Adaptive Rigid Multi-region Selection for 3D face recognition K. Chang, K. Bowyer, P. Flynn Paper presentation Kin-chung (Ryan) Wong 2006/7/27.
Event retrieval in large video collections with circulant temporal encoding CVPR 2013 Oral.
Using simplified meshes for crude registration of two partially overlapping range images Mercedes R.G.Márquez Wu Shin-Ting State University of Matogrosso.
Timo Ahonen, Abdenour Hadid, and Matti Pietikainen
Face Detection Using Neural Network By Kamaljeet Verma ( ) Akshay Ukey ( )
A Part-aware Surface Metric for Shape Analysis Rong Liu 1, Hao Zhang 1, Ariel Shamir 2, and Daniel Cohen-Or 3 1 Simon Fraser University, Canada 2 The Interdisciplinary.
1 A Methodology for automatic retrieval of similarly shaped machinable components Mark Ascher - Dept of ECE.
Mobile Robot Localization and Mapping Using Range Sensor Data Dr. Joel Burdick, Dr. Stergios Roumeliotis, Samuel Pfister, Kristo Kriechbaum.
Face Detection & Recognition
Deformation Modeling for Robust 3D Face Matching Xioguang Lu and Anil K. Jain Dept. of Computer Science & Engineering Michigan State University.
Active Flattening of Curved Document Images via Two Structured Beams
CSE 554 Lecture 8: Alignment
Spectral Methods for Dimensionality
Guillaume-Alexandre Bilodeau
Recognizing Deformable Shapes
PRESENTED BY Yang Jiao Timo Ahonen, Matti Pietikainen
Fast Preprocessing for Robust Face Sketch Synthesis
Object Recognition in the Dynamic Link Architecture
Unsupervised Face Alignment by Robust Nonrigid Mapping
3D Models and Matching particular matching techniques
Computer Vision Lecture 16: Texture II
Scale-Space Representation for Matching of 3D Models
Presentation transcript:

IRCDL'07 - Padova, January D Face Recognition using iso-Geodesic Surfaces Stefano Berretti, Alberto Del Bimbo, Pietro Pala Dipartimento di Sistemi e Informatica University of Firenze, Firenze, ITALY

Motivation  Detection and identification of human faces have been largely addressed mainly focussing on 2D still images and videos.  Recently, the increasing availability of three-dimensional (3D) data, has renewed the interest in 3D face models to improve the effectiveness of face recognition systems.  Solutions based on 3D face models are expected to feature less sensitivity to lighting conditions and pose variations.  3D face models are acquired through laser scanners, structured ligth scanners or multi camera systems, in the form of range images, cloud of points or polygonal meshes.  Usually texture data is not available.

3D face recognition approaches  2D to 2D (using deformations of a 3D model):  The input is 2D image; the database contains 2D images. Recognition is performed by comparing the deformations of a 3D face model over the database images and the query image (using morphing) [Blanz, TPAMI03].  3D to 2D:  3D to 2D transformation. Techniques for 2D image matching (e.g., PCA, Gabor filters, wavelets, etc.) are used to perform recognition in 2D [Pan, CVPR’05], [Wang, CVPR06].  Model based solutions. A 3D annotated face model is deformed to fit a 3D query model. The deformed model is projected onto a 2D image. Recognition is performed by comparing 2D projected images [Jain, CVPR06], [Kakadiaris, BMVC06].  2D+3D:  3D face queries are matched with 3D face models; 2D face images are matched with 2D face image models, separately. Results of the two matchings are combined to achieve better recognition accuracy [Beumier, PRL01], [Wang, IVC05].  3D to 3D (point clouds):  3D face rigid surfaces acquired as point clouds are compared using the Iterative Closest Point (ICP) algorithm [Mian, BMVC05], [Chang, TPAMI06].   Current research challenges:  Recognition in the presence of expression variations.  Matching efficiency.

Proposed solution  We propose a 3D to 3D face recognition approach in which the structural information of a 3D face model, given as a mesh, is directly captured in the 3D space by modeling the basic 3D shape and the relative arrangement of iso-geodesic stripes identified on the model surface.  Three steps:  Extraction of iso-geodesic stripes according to the geodesic distances from fiducial points.  Modeling of the shape and 3D spatial relationships between stripes (surfaces in the 3D space).  Construction and matching of a graph based face model.

Geodesic distance from fiducial points  Iso-geodesic stripes are defined with reference to a function f defined on the face surface.  Value of f at a generic vertex v is defined as the geodesic distance  (v,u) between v and a fiducial vertex u (the nose tip).  Due to the polygonal approximation of the model surface, the geodesic distance  is computed as the shortest path on the edges of the mesh by using the Dijkstra’s algorithm.  For regular meshes, this provides a fair approximation of the real geodesic distance.  Normalized values of the distance are obtained dividing the geodesic distance by the Euclidean eye-to-nose distance.  This guarantees invariance w.r.t. scaling and pose of the face.  Since the eye-to-nose distance is invariant to face expressions, this normalization does not bias values of  under expression changes. Euclidean distance Geodesic distance v u

 The nose tip and the eyes cavities are identified using mean and Gaussian curvatures according to state of the art solutions*.  Models are aligned with the nose tip pointing along the z axis.  Distances from the nose tip are quantized into n intervals c 1,..., c n. Accordingly, n level set stripes are identified on the model surface, the i-th stripe corresponding to the set of surface points on which the value of the distance falls in the limits of interval c i.  Stripes are surfaces in the 3D space.  Information captured by the stripes is modeled by their 3D mutual spatial relationships using the weighted walkthroughs approach. Stripes extraction (*) K.I. Chang, K.W. Bowyer, P.J. Flynn. “Multiple Nose Region Matching for 3D Face Recognition Under Varying Facial Expression,” IEEE Trans. on Pattern Analysis and Machine Intelligence, October 2006.

 In a 3D reference system, with coordinate axes X, Y, Z, the projections of two points a = and b = on each axis, can take three different orders: before, coincident, after.  The combination of the three projections results in 27 different 3D displacements which can be encoded by a triple of indexes :  In general, points of 3D extended sets A, B can be connected through multiple different displacements.  The triple, is a walkthrough from A to B if it encodes the displacement between at least one pair of points a  A, b  B. Modeling 3D spatial relationships x a b z y zaza xaxa yaya zbzb xbxb ybyb i=+1 k=+1 j=-1 z y j A B x k i

3D weighted walkthroughs (WW)  Relevance of each walkthrough is measured by a weight w i,j,k (A,B) evaluating the number of pairs of points a  A, b  B, whose displacement is captured by the direction.  Formally, weights w i,j,k (A,B) are evaluated as integral measures over the six-dimensional set of point pairs in A and B: K ijk acts as dimensional normalization factor; C ±1 (.) are the characteristic functions of the positive and negative real semi-axis (0,+oo) and (-oo,0), respectively. C 0 (.)= δ(.) denotes the Dirac's function and acts as the characteristic function of the singleton set {0}.  Weights between A and B are organized in a 3x3x3 matrix w(A,B), of indexes i, j, k. W 011 W -111 W 111 W 001 W -101 W 101 W 0-11 W W j i W 010 W -110 W 110 W 100 W 1-10 W 01-1 W W 11-1 W 10-1 W k0k w(A,B)

Compositional computation  Integral properties ensure that the 27 weights are reflexive (w i,j,k (A,B) = w -i,-j,-k (B,A)), and invariant with respect to shifting and scaling. In addition, they are compositional:  According to this, any complex entity is decomposed in a set of 3D rectangular voxels of a uniform partitioning of the space.  The integral computation is cast to the linear combination of a set of sub-integrals computed in closed form on rectangular domains arranged in a fixed number of possible different spatial positions w k=1 (a n,b m ) = ¼ ½ ¼ w k=0 (a n,b m ) = ½ 1 ½ w k=-1 (a n,b m ) = ¼ ½ ¼ z y anan x bmbm anan bmbm

WW between surfaces (stripes) in 3D  Computation of 3D spatial relationships between surfaces directly descends from the general case.  Volumetric integrals become surface integrals extended to the points of two surfaces.  Numerical computation of the surface integrals is avoided by exploiting the property of compositionality, and regarding surface entites as volumetric entities with infinitesimal thickness.  Linear combination of integrals computed in closed form between elementary rectangular voxels of a uniform partitioning of the space.

 Given the w(A,B) matrix, measuring the relationship between entities A and B, three directional weights, taking values in [0,1], are computed on the eight corner weights:  These account for the degree by which B is on the right ( ), up ( ) and in front of A ( ).  Similarly, three weights account for the alignment along each of the three reference directions of the space:  Finally, three weights account for the joint alignment along two directions of the space: Directional and alignment weights w(A,B) 3x3x3 matrix w -111 w 111 w w 1-11 w -111 w 111 w w 11-1 w 111 w 1-11 w 11-1 w 1-1-1

Distance between WW  Dis-similarity in the arrangement of pairs of entities and captured by matrices w(A,B)=w and w(A’,B’)=w’, is evaluated by a distance D(w,w') which combines the differences between homologous coefficients in the space of 27-tuples of WW.  Using the directional and alignment weights, this is expressed as: where are non negative numbers with sum equal to 1.

Graph representation  WW are computed for every pair of stripes, including the pair composed by a stripe and itself.  A face with N stripes is represented by N*(N+1)/2 relationship matrixes. This model is cast to a graph representation by regarding face stripes as graph nodes and their mutual spatial relationships as graph edges:   assigns to a node the WW computed between the node and itself;  associates to every edge the WW computed between the two nodes the edge connects to.

Efficient matching  Due to the normalization of geodesic distances, homologous stripes in different models capture the same distance interval.  Face matching is reduced to the comparison between homologous iso-geodesic stripes and their relationships: An interpretation Γ associates homologous nodes in the two representations, that is Γ(p k ) = r k, k=1,...., N, being p k and r k nodes of the template and reference face graphs: t1t1 t2t2 t3t3 t4t4 t5t5 t6t6 t7t7 r1r1 r2r2 r3r3 r4r4 r5r5 r6r6 r7r7

3D face database  Face recognition experiments have been conducted on the 3D facial models of the GavabDB database (publicly available at  The database includes 61 people (45 male and 16 female).  They are Caucasian and most of them are aged between 18 and 40.  For each person, 7 different models are taken differing in pose or facial expression, resulting in 427 facial models.  2 frontal and 2 rotated models with neutral facial expression.  3 frontal models in which the person laughs, smiles or exhibits a random gesture.

Recognition: neutral expression  In the recognition experiments, for each person one of the two frontal models with neutral expression is used as reference (gallery) model.  A gallery of 61 neutral models is obtained.  The remaining 366 models are used as templates for the recognition (probes).  A stripe extent of 0.08 is used in the experiments. Dynamic of the distance using the first 8 stripes #12#8#4

Recognition: non-neutral expression  The proposed approach (IGS) provides a relatively high recognition accuracy also for variations in face expression.  This can be intuitively motivated by the fact that the geodesic distance between vertices of the mesh is almost invariant to moderate facial expression changes.  Large variability inside the given categories of face expressions.  In the “random” category, each subject is free to provide a generic facial expression.

Baseline comparison  Baseline comparison against the Iterative Closest Point (ICP) registration algorithm is performed.  ICP monotonically converges to a local minima with respect to the mean-square distance objective function computed between vertices of two models.  To avoid local minima in the ICP registration, the algorithm is initialized with the correspondence between fiducial points of the two models.

Current and future work  Facial stripes computed with respect to different/multiple fiducial points.  Improvement of recognition accuracy w.r.t. expression variations.  Experiments on larger and heterogeneous face databases.  Florida State University database (115 persons with 6 different expressions each).  The Face Recognition Grand Challenge database - FRGC v.2 (4007 3D scans of 466 persons, includes expression changes).  The BJUT-3D large scale chinese face database (500 models, neutral expression). Useful to study ethnic variation.  Comparison w.r.t. different 3D to 3D recognition approaches on common benchmark databases.