Université du Québec École de technologie supérieure Face Recognition in Video Using What- and-Where Fusion Neural Network Mamoudou Barry and Eric Granger.

Slides:



Advertisements
Similar presentations
Real-Time Detection, Alignment and Recognition of Human Faces
Advertisements

TWO STEP EQUATIONS 1. SOLVE FOR X 2. DO THE ADDITION STEP FIRST
University of Karlsruhe September 30th, 2004 Masayuki Fujita
You have been given a mission and a code. Use the code to complete the mission and you will save the world from obliteration…
Video Surveillance E Senior/Feris/Tian 1 Emerging Topics in Video Surveillance Rogerio Feris IBM TJ Watson Research Center
Algorithms compared Bicubic Interpolation Mitra's Directional Filter
1 Copyright © 2010, Elsevier Inc. All rights Reserved Fig 2.1 Chapter 2.
By D. Fisher Geometric Transformations. Reflection, Rotation, or Translation 1.
Business Transaction Management Software for Application Coordination 1 Business Processes and Coordination.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
0 - 0.
DIVIDING INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
MULTIPLYING MONOMIALS TIMES POLYNOMIALS (DISTRIBUTIVE PROPERTY)
SUBTRACTING INTEGERS 1. CHANGE THE SUBTRACTION SIGN TO ADDITION
MULT. INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
Addition Facts
ZMQS ZMQS
Photo Composition Study Guide Label each photo with the category that applies to that image.
BT Wholesale October Creating your own telephone network WHOLESALE CALLS LINE ASSOCIATED.
Real-Time Projector Tracking on Complex Geometry Using Ordinary Imagery Tyler Johnson and Henry Fuchs University of North Carolina – Chapel Hill ProCams.
LOGO Regression Analysis Lecturer: Dr. Bo Yuan
1 Quality of Service Issues Network design and security Lecture 12.
ABC Technology Project
Matthias Wimmer, Bernd Radig, Michael Beetz Chair for Image Understanding Computer Science TU München, Germany A Person and Context.
1 Human Gesture Recognition by Mohamed Bécha Kaâniche 11/02/2009.
Active Appearance Models
Computer vision: models, learning and inference
IPIM, IST, José Bioucas, Convolution Operators Spectral Representation Bandlimited Signals/Systems Inverse Operator Null and Range Spaces Sampling,
1 Evaluations in information retrieval. 2 Evaluations in information retrieval: summary The following gives an overview of approaches that are applied.
Squares and Square Root WALK. Solve each problem REVIEW:
Vogler and Metaxas University of Toronto Computer Science CSC 2528: Handshapes and Movements: Multiple- channel ASL recognition Christian Vogler and Dimitris.
Machine Learning: Intro and Supervised Classification
Chapter 5 Test Review Sections 5-1 through 5-4.
Addition 1’s to 20.
Probabilistic Tracking and Recognition of Non-rigid Hand Motion
25 seconds left…...
Princess Nora University Artificial Intelligence Artificial Neural Network (ANN) 1.
Test B, 100 Subtraction Facts
Week 1.
Group Meeting Presented by Wyman 10/14/2006
We will resume in: 25 Minutes.
1 Unit 1 Kinematics Chapter 1 Day
Does one size really fit all? Evaluating classifiers in Bag-of-Visual-Words classification Christian Hentschel, Harald Sack Hasso Plattner Institute.
People Counting and Human Detection in a Challenging Situation Ya-Li Hou and Grantham K. H. Pang IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART.
Eigenfaces for Recognition Presented by: Santosh Bhusal.
Tracking Learning Detection
Robust 3D Head Pose Classification using Wavelets by Mukesh C. Motwani Dr. Frederick C. Harris, Jr., Thesis Advisor December 5 th, 2002 A thesis submitted.
Vision Based Control Motion Matt Baker Kevin VanDyke.
Robust Object Tracking via Sparsity-based Collaborative Model
A Study of Approaches for Object Recognition
4EyesFace-Realtime face detection, tracking, alignment and recognition Changbo Hu, Rogerio Feris and Matthew Turk.
Particle Filtering for Non- Linear/Non-Gaussian System Bohyung Han
Overview and Mathematics Bjoern Griesbach
West Virginia University
Spatio-temporal constraints for recognizing 3D objects in videos Nicoletta Noceti Università degli Studi di Genova.
MURI: Integrated Fusion, Performance Prediction, and Sensor Management for Automatic Target Exploitation 1 Dynamic Sensor Resource Management for ATE MURI.
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
Sparse Bayesian Learning for Efficient Visual Tracking O. Williams, A. Blake & R. Cipolloa PAMI, Aug Presented by Yuting Qi Machine Learning Reading.
Max-Confidence Boosting With Uncertainty for Visual tracking WEN GUO, LIANGLIANG CAO, TONY X. HAN, SHUICHENG YAN AND CHANGSHENG XU IEEE TRANSACTIONS ON.
Guillaume-Alexandre Bilodeau
Recognizing Deformable Shapes
Tracking Objects with Dynamics
Recovery from Occlusion in Deep Feature Space for Face Recognition
An Improved Neural Network Algorithm for Classifying the Transmission Line Faults Slavko Vasilic Dr Mladen Kezunovic Texas A&M University.
Department of Electrical Engineering
Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen, Zne-Jung Lee
Introduction to Object Tracking
Presentation transcript:

Université du Québec École de technologie supérieure Face Recognition in Video Using What- and-Where Fusion Neural Network Mamoudou Barry and Eric Granger Laboratoire dimagerie, de vision et dintelligence artificielle École de technologie supérieure Montreal, Canada

Université du Québec École de technologie supérieure 2 Overview 1. Introduction 2. What-and-Where fusion neural network 3. Experimental methodology 4. Results 5. Conclusion

Université du Québec École de technologie supérieure 3 1. Introduction Challenges of video-based face recognition low quality and resolution of frames. low quality and resolution of frames. uncontrolled environments: variation in poses, orientation, expressions, illumination, occlusion, etc. uncontrolled environments: variation in poses, orientation, expressions, illumination, occlusion, etc.

Université du Québec École de technologie supérieure 4 1. Introduction General system for face recognition in video

Université du Québec École de technologie supérieure 5 1. Introduction State of the art 1. Methods based on static images –exploit quality metric, and recognize only high quality ROIs 2. Spatiotemporal approaches –track faces in the environment, and recognize individuals over several samples

Université du Québec École de technologie supérieure 6 1. Introduction Objectives Observe the effectiveness of the What-and-Where fusion neural network in video-based face recognition Observe the effectiveness of the What-and-Where fusion neural network in video-based face recognition Robust operation in uncontrolled environments Robust operation in uncontrolled environments

Université du Québec École de technologie supérieure 7 2. What-and-Where Fusion Neural Network (Granger et al., 2001) Division of data streams 1. What data : intrinsic properties of a face (to classifier) 2. Where data : contextual information (to tracker)

Université du Québec École de technologie supérieure 8 Tracker: bank of Kalman filters stimates the future position of faces in a scene estimates the future position of faces in a scene Classifier: fuzzy ARTMAP classifies faces detected in a scene classifies faces detected in a scene neural network architecture capable of fast, stable, online, unsupervised or supervised, incremental learning, classification and prediction neural network architecture capable of fast, stable, online, unsupervised or supervised, incremental learning, classification and prediction 2. What-and-Where Fusion Neural Network

Université du Québec École de technologie supérieure 9 2. What-and-Where fusion neural network Evidence accumulation

Université du Québec École de technologie supérieure 10 Sequential evidence accumulation Fusion of responses from classifier and tracker 1. accumulation rule: 2. prediction of the recognition system: 2. What-and-Where Fusion Neural Network

Université du Québec École de technologie supérieure Experimental methodology Data set (D. Gorodnichy, CNRC, 2005) Video-based framework for face recognition in video Task: recognize the user of a PC 11 individuals: 11 individuals: 2 video sequences per individual, one dedicated for training and the other for testing

Université du Québec École de technologie supérieure Experimental methodology Data set different scenarios : pose, expression, orientation, motion, proximity, resolution and partial occlusion.

Université du Québec École de technologie supérieure Experimental methodology Protocol for experiments train: train fuzzy ARTMAP with What data, using two training strategies train: train fuzzy ARTMAP with What data, using two training strategies Hold-Out Validation (HV) Hold-Out Validation (HV) Particle Swarm Optimization (PSO) to optimize hyper- parameters (Granger et al., 2007) Particle Swarm Optimization (PSO) to optimize hyper- parameters (Granger et al., 2007) test : classify What data with fuzzy ARTMAP and track Where data with Kalman filters test : classify What data with fuzzy ARTMAP and track Where data with Kalman filters

Université du Québec École de technologie supérieure Experimental methodology Performance measures accuracy: average classification error (estimate of generalization error) accuracy: average classification error (estimate of generalization error) resource requirements: resource requirements: compression: average number of training patterns per category compression: average number of training patterns per category convergence time: average number of epochs required to complete learning. convergence time: average number of epochs required to complete learning.

Université du Québec École de technologie supérieure Results Examples of Face Detections

Université du Québec École de technologie supérieure Results Average error and compression vs. ROI scaling size (with 100% of training data)

Université du Québec École de technologie supérieure Results Average error and compression vs. training subset size (with a |ROI| =10x10)

Université du Québec École de technologie supérieure Results Average convergence time fuzzy ARTMAP with HV: ~1 epoch fuzzy ARTMAP with HV: ~1 epoch fuzzy ARTMAP with PSO: ~543 epochs fuzzy ARTMAP with PSO: ~543 epochs (60 particles x ~8.9 iterations x 1 epoch)

Université du Québec École de technologie supérieure Results Average confusion matrix

Université du Québec École de technologie supérieure 20 Example of prediction errors over time 4. Results

Université du Québec École de technologie supérieure 21 Effectiveness of the What-and-Where fusion neural network in improving the accuracy on complex video data (about 50% over fuzzy ARTMAP alone, and k-NN). Effectiveness of the What-and-Where fusion neural network in improving the accuracy on complex video data (about 50% over fuzzy ARTMAP alone, and k-NN). The system is less sensitive to noise: attenuation of fuzzy ARTMAP poor predictions. The system is less sensitive to noise: attenuation of fuzzy ARTMAP poor predictions. Optimizing the network internal parameters using PSO learning strategy improves the accuracy of the system. Optimizing the network internal parameters using PSO learning strategy improves the accuracy of the system. Fuzzy ARTMAP yields a higher compression than k-NN: suitable for real time and ressource limited applications. Fuzzy ARTMAP yields a higher compression than k-NN: suitable for real time and ressource limited applications. 5. Conclusion

Université du Québec École de technologie supérieure Future work Explore different ARTMAP models to improve the classification rate. Explore different ARTMAP models to improve the classification rate. Explore other representations (features) of face based on biological vision perception. Explore other representations (features) of face based on biological vision perception. Investigate for more robust tracking algorithms such as Extended Kalman filter, Particle filters, etc., for non linear tracking. Investigate for more robust tracking algorithms such as Extended Kalman filter, Particle filters, etc., for non linear tracking.