Using spatio-temporal probabilistic framework for object tracking By: Guy Koren-Blumstein Supervisor: Dr. Hayit Greenspan Emphasis on Face Detection &

Slides:



Advertisements
Similar presentations
Bayesian Decision Theory Case Studies
Advertisements

1 Approximated tracking of multiple non-rigid objects using adaptive quantization and resampling techniques. J. M. Sotoca 1, F.J. Ferri 1, J. Gutierrez.
Segmentation via Maximum Entropy Model. Goals Is it possible to learn the segmentation problem automatically? Using a model which is frequently used in.
Introduction To Tracking
Activity Recognition Aneeq Zia. Agenda What is activity recognition Typical methods used for action recognition “Evaluation of local spatio-temporal features.
Probabilistic Robotics
Automatic Identification of Bacterial Types using Statistical Image Modeling Sigal Trattner, Dr. Hayit Greenspan, Prof. Shimon Abboud Department of Biomedical.
Multiple People Detection and Tracking with Occlusion Presenter: Feifei Huo Supervisor: Dr. Emile A. Hendriks Dr. A. H. J. Stijn Oomes Information and.
Texture Segmentation Based on Voting of Blocks, Bayesian Flooding and Region Merging C. Panagiotakis (1), I. Grinias (2) and G. Tziritas (3)
Computer Vision Optical Flow
Formation et Analyse d’Images Session 8
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
A New Block Based Motion Estimation with True Region Motion Field Jozef Huska & Peter Kulla EUROCON 2007 The International Conference on “Computer as a.
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
Unsupervised Image Clustering using Probabilistic Continuous Models and Information Theoretic Principles Shiri Gordon Electrical Engineering – System,
Rodent Behavior Analysis Tom Henderson Vision Based Behavior Analysis Universitaet Karlsruhe (TH) 12 November /9.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Incremental Learning of Temporally-Coherent Gaussian Mixture Models Ognjen Arandjelović, Roberto Cipolla Engineering Department, University of Cambridge.
Segmentation by Clustering Reading: Chapter 14 (skip 14.5) Data reduction - obtain a compact representation for interesting image data in terms of a set.
1 Integration of Background Modeling and Object Tracking Yu-Ting Chen, Chu-Song Chen, Yi-Ping Hung IEEE ICME, 2006.
Presented by Zeehasham Rasheed
A Probabilistic Framework For Segmentation And Tracking Of Multiple Non Rigid Objects For Video Surveillance Aleksandar Ivanovic, Tomas S. Huang ICIP 2004.
MULTIPLE MOVING OBJECTS TRACKING FOR VIDEO SURVEILLANCE SYSTEMS.
A Probabilistic Framework for Video Representation Arnaldo Mayer, Hayit Greenspan Dept. of Biomedical Engineering Faculty of Engineering Tel-Aviv University,
Particle Filtering for Non- Linear/Non-Gaussian System Bohyung Han
Presented by Pat Chan Pik Wah 28/04/2005 Qualifying Examination
Learning and Recognizing Activities in Streams of Video Dinesh Govindaraju.
1 Video Surveillance systems for Traffic Monitoring Simeon Indupalli.
Feature and object tracking algorithms for video tracking Student: Oren Shevach Instructor: Arie nakhmani.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Mean-shift and its application for object tracking
BraMBLe: The Bayesian Multiple-BLob Tracker By Michael Isard and John MacCormick Presented by Kristin Branson CSE 252C, Fall 2003.
TP15 - Tracking Computer Vision, FCUP, 2013 Miguel Coimbra Slides by Prof. Kristen Grauman.
Learning and Recognizing Human Dynamics in Video Sequences Christoph Bregler Alvina Goh Reading group: 07/06/06.
S EGMENTATION FOR H ANDWRITTEN D OCUMENTS Omar Alaql Fab. 20, 2014.
Computer Vision, Robert Pless Lecture 11 our goal is to understand the process of multi-camera vision. Last time, we studies the “Essential” and “Fundamental”
1 Webcam Mouse Using Face and Eye Tracking in Various Illumination Environments Yuan-Pin Lin et al. Proceedings of the 2005 IEEE Y.S. Lee.
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
Presented by Anuradha Venkataraman.  Introduction  Existing Approaches  Video Representation  Volume Based Video Representation  Events and Actions.
Motion Analysis using Optical flow CIS750 Presentation Student: Wan Wang Prof: Longin Jan Latecki Spring 2003 CIS Dept of Temple.
Vehicle Segmentation and Tracking From a Low-Angle Off-Axis Camera Neeraj K. Kanhere Committee members Dr. Stanley Birchfield Dr. Robert Schalkoff Dr.
A DISTRIBUTION BASED VIDEO REPRESENTATION FOR HUMAN ACTION RECOGNITION Yan Song, Sheng Tang, Yan-Tao Zheng, Tat-Seng Chua, Yongdong Zhang, Shouxun Lin.
Expectation-Maximization (EM) Case Studies
1 Motion Analysis using Optical flow CIS601 Longin Jan Latecki Fall 2003 CIS Dept of Temple University.
Final Review Course web page: vision.cis.udel.edu/~cv May 21, 2003  Lecture 37.
Object Tracking - Slide 1 Object Tracking Computer Vision Course Presentation by Wei-Chao Chen April 05, 2000.
Presented by: Idan Aharoni
Visual Tracking by Cluster Analysis Arthur Pece Department of Computer Science University of Copenhagen
Using decision trees to build an a framework for multivariate time- series classification 1 Present By Xiayi Kuang.
Representing Moving Images with Layers J. Y. Wang and E. H. Adelson MIT Media Lab.
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
Computational Intelligence: Methods and Applications Lecture 26 Density estimation, Expectation Maximization. Włodzisław Duch Dept. of Informatics, UMK.
Detection, Tracking and Recognition in Video Sequences Supervised By: Dr. Ofer Hadar Mr. Uri Perets Project By: Sonia KanOra Gendler Ben-Gurion University.
Face Detection In Color Images Wenmiao Lu Shaohua Sun Group 3 EE368 Project.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Motion Estimation of Moving Foreground Objects Pierre Ponce ee392j Winter March 10, 2004.
Bayesian Decision Theory Case Studies CS479/679 Pattern Recognition Dr. George Bebis.
Computer vision: models, learning and inference
Traffic Sign Recognition Using Discriminative Local Features Andrzej Ruta, Yongmin Li, Xiaohui Liu School of Information Systems, Computing and Mathematics.
A Forest of Sensors: Using adaptive tracking to classify and monitor activities in a site Eric Grimson AI Lab, Massachusetts Institute of Technology
Classification of unlabeled data:
Particle Filtering for Geometric Active Contours
Vehicle Segmentation and Tracking in the Presence of Occlusions
Representing Moving Images with Layers
Image Segmentation Techniques
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Announcements Questions on the project? New turn-in info online
EM Algorithm and its Applications
Presentation transcript:

Using spatio-temporal probabilistic framework for object tracking By: Guy Koren-Blumstein Supervisor: Dr. Hayit Greenspan Emphasis on Face Detection & Tracking

Agenda ► Previous research overview (PGMM) ► Under-segmentation problem ► Face tracking using PGMM  Modeling skin color in [L,a,b] color space – over-segmentation problem ► Optical flow – overview ► Approaches for using optical flow ► Examples

Previous research ► Complementary research to an M.Sc. Thesis research conducted by A.Mayer under the supervision of Dr. H.Greenspan and Dr. J. Goldberger. ► Research Goal: Building a probabilistic framework for spatio-temporal video representation. ► Useful for:  Offline – automatic search in video databases  Online – characterization of events and alerting on those that are defined as ‘suspicious’

Previous research Parsing clip to BOF Build feature Space [L a b] Build GMM modelGMM model In [Lab] space Label BOF pixels Connect. Comp. On [L,a,b,x,y,t] Learn GMM model On [L,a,b,x,y,t] Source ClipBOF 1Labeled BOFBlob Extraction Under segmentation problem…

Building GMM model ► Each sample in the [L,a,b] feature space is a sample of random vector x  R d with the following PDF:

Building GMM model ► Given a set of feature vectors x1,..., xn the maximum likelihood estimation of  is : ► Obtaining  ML by using 2 stages iterative EM algorithm: Expectation step Maximization step

Labeling pixels ► The Labeling of a pixel related to the feature vector x is chosen as the maximum a posteriori probability, as follows:

Face Detection & Tracking ► Most of the known techniques can be divided into two categories :  Search for skin color and apply shape analysis to distinguish between facial and non-facial objects.  Search for facial features regardless of pixel color (eyes,nose,mouth,chin,symmetry etc.)

Apply framework to track faces ► The framework can extract and track after objects in an image sequence. ► Applying shape analysis to each skin- colored-blob can label the blob as ‘face’ or ‘non-face’. ► The face will be tracked by virtue of the tracking capabilities of the framework

Skin color in [L a b] ► Skin color is modeled in [a b] components only ► Supplies very good discriminability between ‘skin’ pixels and ‘not-skin’ pixels (high rate of True-Negative) ► Not optimal in terms of True-Positive (leads to mis- detection of skin color pixels)

Over-segmentation of faces ► Building blobs is done in [L a b] color space. ► More than one blob might have skin color [a b] components ► Solution : Unite all blobs whose [a b] are close enough to the skin color model (adaptive TH can be used)

Under Segmentation ► Faces moving in front of skin-color background are not extracted well. ► Applying shape analysis on the middle map yields mis-detection of faces.

Employing motion information ► Motion information helps to distinguish between foreground dynamic objects and static background ► 2 levels of motion information  Binary – indicates for each pixel whether it is in motion or not. Does not supply motion vector. Feature space: [L a b x y t m] where m={0,1}  Optical flow - supplies motion vector according to a given model. Feature space: [L a b x y t V x V y ]

Is binary information good enough?

Optical Flow ► Optical flow is an apparent motion of image brightness ► If I(x,y,t) is the brightness, two main assumptions can be made:  I(x,y,t) depends on coordinates x,y in greater part of the image  Brightness of every point of moving object does not change in time

Optical Flow ► If object is moving during time dt and its displacement is (dx,dy) then using Taylor series ► According to assumption 2: ► Dividing by dt gives the optical flow equation:

Optical Flow – Block Matching ► Does not use the equation directly. ► Divides the image to blocks ► For every block in I t it search for the best matching block in I t-1. ► Matching criteria: Cross Correlation, Square Difference, SAD etc.

Working with 8-D feature space ► Connected component analysis:  Does not require initialization of the order of the model  Hard decision prone ► GMM model via EM:  Initialized by K means. Requires initialization of K.  Impose elliptic shape on the objects  Soft Decision prone Parsing clip to BOF Build feature Space [L a b] Build GMM model In [Lab] space Label BOF pixels Connect. Comp. On [x,y,t,V x,V y ] Learn GMM model [x,y,t,V x,V y ] Frame By Frame Tracking

Frame by frame tracking ► Widely used in the literature ► Can handle variations in object’s velocity ► Tracking can be improved by employing Kalman filter to predict object’s location and velocity Predict params for next frame merge blobs Create new blobs split blobs Kill old blobs Label by updated parameters Update blob’s params Label by predicted parameters

Examples ► Opposite directions:  Optical Flow, Connected component (Extracted Faces), GMM Optical FlowConnected componentExtracted Faces GMM Optical FlowConnected componentExtracted Faces GMM ► Same direction, different velocity  Optical Flow, Connected component, GMM (Faces) Optical FlowConnected componentGMM Faces Optical FlowConnected componentGMM Faces ► Different directions – complex background  Optical Flow, Connected component, GMM: K=5,K=3,Faces Optical FlowConnected component K=5K=3Faces Optical FlowConnected component K=5K=3Faces ► Variable velocity  Optical Flow, Connected component, GMM, Frame By Frame Optical FlowConnected componentGMMFrame By Frame Optical FlowConnected componentGMMFrame By Frame

Real World Sequences ► Face tracking  Optical Flow Optical Flow Optical Flow  No motion info No motion info No motion info  Connected component Connected component Connected component  GMM GMM  Frame By Frame Frame By Frame By Frame ► Car Tracking  Optical Flow Optical Flow Optical Flow  No Motion info No Motion info No Motion info  GMM GMM ► Flower garden  Optical Flow Optical Flow  No motion info No motion info  Connected component Connected component  GMM GMM

Summary ► Applying probabilistic framework to track faces in video clips ► Working in [L,a,b] color space to detect faces ► Handling over segmentation ► Handling under segmentation by employing optical flow information in 3 different ways:  Connected Component Analysis  Learning GMM model  Frame By Frame tracking

Further Research ► Adaptive face color model ► Variable length BOF (using MDL) ► Using more complex motion model

Thank you for listening Questions ?