Tracking Video Objects in Cluttered Background

Slides:



Advertisements
Similar presentations
IEEE CDC Nassau, Bahamas, December Integration of shape constraints in data association filters Integration of shape constraints in data.
Advertisements

RGB-D object recognition and localization with clutter and occlusions Federico Tombari, Samuele Salti, Luigi Di Stefano Computer Vision Lab – University.
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
1 Approximated tracking of multiple non-rigid objects using adaptive quantization and resampling techniques. J. M. Sotoca 1, F.J. Ferri 1, J. Gutierrez.
電腦視覺 Computer and Robot Vision I Chapter2: Binary Machine Vision: Thresholding and Segmentation Instructor: Shih-Shinh Huang 1.
Vision Based Control Motion Matt Baker Kevin VanDyke.
MPEG-4 Objective Standardize algorithms for audiovisual coding in multimedia applications allowing for Interactivity High compression Scalability of audio.
Robust Foreground Detection in Video Using Pixel Layers Kedar A. Patwardhan, Guillermoo Sapire, and Vassilios Morellas IEEE TRANSACTION ON PATTERN ANAYLSIS.
Tracking Multiple Occluding People by Localizing on Multiple Scene Planes Professor :王聖智 教授 Student :周節.
Robust Object Tracking via Sparsity-based Collaborative Model
A KLT-Based Approach for Occlusion Handling in Human Tracking Chenyuan Zhang, Jiu Xu, Axel Beaugendre and Satoshi Goto 2012 Picture Coding Symposium.
Segmentation and Region Detection Defining regions in an image.
Snakes : Active Contour models
São Paulo Advanced School of Computing (SP-ASC’10). São Paulo, Brazil, July 12-17, 2010 Looking at People Using Partial Least Squares William Robson Schwartz.
Local or Global Minima: Flexible Dual-Front Active Contours Hua Li Anthony Yezzi.
Broadcast Court-Net Sports Video Analysis Using Fast 3-D Camera Modeling Jungong Han Dirk Farin Peter H. N. IEEE CSVT 2008.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
EE 7730 Image Segmentation.
EE663 Image Processing Edge Detection 5 Dr. Samir H. Abdul-Jauwad Electrical Engineering Department King Fahd University of Petroleum & Minerals.
Efficient Moving Object Segmentation Algorithm Using Background Registration Technique Shao-Yi Chien, Shyh-Yih Ma, and Liang-Gee Chen, Fellow, IEEE Hsin-Hua.
A Study of Approaches for Object Recognition
Segmentation Divide the image into segments. Each segment:
Rodent Behavior Analysis Tom Henderson Vision Based Behavior Analysis Universitaet Karlsruhe (TH) 12 November /9.
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
Region-Level Motion- Based Background Modeling and Subtraction Using MRFs Shih-Shinh Huang Li-Chen Fu Pei-Yung Hsiao 2007 IEEE.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
Recovering Articulated Object Models from 3D Range Data Dragomir Anguelov Daphne Koller Hoi-Cheung Pang Praveen Srinivasan Sebastian Thrun Computer Science.
Effective Gaussian mixture learning for video background subtraction Dar-Shyang Lee, Member, IEEE.
Augmented Reality: Object Tracking and Active Appearance Model
Shadow Removal Seminar
MULTIPLE MOVING OBJECTS TRACKING FOR VIDEO SURVEILLANCE SYSTEMS.
Presented by Pat Chan Pik Wah 28/04/2005 Qualifying Examination
A Real-Time for Classification of Moving Objects
IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 11, NOVEMBER 2011 Qian Zhang, King Ngi Ngan Department of Electronic Engineering, the Chinese university.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
1 Activity and Motion Detection in Videos Longin Jan Latecki and Roland Miezianko, Temple University Dragoljub Pokrajac, Delaware State University Dover,
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Olga Zoidi, Anastasios Tefas, Member, IEEE Ioannis Pitas, Fellow, IEEE
1 Mean shift and feature selection ECE 738 course project Zhaozheng Yin Spring 2005 Note: Figures and ideas are copyrighted by original authors.
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
2004, 9/1 1 Optimal Content-Based Video Decomposition for Interactive Video Navigation Anastasios D. Doulamis, Member, IEEE and Nikolaos D. Doulamis, Member,
1 Lucia Maddalena and Alfredo Petrosino, Senior Member, IEEE A Self-Organizing Approach to Background Subtraction for Visual Surveillance Applications.
Spatio-temporal constraints for recognizing 3D objects in videos Nicoletta Noceti Università degli Studi di Genova.
1 Webcam Mouse Using Face and Eye Tracking in Various Illumination Environments Yuan-Pin Lin et al. Proceedings of the 2005 IEEE Y.S. Lee.
Background Subtraction based on Cooccurrence of Image Variations Seki, Wada, Fujiwara & Sumi Presented by: Alon Pakash & Gilad Karni.
November 13, 2014Computer Vision Lecture 17: Object Recognition I 1 Today we will move on to… Object Recognition.
Recognizing Action at a Distance Alexei A. Efros, Alexander C. Berg, Greg Mori, Jitendra Malik Computer Science Division, UC Berkeley Presented by Pundik.
Vision-based human motion analysis: An overview Computer Vision and Image Understanding(2007)
Computer Graphics and Image Processing (CIS-601).
Advances in digital image compression techniques Guojun Lu, Computer Communications, Vol. 16, No. 4, Apr, 1993, pp
Efficient Visual Object Tracking with Online Nearest Neighbor Classifier Many slides adapt from Steve Gu.
2005/12/021 Content-Based Image Retrieval Using Grey Relational Analysis Dept. of Computer Engineering Tatung University Presenter: Tienwei Tsai ( 蔡殿偉.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Text From Corners: A Novel Approach to Detect Text and Caption in Videos Xu Zhao, Kai-Hsiang Lin, Yun Fu, Member, IEEE, Yuxiao Hu, Member, IEEE, Yuncai.
Suspicious Behavior in Outdoor Video Analysis - Challenges & Complexities Air Force Institute of Technology/ROME Air Force Research Lab Unclassified IED.
Tracking Groups of People for Video Surveillance Xinzhen(Elaine) Wang Advisor: Dr.Longin Latecki.
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
Occlusion Tracking Using Logical Models Summary. A Variational Partial Differential Equations based model is used for tracking objects under occlusions.
Color Image Segmentation Mentor : Dr. Rajeev Srivastava Students: Achit Kumar Ojha Aseem Kumar Akshay Tyagi.
Learning and Removing Cast Shadows through a Multidistribution Approach Nicolas Martel-Brisson, Andre Zaccarin IEEE TRANSACTIONS ON PATTERN ANALYSIS AND.
Portable Camera-Based Assistive Text and Product Label Reading From Hand-Held Objects for Blind Persons.
ENTERFACE 08 Project 9 “ Tracking-dependent and interactive video projection ” Mid-term presentation August 19th, 2008.
V. Mezaris, I. Kompatsiaris, N. V. Boulgouris, and M. G. Strintzis
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Object Recognition Today we will move on to… April 12, 2018
Presentation transcript:

Tracking Video Objects in Cluttered Background Andrea Cavallaro, Member, IEEE, Olivier Steiger, Member, IEEE, and Touradj Ebrahimi, Member, IEEE

Outline Introduction Related works Hybrid video object tracking Experimental results Conclusion

Introduction(1/2) Goal: Tracking multiple video objects in cluttered background Using object and region information to solve the problem Appearance and disappearance of objects, splitting and partial occlusions are resolved Video object extraction Video object segmentation Identifying objects in the scene and separating them from the background. Video object tracking Following video objects in the scene and at updating their two-dimensional (2-D) shape from frame to frame.

Introduction(2/2) Establish stable track for dynamic scene  Effective track management Track initiation Track update Track termination Temporal variation of the 2-D shape of video objects Nonrigid objects Occlusion Splitting Appearance and disappearance of objects

Related works Object tracking methods can be classified into five groups: model-based, appearance-based, contour- and mesh-based, feature-based, and hybrid methods. Model-based: Exploit the a priori knowledge of the shape of typical objects in a given scene Computationally expensive Need for object models with detailed geometry for all objects that could be found in the scene Lack of generality.

Related works Appearance-based: Contour- and mesh-based Feature-based Relies on information provided by the entire region Cannot usually cope with complex deformation. Contour- and mesh-based Track only the contour of the object. Computational complexity is high Large nonrigid movements cannot be handled by the method Active contour models (snakes) [8]–[10] Feature-based Uses features of a video object to track parts of the object. Stable tracks for the features under analysis even in case of partial occlusion of the object The problem is how to group the features to determine which of them belong to the same object

Related works Hybrid methods A hybrid between a region-based and a feature-based technique Track video objects based on interactions between different levels of the hierarchy. Higher computational complexity This paper uses a hybrid object tracking algorithm No need for computationally expensive motion models. It can cope with deformation and complex motion

Hybrid video object tracking : Object partition of frame n Object partition is generated by the method presented in [16]. : Region partition of objects in frame n Take spatiotemporal properties into account and extracts homogeneous regions.[17]

Region partition Feature space used here is composed of spatial and temporal features. color features, texture ,displacement vectors

Region descriptor Create a tracking mechanism for deformable objects. Each region is represented by a region descriptor. The region descriptor summarizes the value of the features in the corresponding region. Next, the tracking mechanism operates on region descriptors.

Region tracking Region tracking is based on a flexible procedure Projects the region descriptors from the current frame onto the next frame, Refines the region partition so as to naturally create the updated 2-D topology The region descriptor is defined as is the number of features in frame represent the position of the region descriptor, is motion vector

In the specific implementation, Ki(n) = 8 , and represents the mean value of the three color components in the corresponding region, is the mean value of the texture feature The position predicted through motion compensation Predicted region descriptor

Multilevel Region-Object Tracking The joint region-object tracking mechanism is organized in two major steps: Object partition validation Results in a tentative correspondence Data association generates the final correspondence.

Object Partition Validation Initializes the tracking process and improves the accuracy of the object partition Before initializing the tracking procedure… video object is decomposed into a set of regions Each region is characterized by its region descriptor Track initiation: Each region descriptor is associated to the corresponding object Takes place at the beginning of the tracking process and every time a new video object appears

Object Partition Validation After the initialization: Region descriptors are projected into the next frame. The predicted region is defined as After the projection A bottom-up feedback from the region partition refines the topology of the object partition. The feedback generates a tentative correspondence by labeling the object partition according to the predicted region partition

Object Partition Validation Once all the pixels in the next object partition are associated to the projected regions, we have a prediction as follows: This procedure is straightforward in case each set of connected pixels receives projected region descriptors, and receives them from one object only In reality, multiple simultaneous objects may occlude each other and therefore be included in the same set of connected pixels. The bottom-up interaction is used to improve the object labeling in these cases

bottom-up interaction A new object A connected set of pixels in does not get any region descriptor from the projection mechanism. The detection of a new object triggers a track initiation An occlusion A connected set of pixels in receives projected region descriptors from several objects

bottom-up interaction A splitting Two different disconnected sets of pixels get region descriptors projected from the same video object. The predicted partition may not cover all the pixels If a connected component of receives region descriptors from one object only Unclassified pixels are assigned to that object. If a connected set of receives region descriptors from several objects The unclassified pixels are assigned to the closest projected region.

Data Association Data association validates the track of each region descriptor, and this step is particularly important when faced with track management issues. To verify the correctness of the tentative correspondence obtained with region descriptors projection, we consider the proximity between region descriptors in and in

Data Association The proximity is computed by measuring the distance in the feature space between the region descriptors Before distance computation,using preselection to eliminate the computation The Mahalanobis distance The distance computeed within each category A tentative correspondence between the pth region description frame and the qth region descriptor in frame n+1 is confirmed if

Experimental results

Experimental results

Experimental results

Experimental results

Conclusion We presented an automatic tracking algorithm based on interactions between video objects and their regions Using region descriptor to track the corresponding video object Future works Simpler region segmentation algorithm Initialization of a track when groups of objects enter the scene and total occlusions The data association step may operate on a longer temporal window