Presented by: Idan Aharoni

Slides:



Advertisements
Similar presentations
SE263 Video Analytics Course Project Initial Report Presented by M. Aravind Krishnan, SERC, IISc X. Mei and H. Ling, ICCV’09.
Advertisements

For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Human Identity Recognition in Aerial Images Omar Oreifej Ramin Mehran Mubarak Shah CVPR 2010, June Computer Vision Lab of UCF.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Mapping: Scaling Rotation Translation Warp
Foreground Modeling The Shape of Things that Came Nathan Jacobs Advisor: Robert Pless Computer Science Washington University in St. Louis.
Tracking Multiple Occluding People by Localizing on Multiple Scene Planes Professor :王聖智 教授 Student :周節.
Multiple People Detection and Tracking with Occlusion Presenter: Feifei Huo Supervisor: Dr. Emile A. Hendriks Dr. A. H. J. Stijn Oomes Information and.
Human-Computer Interaction Human-Computer Interaction Segmentation Hanyang University Jong-Il Park.
Learning Semantic Scene Models From Observing Activity in Visual Surveillance Dimitios Makris and Tim Ellis (2005) Presented by Steven Wilson.
Formation et Analyse d’Images Session 8
1 Robust Video Stabilization Based on Particle Filter Tracking of Projected Camera Motion (IEEE 2009) Junlan Yang University of Illinois,Chicago.
Motion Detection And Analysis Michael Knowles Tuesday 13 th January 2004.
Multi video camera calibration and synchronization.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Motion Analysis (contd.) Slides are from RPI Registration Class.
Contents Description of the big picture Theoretical background on this work The Algorithm Examples.
Segmentation Divide the image into segments. Each segment:
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
Detecting and Tracking Moving Objects for Video Surveillance Isaac Cohen and Gerard Medioni University of Southern California.
CS CS 175 – Week 2 Processing Point Clouds Registration.
Object Detection and Tracking Mike Knowles 11 th January 2005
CSSE463: Image Recognition Day 30 Due Friday – Project plan Due Friday – Project plan Evidence that you’ve tried something and what specifically you hope.
Fitting a Model to Data Reading: 15.1,
MULTIPLE MOVING OBJECTS TRACKING FOR VIDEO SURVEILLANCE SYSTEMS.
Multi-camera Video Surveillance: Detection, Occlusion Handling, Tracking and Event Recognition Oytun Akman.
1 Real Time, Online Detection of Abandoned Objects in Public Areas Proceedings of the 2006 IEEE International Conference on Robotics and Automation Authors.
Computer Vision - A Modern Approach Set: Segmentation Slides by D.A. Forsyth Segmentation and Grouping Motivation: not information is evidence Obtain a.
ICBV Course Final Project Arik Krol Aviad Pinkovezky.
Automatic Camera Calibration
CSE 185 Introduction to Computer Vision
Feature and object tracking algorithms for video tracking Student: Oren Shevach Instructor: Arie nakhmani.
Tracking by Sampling Trackers Junseok Kwon* and Kyoung Mu lee Computer Vision Lab. Dept. of EECS Seoul National University, Korea Homepage:
CSSE463: Image Recognition Day 30 This week This week Today: motion vectors and tracking Today: motion vectors and tracking Friday: Project workday. First.
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
MESA LAB Multi-view image stitching Guimei Zhang MESA LAB MESA (Mechatronics, Embedded Systems and Automation) LAB School of Engineering, University of.
Motion Segmentation By Hadas Shahar (and John Y.A.Wang, and Edward H. Adelson, and Wikipedia and YouTube) 1.
Latent SVM 1 st Frame: manually select target Find 6 highest weighted areas in template Area of 16 blocks Train 6 SVMs on those areas Train 1 SVM on entire.
Single View Geometry Course web page: vision.cis.udel.edu/cv April 9, 2003  Lecture 20.
Vehicle Segmentation and Tracking From a Low-Angle Off-Axis Camera Neeraj K. Kanhere Committee members Dr. Stanley Birchfield Dr. Robert Schalkoff Dr.
Stable Multi-Target Tracking in Real-Time Surveillance Video
Robust Object Tracking by Hierarchical Association of Detection Responses Present by fakewen.
Expectation-Maximization (EM) Case Studies
Boosted Particle Filter: Multitarget Detection and Tracking Fayin Li.
By: David Gelbendorf, Hila Ben-Moshe Supervisor : Alon Zvirin
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Visual Tracking by Cluster Analysis Arthur Pece Department of Computer Science University of Copenhagen
CS 376b Introduction to Computer Vision 03 / 31 / 2008 Instructor: Michael Eckmann.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Robotics Chapter 6 – Machine Vision Dr. Amit Goradia.
Color Image Segmentation Mentor : Dr. Rajeev Srivastava Students: Achit Kumar Ojha Aseem Kumar Akshay Tyagi.
MOTION Model. Road Map Motion Model Non Parametric Motion Field : Algorithms 1.Optical flow field estimation. 2.Block based motion estimation. 3.Pel –recursive.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Motion Estimation of Moving Foreground Objects Pierre Ponce ee392j Winter March 10, 2004.
SIFT Scale-Invariant Feature Transform David Lowe
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
University of Ioannina
Paper – Stephen Se, David Lowe, Jim Little
Motion Detection And Analysis
A special case of calibration
Fast and Robust Object Tracking with Adaptive Detection
ISOMAP TRACKING WITH PARTICLE FILTERING
Vehicle Segmentation and Tracking in the Presence of Occlusions
Vehicle Segmentation and Tracking from a Low-Angle Off-Axis Camera
Effective and Efficient Detection of Moving Targets From a UAV’s Camera
CSSE463: Image Recognition Day 30
CSSE463: Image Recognition Day 30
Calibration and homographies
Tracking Many slides adapted from Kristen Grauman, Deva Ramanan.
Presentation transcript:

Presented by: Idan Aharoni Homography Based Multiple Camera Detection and Tracking of People in a Dense Crowd Ran Eshel and Yael Moses Presented by: Idan Aharoni

Motivation Usually for surveillance, but not only. Many cameras, create enormous amount of data, impossible to track manually. Many real life scenes are crowded.

Single camera tracking Many papers about this issue, some of them were presented in this course: Floor fields for tracking in high density crowds Unsupervised Bayesian detection of independent motion in crowds Particle Filters. etc.

Single camera tracking problems Not isolated body parts (human shape trackers) Targets interactions Target blocking each other …

Algorithm Overview Combining data from a set of cameras over looking the same scene. Based on that data, try to detect human head tops. Track after detected head tops by using assumptions on the expected trajectory.

Scene Example

What Is a Homography? Homography is a coordinates transformation from one image to another – represented by a 3x3 matrix. Possible in 2 case only. Camera rotation. Same plane

More Homographys Translation: Rotation: Affine:

More Homographys Projection: It describes what happens to the perceived positions of observed objects when the point of view of the observer changes. Need only 4 points to calculate. (defined up to scaling factor)

Not a Homography! Barrel Correction:

Homography Points Detection For each camera, we want to slice find 4 points in each height plane.

Homography Points Detection

Height Calculation Cross ratio of 4 pixels:

Floor Plane Projection We can define a homography from the image to itself, that will transform a height plane to another height plane. Again, all we need are 4 points of each height.

Head Top Detection Head Top – The highest 2D patch of a person. The detection is based on co-temporal frames – frames that were taken at the same times, from different cameras.

Head Top Detection Camera A Camera B B projected onto A plane A on B

Background Subtraction First stage of the algorithm. All the next stages are performed on foreground pixels only. Subtract each frame from an offline background sample.

What is Hyper-Pixel? A hyper pixel is a Nx1 vector (N denotes the number of cameras) q – Reference image pixel. - Homography related pixels in the rest of the images. - Homography transformation of image i onto the reference image (opposite of ). I – Intensity level.

Hyper-Pixel Usage Hyper pixel is calculated for each foreground pixel of the reference image. By using the hyper pixel intensity variance we can estimate the correlation between the pixels from the different image.

Hyper Pixel Variance Low Variance Low Variance High Variance

2D patches Now we have a map of variances, for each pixel. We need to obtain candidates for real projected pixels. Use variance thresholds and head size clustering (K-Means).

K-Means Clustering Partition N observations into K clusters Each observation belongs to the cluster with the nearest mean. Repeat until convergence… Thanks Wiki!

Back to Floor Projection… A person can be detected on more than one height plane. All heights are projected to the floor, and only highest patch is taken… A Head!

Example Reference foreground Projected foregrounds Variance map Single height detection All heights detection Track

Tracking So far we have a map of potential heads and heights. Tracking should remove false positives and false negatives. For that we define a few prior based measurements.

Tracking – First Stage In this stage, we aim to remove false negatives. For that we have two head maps. One with high threshold, and one with low threshold, (projected to the floor) High threshold yields less false positives, but more false negatives.

Tracking – First Stage High threshold map: If we have a hole, we try to make fill it in the low threshold map.

Tracking – First Stage If no match could be found in high and low maps, we stop the tracking after this track.

Tracking – Second Stage Now we have a list of fractioned tracks Very easy for a human to figure out which one goes where…

Tracking – Second Stage In this stage we aim to connect fragmented tracks, by using priors of how people move. For that we define a score, which is calculated out of 6 parameters, for each pair of time overlapped tracks.

Second Stage - Scores The difference in direction 2. Direction change required

Second Stage - Scores 3) Amount of overlap between tracks (4). 4) Minimal distance along the tracks (3). 5) Average distance along the tracks.

Second Stage - Scores 6) Height change – Not very likely in a tracking time frame…

Tracking - Scores Score calculation: : Maximum expected value of score

Tracking – Final stage We now have full length set of trajectories. In this stage, tracks that are suspected as false positives are removed.

Tracking – Final stage For each trajectory, we use a consistency score between each 2 consecutive frames Consistency score is made of weighted average of: Un-natural speed changes. Un-natural direction changes. Changes in height. Too short track Length.

Results - scene Cameras: Scene: 3 - 9 grey level cameras. 15 fps, 640x512. 30⁰ related to each other. 45⁰ below horizon. Scene: 3x6 meters

Criteria True Positive (TP): 75% - 100% of the trajectory is tracked (might be with IDC) Perfect True Positive (PTP) – 100% of the trajectory is tracked (no IDC). Detection Rate (DR): percent of frames tracked compare to full trajectory. ID Change (IDC) False Negative (FN): less than 75% of the trajectory is tracked. False Positive (FP): Track with no real trajectory.

Results – Summary Seq GT TP PTP IDC DR% FN FP S1 27 26 23 3 98.7 1 6 42 41 39 97.9 5 S3a 19 100 S3b 18 2 S3c 21 20 99.1 S4 22 S5 24 14 12 94.4 Total 174 171 155 16 98.4

Varying the number of cameras It seems like we need at least 8-9 cameras…

Questions?