Trinity College Dublin PixelGT: A new Ground Truth specification for video surveillance Dr. Kenneth Dawson-Howe, Graphics, Vision and Visualisation Group.

Slides:



Advertisements
Similar presentations
24/10/02AutoArch Overview Computer Vision and Media Group: Selected Previous Work David Gibson, Neill Campbell Colin Dalton Department of Computer Science.
Advertisements

Dawn Pedersen Art Institute. Introduction Prior to Flash CS3, the movement of Illustrator and Photoshop documents into Flash was difficult. At that time,
ETISEO Project Corpus data - Video sequences contents - Silogic provider.
DDDAS: Stochastic Multicue Tracking of Objects with Many Degrees of Freedom PIs: D. Metaxas, A. Elgammal and V. Pavlovic Dept of CS, Rutgers University.
ETISEO, Nice, May PETS International Workshops on Performance Evaluation of Tracking and Surveillance James Ferryman Computational Vision Group.
PETS’05, Beijing, October 16 th 2005 ETISEO Project Ground Truth & Video annotation.
Automated Shot Boundary Detection in VIRS DJ Park Computer Science Department The University of Iowa.
INRETS, Villeneuve d’Ascq, December 15 th -16 th 2005 ETISEO Annotation rules Data structure Annotation tool and format Ground truth creation rules Reference.
Wallflower Principles and Practice of Background Maintenance
Database-Based Hand Pose Estimation CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
Foreground Modeling The Shape of Things that Came Nathan Jacobs Advisor: Robert Pless Computer Science Washington University in St. Louis.
Adviser : Ming-Yuan Shieh Student ID : M Student : Chung-Chieh Lien VIDEO OBJECT SEGMENTATION AND ITS SALIENT MOTION DETECTION USING ADAPTIVE BACKGROUND.
IEEE TCSVT 2011 Wonjun Kim Chanho Jung Changick Kim
Chapter 11 Beyond Bag of Words. Question Answering n Providing answers instead of ranked lists of documents n Older QA systems generated answers n Current.
Guide to Oracle10G1 Introduction To Forms Builder Chapter 5.
1 Static Sprite Generation Prof ︰ David, Lin Student ︰ Jang-Ta, Jiang
A Guide to Oracle9i1 Introduction To Forms Builder Chapter 5.
Region-Level Motion- Based Background Modeling and Subtraction Using MRFs Shih-Shinh Huang Li-Chen Fu Pei-Yung Hsiao 2007 IEEE.
Vigilant Real-time storage and intelligent retrieval of visual surveillance data Dr Graeme A. Jones.
Tracking Video Objects in Cluttered Background
Background Subtraction for Urban Traffic Monitoring using Webcams Master Graduation Project Final Presentation Supervisor: Rein van den Boomgaard Mark.
1 Real Time, Online Detection of Abandoned Objects in Public Areas Proceedings of the 2006 IEEE International Conference on Robotics and Automation Authors.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Image Subtraction for Real Time Moving Object Extraction Shahbe Mat Desa, Qussay A. Salih, CGIV’04.
A Vision-Based System that Detects the Act of Smoking a Cigarette Xiaoran Zheng, University of Nevada-Reno, Dept. of Computer Science Dr. Mubarak Shah,
School of Computer Science and Statistics,
© 2010 Delmar, Cengage Learning Chapter 6: Preparing and Publishing Movies.
Created by: Ian Osborn. Possibilities Of Movie Maker Windows Movie Maker allows users to organize and add effects to media clips that ordinarily would.
Object detection, tracking and event recognition: the ETISEO experience Andrea Cavallaro Multimedia and Vision Lab Queen Mary, University of London
REU WEEK IV Malcolm Collins-Sibley Mentor: Shervin Ardeshir.
Olga Zoidi, Anastasios Tefas, Member, IEEE Ioannis Pitas, Fellow, IEEE
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
EE 492 ENGINEERING PROJECT LIP TRACKING Yusuf Ziya Işık & Ashat Turlibayev Yusuf Ziya Işık & Ashat Turlibayev Advisor: Prof. Dr. Bülent Sankur Advisor:
Scientific Writing Abstract Writing. Why ? Most important part of the paper Number of Readers ! Make people read your work. Sell your work. Make your.
ETISEO Benoît GEORIS, François BREMOND and Monique THONNAT ORION Team, INRIA Sophia Antipolis, France Nice, May th 2005.
1 ETISEO: Video Understanding Performance Evaluation Francois BREMOND, A.T. Nghiem, M. Thonnat, V. Valentin, R. Ma Orion project-team, INRIA Sophia Antipolis,
AUTOMATIC TARGET RECOGNITION OF CIVILIAN TARGETS September 28 th, 2004 Bala Lakshminarayanan.
BING: Binarized Normed Gradients for Objectness Estimation at 300fps
Tracking and event recognition – the Etiseo experience Son Tran, Nagia Ghanem, David Harwood and Larry Davis UMIACS, University of Maryland.
Figure ground segregation in video via averaging and color distribution Introduction to Computational and Biological Vision 2013 Dror Zenati.
Student Name: Honghao Chen Supervisor: Dr Jimmy Li Co-Supervisor: Dr Sherry Randhawa.
Human Activity Recognition at Mid and Near Range Ram Nevatia University of Southern California Based on work of several collaborators: F. Lv, P. Natarajan,
An Effective & Interactive Approach to Particle Tracking for DNA Melting Curve Analysis 李穎忠 DEPARTMENT OF COMPUTER SCIENCE & INFORMATION ENGINEERING NATIONAL.
Active Frame Selection for Label Propagation in Videos Sudheendra Vijayanarasimhan and Kristen Grauman Department of Computer Science, University of Texas.
High Resolution Surface Reconstruction from Overlapping Multiple-Views
Yannick FranckenChris HermansPhilippe Bekaert Hasselt University – tUL – IBBT Expertise Centre for Digital Media, Belgium
Suspicious Behavior in Outdoor Video Analysis - Challenges & Complexities Air Force Institute of Technology/ROME Air Force Research Lab Unclassified IED.
Unit 14 - Desktop Publishing Part 2 Desktop Publishing InDesign – Inside Pages
Adaptive Faceted Browsing in Job Offers Danielle H. Lee
ETISEO François BREMOND ORION Team, INRIA Sophia Antipolis, France.
Animating Objects in PowerPoint XP/2003 Using Motion Paths and Custom Animations to add interest to your classroom presentations.
Preliminary Transformations Presented By: -Mona Saudagar Under Guidance of: - Prof. S. V. Jain Multi Oriented Text Recognition In Digital Images.
Occlusion Tracking Using Logical Models Summary. A Variational Partial Differential Equations based model is used for tracking objects under occlusions.
Portable Camera-Based Assistive Text and Product Label Reading From Hand-Held Objects for Blind Persons.
Motion Estimation of Moving Foreground Objects Pierre Ponce ee392j Winter March 10, 2004.
REAL-TIME DETECTOR FOR UNUSUAL BEHAVIOR
Multimedia Content-Based Retrieval
Gender Classification Using Scaled Conjugate Gradient Back Propagation
Lecture 26 Hand Pose Estimation Using a Database of Hand Images
Machine Vision Acquisition of image data, followed by the processing and interpretation of these data by computer for some useful application like inspection,
Vehicle Segmentation and Tracking in the Presence of Occlusions
Text Detection in Images and Video
Rob Fergus Computer Vision
Introduction Computer vision is the analysis of digital images
Grape Detection in Vineyards Introduction To Computational and Biological Vision final project Kobi Ruham Eli Izhak.
The Image The pixels in the image The mask The resulting image 255 X
Report 4 Brandon Silva.
SFNet: Learning Object-aware Semantic Correspondence
Report 2 Brandon Silva.
Presentation transcript:

Trinity College Dublin PixelGT: A new Ground Truth specification for video surveillance Dr. Kenneth Dawson-Howe, Graphics, Vision and Visualisation Group (GV2), School of Computer Science & Statistics, Trinity College Dublin

Trinity College Dublin Contents  Introduction  Review  Specification  Annotation  Conclusions Introduction Review Specification Annotation Conclusions

Trinity College Dublin Available video sequences  Video sequences with no ground truth  Video sequences with moving object pixel masks  Video sequences with moving object bounding boxes  Video sequences with labeled events Introduction Review Specification Annotation Conclusions

Trinity College Dublin Available video sequences Introduction Review Specification Annotation Conclusions

Trinity College Dublin Video sequences with moving object pixel masks  Video sequences with moving object and moving shadow pixel masks.  Artificial video sequences with moving object pixel masks.  Video sequences with labeled pixels. Introduction Review Specification Annotation Conclusions CamVid dataset VSSN dataset PETS/LIMU dataset ATON dataset

Trinity College Dublin Video sequences with moving object bounding boxes Introduction Review Specification Annotation Conclusions ETISEO dataset CAVIAR dataset

Trinity College Dublin Video sequences with labeled events Introduction Review Specification Annotation Conclusions PETS 2006 dataset

Trinity College Dublin Motivation  Existing ground truth requires a choice:  Binary pixel masks (with no specific object identification) OR  Object identification (with bounding boxes only and no shadows).  Relatively little pixel accurate ground truth is available.  No existing means of dealing with transparency and reflections. Introduction Review Specification Annotation Conclusions

Trinity College Dublin Ground truth Specifcation  Pixel based ground truth  Bit organisation  Meta data  Gradient profiles  Bounding boxes & events Introduction Review Specification Annotation Conclusions

Trinity College Dublin Pixel based ground truth  Object class (8 bits)  Object instance (9 bits)  Shadow level (4 bits)  Transparency level (3 bits) Introduction Review Specification Annotation Conclusions

Trinity College Dublin Bit organisation Introduction Review Specification Annotation Conclusions L = Label class I = ID S = Shadow T = Transparency 1 = LSB 8 = MSB

Trinity College Dublin Meta data  Based on CVML  Includes:  File names  Bit organisation  Meaning of object labels  Gradient profile Introduction Review Specification Annotation Conclusions

Trinity College Dublin Gradient profiles  Width of edges?  Required for assessing performance? Introduction Review Specification Annotation Conclusions

Trinity College Dublin Bounding boxes and events  Bounding boxes for moving objects  Implicitedly included  Can be explicitly specified  Aids comparison of techniques  Events  Some implicitedly included  Can be explicitly specified  Aids comparison of techniques Introduction Review Specification Annotation Conclusions

Trinity College Dublin Creation of ground truth  Difficult to label all pixels in all frames  Short videos?  Not all frames / pixels?  Method  Background annotation  Object and shadow boundary specification  Object class and ID annotation  Transparency, Meta-data, Bounding boxes, Events and Gradient Profile specification  Propagation of ground truth Introduction Review Specification Annotation Conclusions

Trinity College Dublin Specifying object class & ID Introduction Review Specification Annotation Conclusions

Trinity College Dublin Background annotation  For static cameras  Annotate all background pixels  Facilitates automatic event detection Introduction Review Specification Annotation Conclusions

Trinity College Dublin Specifying object and shadow boundaries  Identify boundary edges of objects / shadows  Once shadow boundary is complete shadow level is computed automatically. Introduction Review Specification Annotation Conclusions

Trinity College Dublin Specifying object class & ID  Once boundary is complete  Select object class and (optionally) ID  Click on the object Introduction Review Specification Annotation Conclusions

Trinity College Dublin Transparency, Meta-data, Bounding boxes, Events and Gradient Profile specification  Transparency not yet addressed  Meta data – specified in CVML  Bounding boxes & events – extract automatically from pixel based ground truth  Gradient profile – extract automatically from each image Introduction Review Specification Annotation Conclusions

Trinity College Dublin Propagation of ground truth  Propagation only required for moving objects  Working with every frame results in very little movement between frames (0-5 pixels)  Search for the closest similar contour in the new frame where the pixels on the object side of the contour appear similar…  Need to identify & correct 1.when new objects enter the scene. 2.when objects reappear from behind occlusions. 3.when internal holes appear within an object. 4.when errors occur in the propagation Introduction Review Specification Annotation Conclusions

Trinity College Dublin The positive spin…  PixelGT is a new form of ground truth which will facilitate  Pixel-level assessment of trackers  Semi-automatic generation of object location, bounding boxes and event descriptions.  Comparison of trackers which assess performance using different types of ground truth Introduction Review Specification Annotation Conclusions

Trinity College Dublin Questions  Do we really need this detailed ground truth ?  Should we change anything about the specification ?  Other questions ? Introduction Review Specification Annotation Conclusions