Abdul Jabbar Siddiqui, Abdelhamid Mammeri, and Azzedine Boukerche

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Advertisements

Group Meeting Presented by Wyman 10/14/2006
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Presented by Xinyu Chang
BRISK (Presented by Josh Gleason)
MIT CSAIL Vision interfaces Approximate Correspondences in High Dimensions Kristen Grauman* Trevor Darrell MIT CSAIL (*) UT Austin…
Intelligent Systems Lab. Recognizing Human actions from Still Images with Latent Poses Authors: Weilong Yang, Yang Wang, and Greg Mori Simon Fraser University,
Instructor: Mircea Nicolescu Lecture 15 CS 485 / 685 Computer Vision.
Addressing the Medical Image Annotation Task using visual words representation Uri Avni, Tel Aviv University, Israel Hayit GreenspanTel Aviv University,
Multiple People Detection and Tracking with Occlusion Presenter: Feifei Huo Supervisor: Dr. Emile A. Hendriks Dr. A. H. J. Stijn Oomes Information and.
Discriminative and generative methods for bags of features
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
1 Image Recognition - I. Global appearance patterns Slides by K. Grauman, B. Leibe.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Generic Object Detection using Feature Maps Oscar Danielsson Stefan Carlsson
5/30/2006EE 148, Spring Visual Categorization with Bags of Keypoints Gabriella Csurka Christopher R. Dance Lixin Fan Jutta Willamowski Cedric Bray.
California Car License Plate Recognition System ZhengHui Hu Advisor: Dr. Kang.
Multiple Object Class Detection with a Generative Model K. Mikolajczyk, B. Leibe and B. Schiele Carolina Galleguillos.
Overview Introduction to local features
Exercise Session 10 – Image Categorization
Bag of Video-Words Video Representation
Computer vision.
05 - Feature Detection Overview Feature Detection –Intensity Extrema –Blob Detection –Corner Detection Feature Descriptors Feature Matching Conclusion.
Gwangju Institute of Science and Technology Intelligent Design and Graphics Laboratory Multi-scale tensor voting for feature extraction from unstructured.
Action recognition with improved trajectories
Overview Harris interest points Comparing interest points (SSD, ZNCC, SIFT) Scale & affine invariant interest points Evaluation and comparison of different.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Marcin Marszałek, Ivan Laptev, Cordelia Schmid Computer Vision and Pattern Recognition, CVPR Actions in Context.
Reporter: Fei-Fei Chen. Wide-baseline matching Object recognition Texture recognition Scene classification Robot wandering Motion tracking.
Professor: S. J. Wang Student : Y. S. Wang
南台科技大學 資訊工程系 Automatic Website Summarization by Image Content: A Case Study with Logo and Trademark Images Evdoxios Baratis, Euripides G.M. Petrakis, Member,
Building local part models for category-level recognition C. Schmid, INRIA Grenoble Joint work with G. Dorko, S. Lazebnik, J. Ponce.
Svetlana Lazebnik, Cordelia Schmid, Jean Ponce
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
Group Sparse Coding Samy Bengio, Fernando Pereira, Yoram Singer, Dennis Strelow Google Mountain View, CA (NIPS2009) Presented by Miao Liu July
Pedestrian Detection and Localization
Object Recognition in Images Slides originally created by Bernd Heisele.
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
Kylie Gorman WEEK 1-2 REVIEW. CONVERTING AN IMAGE FROM RGB TO HSV AND DISPLAY CHANNELS.
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
Week 10 Emily Hand UNR.
Object Recognition as Ranking Holistic Figure-Ground Hypotheses Fuxin Li and Joao Carreira and Cristian Sminchisescu 1.
Using decision trees to build an a framework for multivariate time- series classification 1 Present By Xiayi Kuang.
Finding Clusters within a Class to Improve Classification Accuracy Literature Survey Yong Jae Lee 3/6/08.
Recognizing specific objects Matching with SIFT Original suggestion Lowe, 1999,2004.
A distributed PSO – SVM hybrid system with feature selection and parameter optimization Cheng-Lung Huang & Jian-Fan Dun Soft Computing 2008.
Blob detection.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
Face recognition using Histograms of Oriented Gradients
Guillaume-Alexandre Bilodeau
Bag-of-Visual-Words Based Feature Extraction
Data Driven Attributes for Action Detection
Learning Mid-Level Features For Recognition
Performance of Computer Vision
Supervised Time Series Pattern Discovery through Local Importance
Introductory Seminar on Research: Fall 2017
Basic machine learning background with Python scikit-learn
Homework| Homework: Derive the following expression for the derivative of the inverse mapping Arun Das | Waterloo Autonomous Vehicles Lab.
Paper Presentation: Shape and Matching
Feature description and matching
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
A New Approach to Track Multiple Vehicles With the Combination of Robust Detection and Two Classifiers Weidong Min , Mengdan Fan, Xiaoguang Guo, and Qing.
CAP 5415 Computer Vision Fall 2012 Dr. Mubarak Shah Lecture-5
Outline S. C. Zhu, X. Liu, and Y. Wu, “Exploring Texture Ensembles by Efficient Markov Chain Monte Carlo”, IEEE Transactions On Pattern Analysis And Machine.
Dongwook Kim, Beomjun Kim, Taeyoung Chung, and Kyongsu Yi
Feature descriptors and matching
Recognition and Matching based on local invariant features
Presentation transcript:

Real-Time Vehicle Make and Model Recognition Based on a Bag of SURF Features Abdul Jabbar Siddiqui, Abdelhamid Mammeri, and Azzedine Boukerche IEEE Transactions on Intelligent Transportation Systems, Vol. 17, No. 11, November 2016 Presented by: Yang Yu {yuyang@islab.ulsan.ac.kr} Nov. 26, 2016

Overview Real-time automated vehicle make and model recognition (VMMR) based on a bag of speeded-up robust features (BoSURF). Use SURF features of front- or rear-facing images and retain the dominant characteristic features (codewords) in adictionary. Single dictionary, modular dictionary. SURF features to BoSURF histograms Single multiclass SVM and an ensemble of multiclass SVM based on attribute bagging.

Application of VMMR Automated vehicular surveillance: parking lots of public spaces (e.g., malls, stadiums or airports). Search for a specific vehicle type, make, or model. VMMR systems based on license plates can fail due to ambiguity, forgery, damage, or license-plate duplication.

Challenges Toyota Altis, Toyota Camry, Nissan Xtrail: three classes Multiplicity: displays different shapes. Ambiguity: different companies --comparable shapes. Intramake ambiguity: same company --comparable shape or appearance.

Contributions Speed: 7 fps, accuracy: 95% (1) VMMR based on BoSURF, dominant dictionary; (2) Single-Dictionary, Modular-Dictionary : multiplicity and ambiguity problems of VMMR. (3) Optimal dictionary sizes. (4) Single Multi-Class SVM Classifier (SVM) and Attribute Bagging based Ensemble of SVM Classifiers (AB-SVM), Inter-class differences (to solve inter-make and intra-make ambiguity) Intra-class similarities (to solve multiplicity) (5) random training-testing dataset splits of NTOU-MMR Dataset

Vehicle make and model recognition Real-time automated vehicle make and model recognition (VMMR) based on a bag of speeded-up robust features (BoSURF). Architecture of VMMR systems.. MMR module of VMMR system

VMMR Vehicle detection. Features extraction and global features representation . Classification approaches. VMMR targeted environment: gates of a cross-border checkpoint.

Dataset description Speed up to 65 km/h for the oncoming vehicles. 2,846 images for training, and 3,793 images for testing. The total numberof classes is 29. Different viewing angle: −20◦ to 20◦ Daytime and night-time, and under weather conditions. Occluded by irrelevant objects

BOSURF-MMR BoSURF approach for VMMR Extract SURF features from training samples of all classes and retain the dominant ones in a “bag” or dictionary: Bag-of-SURF (BoSURF). 64-dimensional SURF. Dictionary represent vehicles’ images as BoSURF histograms

Offline Dictionary Building Extract their SURF features. The dominant features (codewords) are retained in a “bag” or dictionary I: set of training images of class i . Each j-th image in , SURF features. Single-dictionary(SD) building, modular dictionary(MD) building.

Offline Dictionary Building : p-th SURF feature in image-j : number of SURF features extracted from j-th image of class i. Pool of features

Single Dictionary SD(denoted by D), Kmeans++ clustering Group training features into a number of clusters of similar patterns. Cluster centers are the selected dominant features that make up dictionary, and are referred to as codewords The number of clusters (or codewords) determines overall dictionary size

Modular Dictionary Main dictionary (denoted by DM) by combining individual dictionaries of each class. Each Di is the individual dictionary of class i, Size of the Dictionary Affects processing speed, discriminative capacity and generalizability of the built dictionary

BoSURF features representations Use the dictionary D or DM to embed given images’ local features into global BoSURF representations through Features Quantization BoSURF features: Histogram . of votes to the dictionary codewords BoSURF histogram dist(a,b) is the euclidean distance fpji is the pth SURF feature and pji is the numberof SURF features extracted from the image Ij.

Classification Single Multi-Class SVM Classifier (SVM) Based on BoSURF histogram, binary classifiers in multi-class SVM vote to predicted class. Ensemble of Multi-Class SVM Classifiers Based on Attribute Bagging (AB-SVM) Ensemble of individual multi-class classifiers --train different random feature subspaces---avoid over-fitting Create random feature subspaces, each comprising of 5 attributes. Set of randomly chosen attribute indices for an feature subspace BoSURF feature vector is sub-sampled into subsets based on respective set of attribute indices. Use majority-voting to combine outputs to produce final prediction.

Experiment (1) Offline dictionary building (used for features quantization), (2) BoSURF features generation (3) Classification, which involves classifier training and testing. Repeatedly randomly partition the original dataset to form different training and testing splits for each class. Speed:speeds of the SD- and MD-based BoSURF approaches with SVM are 7.5 fps and 6.99 fps Accuracy: correct classification rates are 94.84% and 93.7% Dictionary Training Time: MD is significantly less than SD

Experiment Some challenging cases of vehicles under occlusion

Conclusion Real-time automated vehicle make and model recognition (VMMR) based on Bag-of-SURF features Two Dictionary Building: multiplicity and ambiguity Optimal dictionary sizes for both dictionaries; Two multi-class classification: accurate and efficient make-model prediction: Single SVM Attribute Bagging based Ensemble of SVMs (AB-SVM) Random training-testing splits of the NTOU-MMR Dataset Future work Enhance BoSURF-VMMR approaches by exploring dictionary pruning methods. Build a comprehensive VMMR dataset

SURF Speeded up robust features (SURF) In scale space to find extreme points (position, scale, direction) Neighborhood gradients histogram peak- direction Local feature detector. Use square-shaped filters as an approximation of Gaussian smoothing. Be evaluated quickly using the integral image

SURF Use a blob detector based on the Hessian matrix to find points of interest. The determinant of Hessian matrix is used as a measure of local change around the point. Points are chosen where this determinant is maximal. Cale-space representation and location of points of interest Scale spaces in SURF are implemented by applying box filters of different sizes. Up-scaling the filter size rather than iteratively reducing the image size.

SIFT SIFT: In scale space to find extreme points (position, scale, direction) Neighborhood gradients histogram peak- direction Direction- 4*4*8 HOG feature Dense SIFT: in 16*16 patch--each position SIFT features Match: Histogram intersection operation.

Thank you for attention!