Ignas Budvytis*, Tae-Kyun Kim*, Roberto Cipolla * - indicates equal contribution Making a Shallow Network Deep: Growing a Tree from Decision Regions of.

Slides:



Advertisements
Similar presentations
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Advertisements

On-line learning and Boosting
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
Data Mining and Machine Learning
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Introduction Training Complexity, Pruning CART vs. ID3 vs. C4.5
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
Data Mining Classification: Alternative Techniques
Silhouette-based Object Phenotype Recognition using 3D Shape Priors Yu Chen 1 Tae-Kyun Kim 2 Roberto Cipolla 1 University of Cambridge, Cambridge, UK 1.
CMPUT 466/551 Principal Source: CMU
Online Multiple Classifier Boosting for Object Tracking Tae-Kyun Kim 1 Thomas Woodley 1 Björn Stenger 2 Roberto Cipolla 1 1 Dept. of Engineering, University.
AdaBoost & Its Applications
Face detection Many slides adapted from P. Viola.
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei Li,
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
Contour Based Approaches for Visual Object Recognition Jamie Shotton University of Cambridge Joint work with Roberto Cipolla, Andrew Blake.
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
HCI Final Project Robust Real Time Face Detection Paul Viola, Michael Jones, Robust Real-Time Face Detetion, International Journal of Computer Vision,
Decision Tree Rong Jin. Determine Milage Per Gallon.
Generic Object Detection using Feature Maps Oscar Danielsson Stefan Carlsson
Announcements  Project proposal is due on 03/11  Three seminars this Friday (EB 3105) Dealing with Indefinite Representations in Pattern Recognition.
Object Recognition with Informative Features and Linear Classification Authors: Vidal-Naquet & Ullman Presenter: David Bradley.
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
Ensemble Learning: An Introduction
Viola and Jones Object Detector Ruxandra Paun EE/CS/CNS Presentation
Sparse vs. Ensemble Approaches to Supervised Learning
1 Accurate Object Detection with Joint Classification- Regression Random Forests Presenter ByungIn Yoo CS688/WST665.
Spatial Pyramid Pooling in Deep Convolutional
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Ensemble Learning (2), Tree and Forest
Foundations of Computer Vision Rapid object / face detection using a Boosted Cascade of Simple features Presented by Christos Stoilas Rapid object / face.
FACE DETECTION AND RECOGNITION By: Paranjith Singh Lohiya Ravi Babu Lavu.
Machine Learning CS 165B Spring 2012
Face Detection using the Viola-Jones Method
Chapter 10. Sampling Strategy for Building Decision Trees from Very Large Databases Comprising Many Continuous Attributes Jean-Hugues Chauchat and Ricco.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Window-based models for generic object detection Mei-Chen Yeh 04/24/2012.
Sign Classification Boosted Cascade of Classifiers using University of Southern California Thang Dinh Eunyoung Kim
Skewing: An Efficient Alternative to Lookahead for Decision Tree Induction David PageSoumya Ray Department of Biostatistics and Medical Informatics Department.
BOOSTING David Kauchak CS451 – Fall Admin Final project.
DIEGO AGUIRRE COMPUTER VISION INTRODUCTION 1. QUESTION What is Computer Vision? 2.
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
ECE738 Advanced Image Processing Face Detection IEEE Trans. PAMI, July 1997.
Human pose recognition from depth image MS Research Cambridge.
Face Detection Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Automated Solar Cavity Detection
Robust Real Time Face Detection
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Lecture 6: Classification – Boosting and SVMs CAP 5415 Fall 2006.
Lecture 09 03/01/2012 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
The Viola/Jones Face Detector A “paradigmatic” method for real-time object detection Training is slow, but detection is very fast Key ideas Integral images.
FACE DETECTION : AMIT BHAMARE. WHAT IS FACE DETECTION ? Face detection is computer based technology which detect the face in digital image. Trivial task.
Classification Ensemble Methods 1
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr.
Notes on HW 1 grading I gave full credit as long as you gave a description, confusion matrix, and working code Many people’s descriptions were quite short.
A Brief Introduction on Face Detection Mei-Chen Yeh 04/06/2010 P. Viola and M. J. Jones, Robust Real-Time Face Detection, IJCV 2004.
COM24111: Machine Learning Decision Trees Gavin Brown
Face detection Many slides adapted from P. Viola.
Hand Detection with a Cascade of Boosted Classifiers Using Haar-like Features Qing Chen Discover Lab, SITE, University of Ottawa May 2, 2006.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Presenter: Jae Sung Park
Mustafa Gokce Baydogan, George Runger and Eugene Tuv INFORMS Annual Meeting 2011, Charlotte A Bag-of-Features Framework for Time Series Classification.
Evaluation of Gender Classification Methods with Automatically Detected and Aligned Faces Speaker: Po-Kai Shen Advisor: Tsai-Rong Chang Date: 2010/6/14.
1 Munther Abualkibash University of Bridgeport, CT.
Ensemble Classifiers.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
k-Nearest neighbors and decision tree
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei.
Presentation transcript:

Ignas Budvytis*, Tae-Kyun Kim*, Roberto Cipolla * - indicates equal contribution Making a Shallow Network Deep: Growing a Tree from Decision Regions of a Boosting Classifier

Introduction Aim – improved classification time of a learnt boosting classifier Shallow network of boosting classifier converted into a “deep” decision tree based structure Applications Real time detection and tracking Object segmentation Design goals Significant speed up Similar accuracy BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge2/22

Speeding up a boosting classifier Creating a cascade of boosting classifiers Robust Real-time Object Detection [Viola & Jones 02] Single path of varying length “Fast exit” [Zhou 05] Sequential probability ratio test [Sochman et. al. 05] Multiple paths of different lengths A binary decision tree implementation of a boosted strong classifier [Zhou 05] Feature sharing between multiple classifiers Sharing visual features [Torralba et. al 07] VectorBoost [Huang et. al 05] Boosted trees AdaTree [Grossmann 05] BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge3/22 Weak classifier Strong classifier

Brief review of boosting classifier Aggregation of weak learners yields a strong classifier Many variations of learning method and weak classifier functions. Anyboost [Mason et al 00] implementation with discrete decision stumps Weak classifiers: Haar-basis like functions (45,396 in total) BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge4/22 Weak classifier Strong classifier

Brief review of boosting classifier Smooth decision regions BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge5/22

Brief review of decision tree classifier BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge6/ category c split nodes leaf nodes v feature vector v split functions f n ( v ) thresholds t n Classifications P n ( c ) feature vector v split functions f n ( v ) thresholds t n Classifications P n ( c ) ≥ < < ≥ Slide taken and modified from Shotton et. al (2008)

Brief review of decision tree classifier Short classification time BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge7/ category c v ≥ < < ≥

Boosting Classifier vs Decision Tree Preserving (smooth) decision regions for good generalisation Short classification time BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge8/22 Decision treeBoosting

Converting boosting classifier to a decision tree – Super Tree BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge9/22 Boosting Preserving (smooth) decision regions for good generalisation Short classification time Super tree

Boolean optimisation formulation For a learnt boosting classifier split a data space into 2 m primitive regions by m binary weak-learners. Code regions R i i=1,..., 2 m by boolean expressions. BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge10/22BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge W2W2 R3 R5 R6 R1 R2 R4 R7 W1W W3W3 W1W1 W2W2 W3W3 C R1000F R2001F R3010F R4011T R5100T R6101T R7110T R8111X Data space Data space as a boolean table

Boolean optimisation formulation Boolean expression minimisation by optimally joining the regions of the same class label or don’t care label. A short tree built from the minimised boolean expression by placing more frequent variables at the top. BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge11/22 W2W2 R3 R5 R6 R1 R2 R4 R7 W1W W3W3 W1W1 W2W2 W3W3 C R1000F R2001F R3010F R4011T R5100T R6101T R7110T R8111X R1,R W1W1 W2W2 W3W3 TF F T R4 R5,R6,R7,R8 R3 Data space Data space as a boolean table Data space as a tree don’t care

Boolean optimisation formulation Optimally short tree is defined in terms of average expected path length of data points as where region prior p(R i )=M i /M. Constraint: tree must duplicate the decision regions of the boosting classifier BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge12/22

Growing a Super Tree Regions of data points R i taken as input s.t. p(R i )>0 A tree grown by maximising the region information gain Where Key ideas – Growing a tree from the decision regions – Using the region prior (data distribution). BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge13/22 Region prior p Entropy H Weak learner w j Region set R n at node n Region prior p Entropy H Weak learner w j Region set R n at node n

Synthetic data exp1 BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge14/22 Examples generated from GMMs

Synthetic data exp2 BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge15/22 Imbalanced cases

Growing a Super Tree BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge16/22 W1W2W3W4W5SumC Weight Region Boundary region Extended region1x1xx When number of weak learners is relatively large, too many regions of no data points maybe assigned to different class labels from the original ones Solution: Extending regions Modifying information gain: “dont’ care” variable

Face detection experiment Training set: MPEG-7 face data set (11,845 faces) Validation set (for boostrapping): BANCA face set (520 faces) + Caltech background dataset (900 images) Total number: Testing set: MIT+CMU face test set (130 images of 507 faces) 21,780 Harr-like features BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge17/22

Face detection experiment The proposed solution is about 3 to 5 times faster than boosting and 1.5 to 2.8 times faster than [Zhou 05], at the similar accuracy. BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge18/22 BoostingFast Exit [Zhou 05]Super Tree No. of weak learners False positive False negative Average path length False positive False negative Average path length False positive False negative Average path length Total test data points = 57507

Face detection experiment For more than 60 weak-learners a boosting cascade is considered. BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge19/22 Total test data points = BoostingFast Exit [Zhou 05]Super Tree No. of weak learners False positive False negative Average path length False positive False negative Average path length False positive False negative Average path length Fast Exit Cascade No. of weak learners False positive rate False negative rate Average path length Class A Super Tree “Fast Exit” Class A Class B

Experiments with tracking and segmentation by ST BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge20/22

Summary Speeded up boosting classifier without sacrificing accuracy Formalized the problem as a boolean optimization task Proposed a boolean optimisation method for a large number of binary variables (~60) Proposed a 2 stage cascade to handle almost any number of weak learners (binary variables) BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge21/22

Questions? BMVC 2010, Budvytis, Kim, Cipolla, University of Cambridge22/22