Boris Babenko 1, Ming-Hsuan Yang 2, Serge Belongie 1 1. University of California, San Diego 2. University of California, Merced OLCV, Kyoto, Japan.

Slides:



Advertisements
Similar presentations
Longin Jan Latecki Temple University
Advertisements

Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Online Multiple Classifier Boosting for Object Tracking Tae-Kyun Kim 1 Thomas Woodley 1 Björn Stenger 2 Roberto Cipolla 1 1 Dept. of Engineering, University.
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Separating Hyperplanes
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei Li,
Lecture 4: Embedded methods
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Lecture 14 – Neural Networks
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Sparse vs. Ensemble Approaches to Supervised Learning
Support Vector Machines for Multiple- Instance Learning Authors: Andrews, S.; Tsochantaridis, I. & Hofmann, T. (Advances in Neural Information Processing.
Chapter 6: Multilayer Neural Networks
Region Based Image Annotation Through Multiple-Instance Learning By: Changbo Yang Wayne State University Department of Computer Science.
Artificial Neural Networks
Collaborative Filtering Matrix Factorization Approach
Neural Networks Lecture 8: Two simple learning algorithms
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Outline Classification Linear classifiers Perceptron Multi-class classification Generative approach Naïve Bayes classifier 2.
Visual Tracking with Online Multiple Instance Learning
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
Machine Learning Chapter 4. Artificial Neural Networks
Recognition using Boosting Modified from various sources including
Mathematical formulation XIAO LIYING. Mathematical formulation.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Selective Block Minimization for Faster Convergence of Limited Memory Large-scale Linear Models Kai-Wei Chang and Dan Roth Experiment Settings Block Minimization.
Benk Erika Kelemen Zsolt
Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.
BOOSTING David Kauchak CS451 – Fall Admin Final project.
Boris 2 Boris Babenko 1 Ming-Hsuan Yang 2 Serge Belongie 1 (University of California, Merced, USA) 2 (University of California, San Diego, USA) Visual.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Ensemble Methods: Bagging and Boosting
Extending the Multi- Instance Problem to Model Instance Collaboration Anjali Koppal Advanced Machine Learning December 11, 2007.
Linear Discrimination Reading: Chapter 2 of textbook.
Non-Bayes classifiers. Linear discriminants, neural networks.
Robust Object Tracking with Online Multiple Instance Learning
Non-Isometric Manifold Learning Analysis and an Algorithm Piotr Dollár, Vincent Rabaud, Serge Belongie University of California, San Diego.
Online Learning Rong Jin. Batch Learning Given a collection of training examples D Learning a classification model from D What if training examples are.
Boris Babenko, Steve Branson, Serge Belongie University of California, San Diego ICCV 2009, Kyoto, Japan.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
ECE 5984: Introduction to Machine Learning Dhruv Batra Virginia Tech Topics: –Ensemble Methods: Bagging, Boosting Readings: Murphy 16.4; Hastie 16.
1 Predictive Learning from Data Electrical and Computer Engineering LECTURE SET 5 Nonlinear Optimization Strategies.
Neural Networks The Elements of Statistical Learning, Chapter 12 Presented by Nick Rizzolo.
Learning by Loss Minimization. Machine learning: Learn a Function from Examples Function: Examples: – Supervised: – Unsupervised: – Semisuprvised:
Page 1 CS 546 Machine Learning in NLP Review 2: Loss minimization, SVM and Logistic Regression Dan Roth Department of Computer Science University of Illinois.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
Strong Supervision from Weak Annotation: Interactive Training of Deformable Part Models S. Branson, P. Perona, S. Belongie.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Neural networks and support vector machines
Predictive Learning from Data
Multilayer Perceptrons
Dan Roth Department of Computer and Information Science
Restricted Boltzmann Machines for Classification
Boosting and Additive Trees
A Simple Artificial Neuron
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Classification with Perceptrons Reading:
Asymmetric Gradient Boosting with Application to Spam Filtering
Neural Networks and Backpropagation
Machine Learning Today: Reading: Maria Florina Balcan
Goodfellow: Chap 6 Deep Feedforward Networks
Disadvantages of Discrete Neurons
Collaborative Filtering Matrix Factorization Approach
CSCI B609: “Foundations of Data Science”
Logistic Regression & Parallel SGD
Predictive Learning from Data
Softmax Classifier.
Boris Babenko, Steve Branson, Serge Belongie
Presentation transcript:

Boris Babenko 1, Ming-Hsuan Yang 2, Serge Belongie 1 1. University of California, San Diego 2. University of California, Merced OLCV, Kyoto, Japan

Extending online boosting beyond supervised learning Some algorithms exist (i.e. MIL, Semi- Supervised), but would like a single framework [Oza ‘01, Grabner et al. ‘06, Grabner et al. ‘08, Babenko et al. ‘09]

Goal: learn a strong classifier where is a weak classifier, and is the learned parameter vector

Have some loss function Have Find next weak classifier:

Find some parameter vector that optimizes loss

If loss over entire training data can be split into sum of loss per training example can use the following update:

Recall, we want to solve What if we use stochastic gradient descent to find ?

For any differentiable loss function, can derive boosting algorithm…

Loss: Update rule:

Training data: bags of instances and bag labels Bag is positive if at least one member is positive

Loss: where [Viola et al. ‘05]

Update rule:

So far, only empirical results Compare – OSB – BSB – standard batch boosting algorithm – Linear & non-linear model trained with stochastic gradient descent (BSB with M=1)

[LeCun et al. 98, Kanade et al. ‘00, Huang et al. ‘07

[UCI Repository, Ranganathan et al. ‘08]

LeCun et al. ‘97, Andrews et al ‘02

Friedman’s “Gradient Boosting” framework = gradient descent in function space – OSB = gradient descent in parameter space Similar to Neural Net methods (i.e. Ash et al. ‘89)

Advantages: – Easy to derive new Online Boosting algorithms for various problems / loss functions – Easy to implement Disadvantages: – No theoretic guarantees yet – Restricted class of weak learners

Research supported by: – NSF CAREER Grant # – NSF IGERT Grant DGE – ONR MURI Grant #N