Efficient Discriminative Learning of Parts-based Models M. Pawan Kumar Andrew Zisserman Philip Torr

Slides:



Advertisements
Similar presentations
POSE–CUT Simultaneous Segmentation and 3D Pose Estimation of Humans using Dynamic Graph Cuts Mathieu Bray Pushmeet Kohli Philip H.S. Torr Department of.
Advertisements

Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London Tutorial at GDR (Optimisation Discrète, Graph Cuts.
O BJ C UT M. Pawan Kumar Philip Torr Andrew Zisserman UNIVERSITY OF OXFORD.
Solving Markov Random Fields using Second Order Cone Programming Relaxations M. Pawan Kumar Philip Torr Andrew Zisserman.
Support Vector Machine
Linear Time Methods for Propagating Beliefs Min Convolution, Distance Transforms and Box Sums Daniel Huttenlocher Computer Science Department December,
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Ľubor Ladický1 Phil Torr2 Andrew Zisserman1
Support vector machine
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
Many slides based on P. FelzenszwalbP. Felzenszwalb General object detection with deformable part-based models.
More sliding window detection: Discriminative part-based models Many slides based on P. FelzenszwalbP. Felzenszwalb.
Restrict learning to a model-dependent “easy” set of samples General form of objective: Introduce indicator of “easiness” v i : K determines threshold.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Robust Higher Order Potentials For Enforcing Label Consistency
1-norm Support Vector Machines Good for Feature Selection  Solve the quadratic program for some : min s. t.,, denotes where or membership. Equivalent.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Schedule Introduction Models: small cliques and special potentials Tea break Inference: Relaxation techniques:
P 3 & Beyond Solving Energies with Higher Order Cliques Pushmeet Kohli Pawan Kumar Philip H. S. Torr Oxford Brookes University CVPR 2007.
An Introduction to Kernel-Based Learning Algorithms K.-R. Muller, S. Mika, G. Ratsch, K. Tsuda and B. Scholkopf Presented by: Joanna Giforos CS8980: Topics.
Improved Moves for Truncated Convex Models M. Pawan Kumar Philip Torr.
Efficiently Solving Convex Relaxations M. Pawan Kumar University of Oxford for MAP Estimation Philip Torr Oxford Brookes University.
Constrained Optimization Rong Jin. Outline  Equality constraints  Inequality constraints  Linear Programming  Quadratic Programming.
Object Recognizing We will discuss: Features Classifiers Example ‘winning’ system.
Exploiting Duality (Particularly the dual of SVM) M. Pawan Kumar VISUAL GEOMETRY GROUP.
Support Vector Machines and Kernel Methods
Learning to Segment from Diverse Data M. Pawan Kumar Daphne KollerHaithem TurkiDan Preston.
Hierarchical Graph Cuts for Semi-Metric Labeling M. Pawan Kumar Joint work with Daphne Koller.
Invariant Large Margin Nearest Neighbour Classifier M. Pawan Kumar Philip Torr Andrew Zisserman.
Optimization Theory Primal Optimization Problem subject to: Primal Optimal Value:
MAP Estimation Algorithms in M. Pawan Kumar, University of Oxford Pushmeet Kohli, Microsoft Research Computer Vision - Part I.
Latent Boosting for Action Recognition Zhi Feng Huang et al. BMVC Jeany Son.
Probabilistic Inference Lecture 4 – Part 2 M. Pawan Kumar Slides available online
“Study on Parallel SVM Based on MapReduce” Kuei-Ti Lu 03/12/2015.
计算机学院 计算感知 Support Vector Machines. 2 University of Texas at Austin Machine Learning Group 计算感知 计算机学院 Perceptron Revisited: Linear Separators Binary classification.
Learning a Small Mixture of Trees M. Pawan Kumar Daphne Koller Aim: To efficiently learn a.
Lecture 31: Modern recognition CS4670 / 5670: Computer Vision Noah Snavely.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Discrete Optimization Lecture 4 – Part 2 M. Pawan Kumar Slides available online
Discrete Optimization in Computer Vision M. Pawan Kumar Slides will be available online
Discrete Optimization Lecture 3 – Part 1 M. Pawan Kumar Slides available online
Probabilistic Inference Lecture 5 M. Pawan Kumar Slides available online
Update any set S of nodes simultaneously with step-size We show fixed point update is monotone for · 1/|S| Covering Trees and Lower-bounds on Quadratic.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Biointelligence Laboratory, Seoul National University
Linear Models for Classification
O BJ C UT M. Pawan Kumar Philip Torr Andrew Zisserman UNIVERSITY OF OXFORD.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Discrete Optimization Lecture 2 – Part 2 M. Pawan Kumar Slides available online
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Inference for Learning Belief Propagation. So far... Exact methods for submodular energies Approximations for non-submodular energies Move-making ( N_Variables.
Probabilistic Inference Lecture 2 M. Pawan Kumar Slides available online
Maximum Entropy Discrimination Tommi Jaakkola Marina Meila Tony Jebara MIT CMU MIT.
Discrete Optimization Lecture 1 M. Pawan Kumar Slides available online
Efficient Belief Propagation for Image Restoration Qi Zhao Mar.22,2006.
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 1. Stat 231. A.L. Yuille. Fall 2004 Linear Separation and Margins. Non-Separable and.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
MAP Estimation of Semi-Metric MRFs via Hierarchical Graph Cuts M. Pawan Kumar Daphne Koller Aim: To obtain accurate, efficient maximum a posteriori (MAP)
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Linear Discriminant Functions Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Support vector machines
Introduction of BP & TRW-S
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
Learning a Region-based Scene Segmentation Model
Efficiently Selecting Regions for Scene Understanding
Support Vector Machines Introduction to Data Mining, 2nd Edition by
Support vector machines
MAP Estimation of Semi-Metric MRFs via Hierarchical Graph Cuts
Support vector machines
Support vector machines
Presentation transcript:

Efficient Discriminative Learning of Parts-based Models M. Pawan Kumar Andrew Zisserman Philip Torr Aim: To efficiently learn parts-based models which discriminate between positive and negative poses of the object category Results - Sign Language Efficient Reformulation Results - Buffy Parts-based Model G = (V, E)Restricted to Tree The Learning Problem Q(f) = ∑ Q a (f(a)) + ∑ Q ab (f(a), f(b)) f : V Pose of V (h values) Q a (f(a)) : Unary potential for f(a) Computed using features Q ab (f(a), f(b)): Pairwise potential for validity of (f(a),f(b)) Restricted to Potts Q a (f(a)) : w a T  (f(a))Q a (f(a),f(b)) : w ab T  (f(a),f(b)) Q(f) : w T  (f) min ||w|| + C∑  i w T  (f + i ) +  ≥ 1 -  + i w T  (f - ij ) +  ≤ -1 +  - i Maximize margin, minimize hinge loss High energy for all positive examples Low energy for all negative examples Related Work Local Iterative Support Vector Machine (ISVM-1) Start with a small subset of negative examples (1 per image) Solve for w and b Replace negative examples with current MAP estimates Converges to local optimum Start with a small subset of negative examples (1 per image) Solve for w and b Add current MAP estimates to set of negative examples Converges to global optimum Global Iterative Support Vector Machine (ISVM-2) Drawback: Requires obtaining MAP estimate of each image at each iteration (computationally expensive) Our: 86.4% Buehler et al.,2008: 87.7% Our: 39.2% Ferrari et al.,2008: 41.0% 100 training images, 95 test images ISVM-1 ISVM-2 Our ISVM-1 ISVM-2 Our 196 training images, 204 test images ISVMs run for twice as long For all j (exponential in |V|) = 1, if (f(a),f(b))  L ab, = 0, otherwise. b a w T  (f - ij ) +  ≤ -1 +  - i, for all j M i ba (k) ≥ w b  b (l), for all l M i ba (k) ≥ w b  b (l) + w ab, for all (k,l)  L ab w a  a (k) + ∑ b M i ba (k) +  ≤ -1 +  - i Exponential in |V| Linear in |V| Linear in h Linear in |L ab | b a b a max  ab T 1 -  ab T K ab  ab s.t.  ab T y = 0,  ab ≥ 0 0 ≤ ∑  i ab (k) + ∑  i ab (k,l) ∑  i ab (k) + ∑  i ab (k,l) ≤ C Problem (1) ∑ k  i ab (k) = ∑ l  i ba (l) Constraint (3)Results in a large minimal problem Dual Decomposition Master Problem(1)Problem (2) minimal problem size = 2 Update Lagrange multiplier of (3) SVM-like problems Modified SVM Light min ∑ i g i (x), subject to x  P min ∑ g i (x i ), s.t. x i  P, x i = x max min ∑ g i (x i ) + i (x i - x), s.t. x i  P KKT Condition: ∑ i = 0 Solve min ∑ g i (x i ) + i x i i = i +  x i * Project Problem (1) learns the unary weight vector w a and pairwise weight w ab Problem (2) learns the unary weight vector w b and pairwise weight w ab M i ba (k) analogous to messages in Belief Propagation (BP) Efficient BP using distance transform: Felzenszwalb and Huttenlocher, 2004 Solving the Dual Implementation Details FeaturesShape: HOG Appearance: (x,x 2 ), x = fraction of skin pixels DataPositive examples: Provided by user Negative examples: All other poses OcclusionEach putative pose can be occluded (twice the number of labels)  (f(a),f(b)) max  ba T 1 -  ba T K ba  ba s.t.  ba T y = 0,  ba ≥ 0 0 ≤ ∑  i ba (k) + ∑  i ba (k,l) ∑  i ba (k) + ∑  i ba (k,l) ≤ C Problem (2)