Schedule 8 30 -9 00 Introduction 9 00 -10 00 Models: small cliques and special potentials 10 00 -10 30 Tea break 10 30 -12 00 Inference: Relaxation techniques:

Slides:



Advertisements
Similar presentations
POSE–CUT Simultaneous Segmentation and 3D Pose Estimation of Humans using Dynamic Graph Cuts Mathieu Bray Pushmeet Kohli Philip H.S. Torr Department of.
Advertisements

Mean-Field Theory and Its Applications In Computer Vision1 1.
Bayesian Belief Propagation
O BJ C UT M. Pawan Kumar Philip Torr Andrew Zisserman UNIVERSITY OF OXFORD.
The Layout Consistent Random Field for detecting and segmenting occluded objects CVPR, June 2006 John Winn Jamie Shotton.
Constrained Approximate Maximum Entropy Learning (CAMEL) Varun Ganapathi, David Vickrey, John Duchi, Daphne Koller Stanford University TexPoint fonts used.
Agenda Introduction Bag-of-words models Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based.
Learning to Combine Bottom-Up and Top-Down Segmentation Anat Levin and Yair Weiss School of CS&Eng, The Hebrew University of Jerusalem, Israel.
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
I Images as graphs Fully-connected graph – node for every pixel – link between every pair of pixels, p,q – similarity w ij for each link j w ij c Source:
The University of Ontario CS 4487/9687 Algorithms for Image Analysis Multi-Label Image Analysis Problems.
Human Pose detection Abhinav Golas S. Arun Nair. Overview Problem Previous solutions Solution, details.
C. Olsson Higher-order and/or non-submodular optimization: Yuri Boykov jointly with Western University Canada O. Veksler Andrew Delong L. Gorelick C. NieuwenhuisE.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
GrabCut Interactive Image (and Stereo) Segmentation Carsten Rother Vladimir Kolmogorov Andrew Blake Antonio Criminisi Geoffrey Cross [based on Siggraph.
GrabCut Interactive Foreground Extraction using Iterated Graph Cuts Carsten Rother Vladimir Kolmogorov Andrew Blake Microsoft Research Cambridge-UK.
GrabCut Interactive Foreground Extraction using Iterated Graph Cuts Carsten Rother Vladimir Kolmogorov Andrew Blake Microsoft Research Cambridge-UK.
Stephen J. Guy 1. Photomontage Photomontage GrabCut – Interactive Foreground Extraction 1.
1 s-t Graph Cuts for Binary Energy Minimization  Now that we have an energy function, the big question is how do we minimize it? n Exhaustive search is.
Graph-based image segmentation Václav Hlaváč Czech Technical University in Prague Faculty of Electrical Engineering Department of Cybernetics Prague, Czech.
GrabCut Interactive Image (and Stereo) Segmentation Joon Jae Lee Keimyung University Welcome. I will present Grabcut – an Interactive tool for foreground.
Advanced Computer Vision (Module 5F16) Carsten Rother Pushmeet Kohli.
Pseudo-Bound Optimization for Binary Energies Presenter: Meng Tang Joint work with Ismail Ben AyedYuri Boykov 1 / 27.
Simultaneous Segmentation and 3D Pose Estimation of Humans or Detection + Segmentation = Tracking? Philip H.S. Torr Pawan Kumar, Pushmeet Kohli, Matt Bray.
Models for Scene Understanding – Global Energy models and a Style-Parameterized boosting algorithm (StyP-Boost) Jonathan Warrell, 1 Simon Prince, 2 Philip.
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
Optimal solution of binary problems Much material taken from :  Olga Veksler, University of Western Ontario
Robust Higher Order Potentials For Enforcing Label Consistency
P 3 & Beyond Solving Energies with Higher Order Cliques Pushmeet Kohli Pawan Kumar Philip H. S. Torr Oxford Brookes University CVPR 2007.
A Closed Form Solution to Natural Image Matting
TextonBoost : Joint Appearance, Shape and Context Modeling for Multi-Class Object Recognition and Segmentation J. Shotton*, J. Winn†, C. Rother†, and A.
The Layout Consistent Random Field for Recognizing and Segmenting Partially Occluded Objects By John Winn & Jamie Shotton CVPR 2006 presented by Tomasz.
Graph Cut based Inference with Co-occurrence Statistics Ľubor Ladický, Chris Russell, Pushmeet Kohli, Philip Torr.
What Energy Functions Can be Minimized Using Graph Cuts? Shai Bagon Advanced Topics in Computer Vision June 2010.
An Iterative Optimization Approach for Unified Image Segmentation and Matting Hello everyone, my name is Jue Wang, I’m glad to be here to present our paper.
Measuring Uncertainty in Graph Cut Solutions Pushmeet Kohli Philip H.S. Torr Department of Computing Oxford Brookes University.
What, Where & How Many? Combining Object Detectors and CRFs
A Trainable Graph Combination Scheme for Belief Propagation Kai Ju Liu New York University.
Graph-based Segmentation
Image Segmentation Rob Atlas Nick Bridle Evan Radkoff.
Reconstructing Relief Surfaces George Vogiatzis, Philip Torr, Steven Seitz and Roberto Cipolla BMVC 2004.
MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/31/15.
CS774. Markov Random Field : Theory and Application Lecture 08 Kyomin Jung KAIST Sep
MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/24/10.
Planar Cycle Covering Graphs for inference in MRFS The Typhon Algorithm A New Variational Approach to Ground State Computation in Binary Planar Markov.
Lena Gorelick joint work with O. Veksler I. Ben Ayed A. Delong Y. Boykov.
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London.
Associative Hierarchical CRFs for Object Class Image Segmentation Ľubor Ladický 1 1 Oxford Brookes University 2 Microsoft Research Cambridge Based on the.
1 Markov Random Fields with Efficient Approximations Yuri Boykov, Olga Veksler, Ramin Zabih Computer Science Department CORNELL UNIVERSITY.
Dynamic Tree Block Coordinate Ascent Daniel Tarlow 1, Dhruv Batra 2 Pushmeet Kohli 3, Vladimir Kolmogorov 4 1: University of Toronto3: Microsoft Research.
Efficient Discriminative Learning of Parts-based Models M. Pawan Kumar Andrew Zisserman Philip Torr
Associative Hierarchical CRFs for Object Class Image Segmentation
O BJ C UT M. Pawan Kumar Philip Torr Andrew Zisserman UNIVERSITY OF OXFORD.
Tractable Higher Order Models in Computer Vision (Part II) Slides from Carsten Rother, Sebastian Nowozin, Pusohmeet Khli Microsoft Research Cambridge Presented.
Jigsaws: joint appearance and shape clustering John Winn with Anitha Kannan and Carsten Rother Microsoft Research, Cambridge.
CS654: Digital Image Analysis Lecture 28: Advanced topics in Image Segmentation Image courtesy: IEEE, IJCV.
TextonBoost: Joint Appearance, Shape and Context Modeling for Multi-Class Object Recognition and Segmentation J. Shotton ; University of Cambridge J. Jinn,
ICCV 2007 National Laboratory of Pattern Recognition Institute of Automation Chinese Academy of Sciences Half Quadratic Analysis for Mean Shift: with Extension.
Image segmentation.
Holistic Scene Understanding Virginia Tech ECE /02/26 Stanislaw Antol.
MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/27/12.
Markov Random Fields Tomer Michaeli Graduate Course
Segmentation of Building Facades using Procedural Shape Priors
Learning a Region-based Scene Segmentation Model
GrabCut Interactive Foreground Extraction using Iterated Graph Cuts Carsten Rother Vladimir Kolmogorov Andrew Blake Microsoft Research Cambridge-UK.
Markov Random Fields with Efficient Approximations
Learning to Combine Bottom-Up and Top-Down Segmentation
“The Truth About Cats And Dogs”
Expectation-Maximization & Belief Propagation
“Traditional” image segmentation
Presentation transcript:

Schedule Introduction Models: small cliques and special potentials Tea break Inference: Relaxation techniques: LP, Lagrangian, Dual Decomposition Models: global potentials and global parameters + discussion

MRF with global potential GrabCut model [Rother et. al. ‘04] F i = -log Pr(z i |θ F )B i = -log Pr(z i |θ B ) Background Foreground G R θ F/B Gaussian Mixture models E(x,θ F,θ B ) = Problem: for unknown x,θ F,θ B the optimization is NP-hard! [Vicente et al. ‘09] Image z Output x ∑ F i (θ F )x i + B i (θ B )(1-x i )+ ∑ |x i -x j | i,j Є N i θ F/B

GrabCut: Iterated Graph Cuts [Rother et al. Siggraph ‘04] Learning of the colour distributions Graph cut to infer segmentation F x min E(x, θ F, θ B ) θF,θBθF,θB B Most systems with global variables work like that e.g. [ObjCut Kumar et. al. ‘05, PoseCut Bray et al. ’06, LayoutCRF Winn et al. ’06] θ F/B

GrabCut: Iterated Graph Cuts Energy after each Iteration Result Guaranteed to converge

Colour Model Background Foreground & Background G R Background Foreground G R Iterated graph cut

Optimizing over θ’s help after convergence [GrabCut ‘04] no iteration [Boykov&Jolly ‘01] Input after convergence [GrabCut ‘04]

Global optimality? GrabCut (local optimum) Global Optimum [Vicente et al. ‘09] Is it a problem of the optimization or the model?

… first attempt to solve it [Lempisky et al. ECCV ‘08] Model a discrete subset: w F = (1,1,0,1,0,0,0,0); w B = (1,0,0,0,0,0,0,1) #solutions: w F *w B = 2 16 Global Optimum: Exhaustive Search: Graph Cuts Branch-and-MinCut: ~ Graph Cuts (depends on image) E(x,θ F,θ B )= ∑ F i (θ F )x i + B i (θ B )(1-x i ) + ∑ w ij |x i -x j | 8 Gaussians whole image G R w F,B

Branch-and-MinCut w F = (1,1,1,1,0,0,0,0) w B = (1,0,1,1,0,1,0,0) w F = (0,0,*,*,*,*,*,*) w B = (0,*,*,*,*,*,*,*) w F = (*,*,*,*,*,*,*,*) w B = (*,*,*,*,*,*,*,*) min E(x,w F,w B ) = min [ ∑ F i (w F )x i + B i (w B )(1-x i ) + ∑ w ij (x i,x j ) ] ≥ min [∑ min F i (w F )x i + min B i (w B )(1-x i ) + ∑ w ij (x i,x j )] x,w F,w B x wBwB wFwF

Results … E=-618 GrabCut E=-624 (speed-up 481) Branch-and-MinCut E=-593 GrabCut E=-584 (speed-up 141) Branch-and-MinCut

Object Recognition & Segmentation Given exemplar shapes: Test: Speed-up ~900; accuracy 98.8% |w| ~ min E(x,w) with: w = Templates x Position w

… second attempt to solve it [Vicente et al. ICCV ‘09] Eliminate global color model θ F,θ B : θF,θBθF,θB E’(x) = min E(x,θ F,θ B )

Eliminate color model E(x,θ F,θ B )= ∑ F i (θ F )x i + B i (θ B )(1-x i ) + ∑ w ij |x i -x j | Image histogram k given x K = 16 3 k θBθB θ F є [0,1] K is a distributions (∑ θ F = 1) (background same) θFθF k background distribution foreground distribution Optimal θ F/B given by empirical histograms: θ F = n F k /n F n F = ∑x i #fgd. pixel n F = ∑x i #fgd. pixel in bin k Optimal θ F/B given by empirical histograms: θ F = n F k /n F n F = ∑x i #fgd. pixel n F = ∑x i #fgd. pixel in bin k k k Image discretized in bins K i Є B k K K

Eliminate color model E’(x)= g(n F ) + ∑ h k (n F ) + ∑ w ij |x i -x j | with n F = ∑x i, n F = ∑x i min θ F,θ B k k E(x,θ F,θ B )= ∑ F i (θ F )x i + B i (θ B )(1-x i ) + ∑ w ij |x i -x j | k nFnF 0n/2n g Prefers “equal area” segmentation Each color either fore- or background hkhk nFnF 0 max k convex concave i Є B k i E(x,θ F,θ B )= ∑ -n F k log θ F k -n B k log θ B k + ∑ w ij |x i -x j | k (θ F = n F k /n F ) K

How to optimize … Dual Decomposition E(x)= g(n F ) + ∑ h k (n F k ) + ∑ w ij |x i -x j | E 1 (x)E 2 (x) min E(x) = min [ E 1 (x) + y T x + E 2 (x) – y T x ] ≥ min [ E 1 (x’) + y T x’ ] + min [E 2 (x) – y T x] =: L(y) Goal: - maximize concave function L(y) using sub-gradient - no guarantees on E (NP-hard) L(y) E(x’) k x’xxx “paramteric maxflow” gives optimal y=λ1 efficiently [Vicente et al. ICCV ‘09] Simple (no MRF) Robust P n Potts

Some results… Global optimum in 61% of cases (GrabCut database) InputGrabCut Global Optimum (DD) Local Optimum (DD)

Insights on the GrabCut model g 0.4 g0.3 g1.5 g hkhk nFnF Each color either fore- or background 0 max k nFnF 0n/2n g Prefers “equal area” segmentation concave convex

Relationship to Soft P n Potts Image Pairwise CRF only TextonBoost [Shotton et al. ‘06] robust P n Potts [Kohli et al ‘08] One super- pixelization another super- pixelization GrabCut: cluster all colors together Just different type of clustering:

Marginal Probability Field (MPF) What is the prior of a MAP-MRF solution: [Woodford et. al. ICCV ‘09] Training image: 60% black, 40% white MRF is a bad prior since ignores shape of the (feature) distribution ! MAP: prior(x) = 0.6 = Others less likely : prior(x) = 0.6 * 0.4 = Introduce a global term, which controls global statistic

Marginal Probability Field (MPF) [Woodford et. al. ICCV ‘09] Optimization done with Dual Decomposition (different ones) max 0 0 MRF True energy

Examples Segmentation: In-painting: Pairwise MRF – Increase Prior strength Ground truth Noisy input Global gradient prior

Schedule Introduction Models: small cliques and special potentials Tea break Inference: Relaxation techniques: LP, Lagrangian, Dual Decomposition Models: global potentials and global parameters + discussion

Open Questions Many exciting future directions – Exploiting latest ideas for applications (object recognition etc.) – Many other higher-order cliques: Topology, Grammars, etc. (this conference). Comparison of inference techniques needed: – Factor graph message passing vs. transformation vs. LP relaxation? Learning higher order Random Fields