Efficient Graph Cut Optimization for Full CRFs with Quantized Edges

Slides:



Advertisements
Similar presentations
Primal-dual Algorithm for Convex Markov Random Fields Vladimir Kolmogorov University College London GDR (Optimisation Discrète, Graph Cuts et Analyse d'Images)
Advertisements

Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London Tutorial at GDR (Optimisation Discrète, Graph Cuts.
Bayesian Belief Propagation
Mean-Field Theory and Its Applications In Computer Vision2
Generating Classic Mosaics with Graph Cuts Y. Liu, O. Veksler and O. Juan University of Western Ontario, Canada Ecole Centrale de Paris, France.
Pose Estimation and Segmentation of People in 3D Movies Karteek Alahari, Guillaume Seguin, Josef Sivic, Ivan Laptev Inria, Ecole Normale Superieure ICCV.
Graph Cut Algorithms for Computer Vision & Medical Imaging Ramin Zabih Computer Science & Radiology Cornell University Joint work with Y. Boykov, V. Kolmogorov,
Lena Gorelick Joint work with Frank Schmidt and Yuri Boykov Rochester Institute of Technology, Center of Imaging Science January 2013 TexPoint fonts used.
Introduction to Markov Random Fields and Graph Cuts Simon Prince
Scene Labeling Using Beam Search Under Mutex Constraints ID: O-2B-6 Anirban Roy and Sinisa Todorovic Oregon State University 1.
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
I Images as graphs Fully-connected graph – node for every pixel – link between every pair of pixels, p,q – similarity w ij for each link j w ij c Source:
The University of Ontario CS 4487/9687 Algorithms for Image Analysis Multi-Label Image Analysis Problems.
Efficient Inference for Fully-Connected CRFs with Stationarity
Proportion Priors for Image Sequence Segmentation Claudia Nieuwenhuis, etc. ICCV 2013 Oral.
1 s-t Graph Cuts for Binary Energy Minimization  Now that we have an energy function, the big question is how do we minimize it? n Exhaustive search is.
Graph-based image segmentation Václav Hlaváč Czech Technical University in Prague Faculty of Electrical Engineering Department of Cybernetics Prague, Czech.
Image Matting and Its Applications Chen-Yu Tseng Advisor: Sheng-Jyh Wang
Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)
1 Can this be generalized?  NP-hard for Potts model [K/BVZ 01]  Two main approaches 1. Exact solution [Ishikawa 03] Large graph, convex V (arbitrary.
Robust Higher Order Potentials For Enforcing Label Consistency
P 3 & Beyond Solving Energies with Higher Order Cliques Pushmeet Kohli Pawan Kumar Philip H. S. Torr Oxford Brookes University CVPR 2007.
Improved Moves for Truncated Convex Models M. Pawan Kumar Philip Torr.
2010/5/171 Overview of graph cuts. 2010/5/172 Outline Introduction S-t Graph cuts Extension to multi-label problems Compare simulated annealing and alpha-
Stereo & Iterative Graph-Cuts Alex Rav-Acha Vision Course Hebrew University.
Graph Cut based Inference with Co-occurrence Statistics Ľubor Ladický, Chris Russell, Pushmeet Kohli, Philip Torr.
Stereo Computation using Iterative Graph-Cuts
Comp 775: Graph Cuts and Continuous Maximal Flows Marc Niethammer, Stephen Pizer Department of Computer Science University of North Carolina, Chapel Hill.
What Energy Functions Can be Minimized Using Graph Cuts? Shai Bagon Advanced Topics in Computer Vision June 2010.
Compact Windows for Visual Correspondence via Minimum Ratio Cycle Algorithm Olga Veksler NEC Labs America.
Measuring Uncertainty in Graph Cut Solutions Pushmeet Kohli Philip H.S. Torr Department of Computing Oxford Brookes University.
Graph-Cut Algorithm with Application to Computer Vision Presented by Yongsub Lim Applied Algorithm Laboratory.
Computer vision: models, learning and inference
Perceptual Organization: Segmentation and Optical Flow.
Extensions of submodularity and their application in computer vision
Automatic User Interaction Correction via Multi-label Graph-cuts Antonio Hernández-Vela, Carlos Primo and Sergio Escalera Workshop on Human Interaction.
Graph-based Segmentation
Image Renaissance Using Discrete Optimization Cédric AllèneNikos Paragios ENPC – CERTIS ESIEE – A²SI ECP - MAS France.
MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/24/10.
Graph Cut Algorithms for Binocular Stereo with Occlusions
City University of Hong Kong 18 th Intl. Conf. Pattern Recognition Self-Validated and Spatially Coherent Clustering with NS-MRF and Graph Cuts Wei Feng.
CS774. Markov Random Field : Theory and Application Lecture 13 Kyomin Jung KAIST Oct
Planar Cycle Covering Graphs for inference in MRFS The Typhon Algorithm A New Variational Approach to Ground State Computation in Binary Planar Markov.
Interactive Graph Cuts for Optimal Boundary & Region Segmentation of Objects in N-D Images (Fri) Young Ki Baik, Computer Vision Lab.
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
Rounding-based Moves for Metric Labeling M. Pawan Kumar Center for Visual Computing Ecole Centrale Paris.
Lena Gorelick joint work with O. Veksler I. Ben Ayed A. Delong Y. Boykov.
Graph Cuts Marc Niethammer. Segmentation by Graph-Cuts A way to compute solutions to the optimization problems we looked at before. Example: Binary Segmentation.
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London.
1 Markov Random Fields with Efficient Approximations Yuri Boykov, Olga Veksler, Ramin Zabih Computer Science Department CORNELL UNIVERSITY.
Lecture 19: Solving the Correspondence Problem with Graph Cuts CAP 5415 Fall 2006.
Presenter : Kuang-Jui Hsu Date : 2011/3/24(Thur.).
Tractable Higher Order Models in Computer Vision (Part II) Slides from Carsten Rother, Sebastian Nowozin, Pusohmeet Khli Microsoft Research Cambridge Presented.
The University of Ontario CS 4487/9687 Algorithms for Image Analysis Multi-Label Image Analysis Problems.
Gaussian Mixture Models and Expectation-Maximization Algorithm.
Machine Learning – Lecture 15
CS654: Digital Image Analysis Lecture 28: Advanced topics in Image Segmentation Image courtesy: IEEE, IJCV.
A global approach Finding correspondence between a pair of epipolar lines for all pixels simultaneously Local method: no guarantee we will have one to.
Energy functions f(p)  {0,1} : Ising model Can solve fast with graph cuts V( ,  ) = T[  ] : Potts model NP-hard Closely related to Multiway Cut Problem.
Image segmentation.
Gaussian Conditional Random Field Network for Semantic Segmentation
Energy minimization Another global approach to improve quality of correspondences Assumption: disparities vary (mostly) smoothly Minimize energy function:
Markov Random Fields Tomer Michaeli Graduate Course
Towards Globally Optimal Normal Orientations for Large Point Clouds
Markov Random Fields with Efficient Approximations
Nonparametric Semantic Segmentation
Exact Inference Continued
Tractable MAP Problems
Automatic User Interaction Correction via Multi-label Graph-cuts
“Traditional” image segmentation
Presentation transcript:

Efficient Graph Cut Optimization for Full CRFs with Quantized Edges Olga Veksler

Contents Introduction Quantized edge Full-CRF CRF with Potts pairwise potentials sparsely connected fully connected CRF (Full-CRF) Gaussian edge weights mean field optimization Quantized edge Full-CRF optimization two label case multi-label case connection to Gaussian Edge CRF application to semantic segmentation

CRF Energy with Potts Potentials high energy low energy Find labeling x minimizing energy Optimization Solved exactly in binary label case with a graph cut NP hard in multi-label case expansion algorithm approximation (factor of 2) [Boykov et.al.’TPAMI01]

Sparse vs. Fully Connected CRF Sparsely connected CRFs 4, 8, or small neighbourhood connected TRWS [Kolmogorov ‘TPAMI2006] or expansion algorithms work well length regularization [Boykov&Kolmogorov’ICCV2003] Fully connected CRFs [Krahenbul&Koltun’NIPS2011] all pixels are neighbors, n pixels, O(n2) edges naïve application of expansion algorithm, TRWS, etc. is not efficient regularization properties?

Full CRF with Uniform Weights Cardinality regularization labels in {0,1} n pixels in the image, k pixels assigned to label 1

Full CRF with Uniform Weights Cardinality regularization labels in {0,1} n pixels in the image, k pixels assigned to label 1 pairwise energy is

Full CRF with Uniform Weights Cardinality regularization labels in {0,1} n pixels in the image, k pixels assigned to label 1 pairwise energy is w · (n - k) · k k n same pairwise cost Efficient algorithm for each k find the k pixels with lowest cost for label 1 compute total energy chose k corresponding to the smallest energy

Fully Connected CRFs Fully connected CRFs [Krahenbul&Koltun’NIPS2011] assumes Gaussian edge weights efficient mean field inference approximate bilateral filter [Paris&Durand, IJCV’2009] mean field is not a very effective optimization method [Weiss’2001]

CRF+CNN Combination CNN can give blurred not pixel precise results Sharpen with CRF Chen et.al. ICLR’2015 as post processing Or unified framework [Zheng et.al., ICCV’2015]

Quantized Edge Fully Connected CRFs Gaussian Edge Weights [Krahenbul&Koltun’NIPS2011] d superpixels Quantized edge weights m

Quantized Edge Fully Connected CRFs Edge weights depend on superpixel membership do not have to be Gaussian weighted superpixels

Quantized Edge Fully Connected CRFs input image superpixels Interior/exterior weights interior weights exterior weights

Optimization for 2 labels: Superpixel Consider one superpixel internal edges weight w n – k pixels 1 w k pixels w Internal pairwise cost is k·(n - k) w· if vary green superpixel labeling, cost changes only with k

Optimization for 2 labels: Two Superpixels Consider two superpixels external edge weight ww 1 k pixels n – k pixels 1 h pixels m – h pixels ww ww External pairwise cost is k·(m - h) + (n - k) ·h ww · [ ] if vary green superpixel labeling, cost changes only with k If k pixels in green superpixel are assigned to label 1, they must be those that have the smallest cost for label 1

Optimization for 2 labels: Overview Convert binary energy in pixel domain to multi-label energy in smaller superpixel domain new variables are the superpixels new cardinality labels are 0,1,…,superpixelSize assume unary cost for label 0 is 0 old labels {0,1} 8 pixels 4 pixels new variables new labels {0,...,4} {0,...,10} {0,...,8} {0,...,11} 10 pixels 11 pixels

Optimization for 2 labels: Conversion Sort pixels in each superpixel by increasing cost of being assigned to label 1 New variables are the superpixels New labels are 0,1,…,superpixelSize Label k assigned to superpixel means k smallest cost pixels in that superpixel are assigned to label 1 in original problem 111 1 1 1 3 1 sort pixels in each superpixel by cost of being assigned to label 1 1 4 2 s1=2 s2=3 s3=6 s4=11

Unary Cost for Transformed Problem Unary cost for green superpixel to have label k account for unary terms of the original binary problem 1 k pixels n – k pixels w

Unary Cost for Transformed Problem Unary cost for green superpixel to have label k account for unary terms of the original binary problem account for internal pairwise terms of original binary problem k pixels n – k pixels w cost of 1 cost of 1 + w·k·(n - k) cost of 1 cost of 1

Pairwise Cost for Transformed Problem green superpixel to have label k, purple superpixel to have label h models external pairwise cost 1 k pixels n – k pixels h pixels m – h pixels ww k·(m - h) + (n - k) ·h ww · [ ]

Optimization for Transformed Problem Pairwise cost for green superpixel to have label k, purple superpixel to have label h k·(m - h) + (n - k) ·h ww · [ ] Can be rewritten as unary terms + (h - k)2 Can optimize exactly with [Ishikawa’TPAMI04] number of edges is quadratic in the number of labels memory inefficient, time complexity almost as bad as the original binary problem Or with [Ajanthan’CVPR2016] Memory efficient, but time complexity almost as bad as the original binary problem

Optimization: Jump Moves Pairwise cost is quadratic (h - k)2 5 4 3 2 1 7 5 3 2 4 7 +1 jump -1 jump 5 2 1 3 4 7 Jump moves [Veksler’99, Kolmogorov &Shioura’09] each move is optimization of binary energy efficient: number of edges is linear in the number of pixels give exact minimum efficiently if unary terms are also convex Our unary terms are not convex jump moves do not work well in practice

Optmization: Expansion Moves 5 4 3 2 1 7 5 2 3 4 7 2-expansion 5 2 1 7 1 - expansion Expansion moves [Boykov et.al., PAMI’2001] each expansion move is optimization of binary energy efficient: number of edges is linear in the number of pixels Not submodular for quadratic potential but does find the optimum in the overwhelming majority of cases

Multi-Label Quantized Full-CRF Apply expansion algorithm each expansion step is optimization of binary energy already know how to optimize 2-label Edge Quantized Full CRF problem meaning of label 0 is not fixed for expansion algorithm solution construct new superpixels according to the current labeling β α γ ε δ old superpixels new superpixels

Final Algorithm, Multi-Label Case for each α ∊ L perform α-expansion 1. compute new superpixels 2. transform binary expansion energy from pixel domain to multi-label energy in superpixel domain 3. for each β ∊ Ltransformed perform β -expansion until convergence

Connection to Gaussian Full-CRF Quantized edge CRF gets close to Gaussian edge CRF as number of superpixels increases as beta increases

Connection to Gaussian Full-CRF Regularization properties of full Gaussian CRF not well understood “all pixels connected”, “preserves fine structure” ground truth Gaussian Full-CRF results Quantized Edge CRF model helps to understand Gaussian CRF If k pixels in a superpixel split from the rest, shape of the split does not matter equal cost labelings

Optimization Results: Full-CRF, 2 labels validation fold of Pascal 2012 dataset reduced to 70x70 pixels 2 most likely labels global optimum with a graph cut our method is exact in 89% of cases running time in seconds

Optimization Results: Full-CRF, multilabel validation fold of Pascal 2012 dataset 21 labels our method is always better than mean-field, ICM running time in seconds

Full CRFs :Semantic Segmentaiton Test fold of Pascal 2012 dataset 21 labels Overall IOU Unary 67.143 Superpixels 65.89 Mean Field 67.3 Ours 67.75

Full CRFs :Semantic Segmentaiton (a) Input image (b) superpixels (c) unary terms (d) our result (e) ground truth

Summary Quantized Edge Full CRF model Approximation to Gaussian Edge CRF Helps to understand properties of Gaussian Edge CRF Efficient optimization of Quantized Edge full CRF with graph cuts Transform the original problem to a smaller domain Optimization quality significantly better than mean field inference