The University of Ontario CS 4487/9687 Algorithms for Image Analysis Multi-Label Image Analysis Problems.

Slides:



Advertisements
Similar presentations
Mean-Field Theory and Its Applications In Computer Vision1 1.
Advertisements

Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London Tutorial at GDR (Optimisation Discrète, Graph Cuts.
Bayesian Belief Propagation
HOPS: Efficient Region Labeling using Higher Order Proxy Neighborhoods Albert Y. C. Chen 1, Jason J. Corso 1, and Le Wang 2 1 Dept. of Computer Science.
Various Regularization Methods in Computer Vision Min-Gyu Park Computer Vision Lab. School of Information and Communications GIST.
Graph Cut Algorithms for Computer Vision & Medical Imaging Ramin Zabih Computer Science & Radiology Cornell University Joint work with Y. Boykov, V. Kolmogorov,
Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
I Images as graphs Fully-connected graph – node for every pixel – link between every pair of pixels, p,q – similarity w ij for each link j w ij c Source:
The University of Ontario CS 4487/9687 Algorithms for Image Analysis Multi-Label Image Analysis Problems.
ICCV 2007 tutorial on Discrete Optimization Methods in Computer Vision part I Basic overview of graph cuts.
C. Olsson Higher-order and/or non-submodular optimization: Yuri Boykov jointly with Western University Canada O. Veksler Andrew Delong L. Gorelick C. NieuwenhuisE.
1 s-t Graph Cuts for Binary Energy Minimization  Now that we have an energy function, the big question is how do we minimize it? n Exhaustive search is.
Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)
Corp. Research Princeton, NJ Computing geodesics and minimal surfaces via graph cuts Yuri Boykov, Siemens Research, Princeton, NJ joint work with Vladimir.
Markov Random Fields (MRF)
1 Can this be generalized?  NP-hard for Potts model [K/BVZ 01]  Two main approaches 1. Exact solution [Ishikawa 03] Large graph, convex V (arbitrary.
Robust Higher Order Potentials For Enforcing Label Consistency
More on Stereo. Outline Fast window-based correlationFast window-based correlation DiffusionDiffusion Energy minimizationEnergy minimization Graph cutsGraph.
P 3 & Beyond Solving Energies with Higher Order Cliques Pushmeet Kohli Pawan Kumar Philip H. S. Torr Oxford Brookes University CVPR 2007.
The University of Ontario University of Bonn July 2008 Optimization of surface functionals using graph cut algorithms Yuri Boykov presenting joint work.
Improved Moves for Truncated Convex Models M. Pawan Kumar Philip Torr.
2010/5/171 Overview of graph cuts. 2010/5/172 Outline Introduction S-t Graph cuts Extension to multi-label problems Compare simulated annealing and alpha-
Stereo & Iterative Graph-Cuts Alex Rav-Acha Vision Course Hebrew University.
MRF Labeling With Graph Cut CMPUT 615 Nilanjan Ray.
The plan for today Camera matrix
Announcements Readings for today:
Lecture 10: Stereo and Graph Cuts
Stereo Computation using Iterative Graph-Cuts
Comp 775: Graph Cuts and Continuous Maximal Flows Marc Niethammer, Stephen Pizer Department of Computer Science University of North Carolina, Chapel Hill.
What Energy Functions Can be Minimized Using Graph Cuts? Shai Bagon Advanced Topics in Computer Vision June 2010.
Measuring Uncertainty in Graph Cut Solutions Pushmeet Kohli Philip H.S. Torr Department of Computing Oxford Brookes University.
Computer vision: models, learning and inference
Extensions of submodularity and their application in computer vision
Graph-based Segmentation
Mutual Information-based Stereo Matching Combined with SIFT Descriptor in Log-chromaticity Color Space Yong Seok Heo, Kyoung Mu Lee, and Sang Uk Lee.
Graph Cut & Energy Minimization
MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/24/10.
Graph Cut Algorithms for Binocular Stereo with Occlusions
CS774. Markov Random Field : Theory and Application Lecture 13 Kyomin Jung KAIST Oct
Planar Cycle Covering Graphs for inference in MRFS The Typhon Algorithm A New Variational Approach to Ground State Computation in Binary Planar Markov.
Multiplicative Bounds for Metric Labeling M. Pawan Kumar École Centrale Paris Joint work with Phil Torr, Daphne Koller.
Lena Gorelick joint work with O. Veksler I. Ben Ayed A. Delong Y. Boykov.
The University of Ontario How to fit a surface to a point cloud? or optimization of surface functionals in computer vision Yuri Boykov TRICS seminar Computer.
CS 4487/6587 Algorithms for Image Analysis
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London.
Discrete Optimization Lecture 3 – Part 1 M. Pawan Kumar Slides available online
1 Markov Random Fields with Efficient Approximations Yuri Boykov, Olga Veksler, Ramin Zabih Computer Science Department CORNELL UNIVERSITY.
Fast and accurate energy minimization for static or time-varying Markov Random Fields (MRFs) Nikos Komodakis (Ecole Centrale Paris) Nikos Paragios (Ecole.
Probabilistic Inference Lecture 5 M. Pawan Kumar Slides available online
Lecture 19: Solving the Correspondence Problem with Graph Cuts CAP 5415 Fall 2006.
Presenter : Kuang-Jui Hsu Date : 2011/3/24(Thur.).
Gaussian Mixture Models and Expectation-Maximization Algorithm.
Machine Learning – Lecture 15
Motion Estimation using Markov Random Fields Hrvoje Bogunović Image Processing Group Faculty of Electrical Engineering and Computing University of Zagreb.
Pseudo-Bound Optimization for Binary Energies Meng Tang 1 Ismail Ben Ayed 2 Yuri Boykov 1 1 University of Western Ontario, Canada 2 GE Healthcare Canada.
CS654: Digital Image Analysis Lecture 28: Advanced topics in Image Segmentation Image courtesy: IEEE, IJCV.
A global approach Finding correspondence between a pair of epipolar lines for all pixels simultaneously Local method: no guarantee we will have one to.
Efficient Belief Propagation for Image Restoration Qi Zhao Mar.22,2006.
Graph Algorithms for Vision Amy Gale November 5, 2002.
Photoconsistency constraint C2 q C1 p l = 2 l = 3 Depth labels If this 3D point is visible in both cameras, pixels p and q should have similar intensities.
Energy functions f(p)  {0,1} : Ising model Can solve fast with graph cuts V( ,  ) = T[  ] : Potts model NP-hard Closely related to Multiway Cut Problem.
Rounding-based Moves for Metric Labeling M. Pawan Kumar École Centrale Paris INRIA Saclay, Île-de-France.
Energy minimization Another global approach to improve quality of correspondences Assumption: disparities vary (mostly) smoothly Minimize energy function:
Markov Random Fields Tomer Michaeli Graduate Course
Graph Cuts vs. Level Sets
Markov Random Fields with Efficient Approximations
Efficient Graph Cut Optimization for Full CRFs with Quantized Edges
Discrete Optimization Methods Basic overview of graph cuts
“Traditional” image segmentation
Presentation transcript:

The University of Ontario CS 4487/9687 Algorithms for Image Analysis Multi-Label Image Analysis Problems

The University of Ontario CS 4487/9687 Algorithms for Image Analysis Topic overview n Multi-label problems: Stereo, restoration, texture synthesis, multi-object segmentation n Types of pair-wise pixel interactions Convex interactions Discontinuity preserving interactions n Energy minimization algorithms: Ishikawa (convex) a-expansions (robust metric interactions) ICM, simulated annealing, message passing, etc. (general) Extra Reading: Szeliski Ch 3.7

The University of Ontario Example of a binary labeling problem Object/background segmentation (topics 4-6) is an example of binary labeling problem feasible labels at any pixel S S S S

The University of Ontario Example of a binary labeling problem Object/background segmentation (topics 4-6) is an example of binary labeling problem feasible labels at any pixel labeling of image pixels or, equivalently, S S S S For convenience this topic uses L p (label at p) instead of S p (segment at p)

The University of Ontario depth map Stereo (topic 7) is an example of image labeling problem with non-binary labels Example of a multi-label problem feasible disparities at any pixel or, equivalently, labeling of image pixels In topic 7 we used equivalent notation

The University of Ontario Remember: stereo with s-t graph cuts [Roy&Cox’98] x y

The University of Ontario Remember: stereo with s-t graph cuts [Roy&Cox’98] s t cut L(p) p “cut” x y labels x y Disparity labels

The University of Ontario Multi-label energy minimization with s-t graph cuts [Ishikawa 1998, 2003] n Exact optimization for convex pair-wise potentials V(dL) dL=Lp-Lq V(dL) dL=Lp-Lq graph construction for linear interactions extends to “convex” interactions works only for 1D labels

The University of Ontario Q: Are convex models good enough for general labeling problems in vision? A: Unfortunately, NO (see the following discussion)

The University of Ontario observed noisy image I I p along one scan line in the image Reconstruction in Vision: (a basic example) image labeling L (restored intensities) L = {L 1, L 2,..., L n } I = {I 1, I 2,..., I n } How to compute L from I ? L

The University of Ontario observed noisy image I I p along one scan line in the image Reconstruction in Vision: (a basic example) image labeling L (restored intensities) L = {L 1, L 2,..., L n } I = {I 1, I 2,..., I n } How to compute L from I ? L

The University of Ontario Energy minimization (discrete approach) n Markov Random Fields (MRF) framework weak membrane model (Geman&Geman’84, Blake&Zisserman’83,87) discontinuity preserving potentials Blake&Zisserman’83,87 spatial regularization data fidelity piece-wise smooth labeling

The University of Ontario Basic pairwise potentials V(α,β) : n Convex regularization gradient descent works exact polynomial algorithms (Ishikawa) n TV regularization (extreme case of convex) a bit harder (non-differentiable) global minima algorithms (Ishikawa, Hochbaum, Nikolova et al.) n Robust regularization (“discontinuity-preserving”) bounded potentials (e.g. truncated convex) NP-hard, many local minima good approximations (message passing, a-expansion) total variation

The University of Ontario Robust pairwise regularization n Robust regularization (“discontinuity-preserving”) bounded potentials (e.g. truncated convex) NP-hard, many local minima good approximations (message passing, a-expansion) piece-wise smooth labeling

The University of Ontario Robust pairwise regularization n Robust regularization (“discontinuity-preserving”) bounded potentials (e.g. Ising or Potts model) NP-hard, many local minima provably good approximations (a-expansion) maxflow/mincut algorithms piece-wise smooth labeling “perceptual grouping” piece-wise constant labeling weak membrane

The University of Ontario Potts model (piece-wise constant labeling) n Robust regularization (“discontinuity-preserving”) bounded potentials (e.g. Ising or Potts model) NP-hard, many local minima provably good approximations (a-expansion) maxflow/mincut algorithms

The University of Ontario Left eye image Right eye image Potts model (piece-wise constant labeling) depth layers n Robust regularization (“discontinuity-preserving”) bounded potentials (e.g. Ising or Potts model) NP-hard, many local minima provably good approximations (a-expansion) maxflow/mincut algorithms

The University of Ontario Potts model (piece-wise constant labeling) C n Robust regularization (“discontinuity-preserving”) bounded potentials (e.g. Ising or Potts model) NP-hard, many local minima provably good approximations (a-expansion) maxflow/mincut algorithms

The University of Ontario Pairwise interactions V: “convex” vs. “discontinuity-preserving” V(dL) dL=Lp-Lq Potts model robust “ discontinuity preserving ” interactions V V(dL) dL=Lp-Lq “ convex ” interactions V V(dL) dL=Lp-Lq V(dL) dL=Lp-Lq “linear” model (TV) “quadratic” model bounded models (truncated convex) piecewise constant labeling piecewise smooth labeling smooth labeling smooth labeling (with some discontinuity robustness) “most robust” among convex good for perceptual grouping see comparison in the next slides

The University of Ontario Pairwise interactions V: “convex” vs. “discontinuity-preserving” NOTE: optimization of restoration energy with quadratic regularization term relates to noise-reduction via mean-filtering (Topic 3) HINT: consider initial labeling L = I. Treat label L p at one pixel p as the only optimization variable reducing our energy to. The optimal value of L p is weighted mean => local optimization of E(L) corresponds to mean-filtering with kernel I L quadratic

The University of Ontario Pairwise interactions V: “convex” vs. “discontinuity-preserving” NOTE: optimization of quadratic regularization model relates to noise-reduction via mean-filtering (Topic 3) HINT: consider initial labeling L = I. Treat label L p at one pixel p as the only optimization variable reducing our energy to. The optimal value of L p is weighted mean => local optimization of E(L) corresponds to mean-filtering with kernel quadratic I L

The University of Ontario I p blurred edge along one scan line in the image Pairwise interactions V: “convex” vs. “discontinuity-preserving” quadratic (convex) may over-smooth (similarly to mean filtering) quadratic I L NOTE: minimizing the sum of quadratic differences prefers to split one large jump into many small ones a large jump

The University of Ontario I p blurred edge along one scan line in the image Pairwise interactions V: “convex” vs. “discontinuity-preserving” histogram equalization linear (TV) I L quadratic (convex) may over-smooth (similarly to mean filtering) linear (TV) NOTE: minimizing the sum of absolute differences does not care how a large jump is split (the sum does not change) => no over-smoothing may create “stair-case” better robustness (similar to median vs. mean ) a large jump

The University of Ontario Pairwise interactions V: “convex” vs. “discontinuity-preserving” Potts model may create “false banding” (next slide) histogram equalization Potts I L I p blurred edge along one scan line in the image quadratic (convex) may over-smooth (similarly to mean filtering) linear (TV) may create “stair-case” bounded (e.g. Potts) NOTE: minimizing the sum of bounded differences prefers one large jump to splitting into smaller ones. => restores sharp boundaries. a large jump

The University of Ontario convex models discontinuity-preserving models linear (TV) quadrat ic truncated linear truncat ed quadrat ic Potts restoration with: Pairwise Regularization Models (comparison) smooth piecewis e smooth piecewise constant noisy images stair-casing over-smoothing stair-casing over-smoothing banding common artifacts: over-smoothing, stair-casing, banding

The University of Ontario code Optimization for “discontinuity preserving” models n NP-hard problem (3 or more labels) two labels can be solved via s-t cuts n a-expansion approximation algorithm [BVZ 1998, PAMI2001] guaranteed approximation quality (Veksler, 2001) –within a factor of 2 from the global minima (Potts model) n Many other (small or large) move making algorithms - a/b swap, jump moves, range moves, fusion moves, etc. n LP relaxations, message passing, e.g. (LBP, TRWS) n Other MRF techniques (simulated annealing, ICM) n Variational methods (e.g. multi-phase level-sets)

The University of Ontario other labels a a-expansion move Basic idea is motivated by methods for multi-way cut problem (similar to Potts model) Break computation into a sequence of binary s-t cuts

The University of Ontario a-expansion (binary move) optimizies sumbodular set function expansions correspond to subsets (shaded area) L current labeling =

The University of Ontario a-expansion (binary move) optimizies sumbodular set function L current labeling 0 1

The University of Ontario a-expansion (binary move) optimizies sumbodular set function L current labeling

The University of Ontario a-expansion (binary move) optimizies sumbodular set function L current labeling Set function is submodular if

The University of Ontario a-expansion (binary move) optimizies sumbodular set function L current labeling Set function is submodular if = 0 triangular inequality for ||a-b||=E(a,b)

The University of Ontario a-expansion (binary move) optimizies sumbodular set function L current labeling a-expansion moves are submodular if is a metric on the space of labels [Boykov, Veksler, Zabih, PAMI 2001]

The University of Ontario examples of metric pairwise interactions FACT (easy to prove): any truncated metric is also a metric But, this is only a special case of L 2 metric for 1D labels. In general, for this metric is defined as We called this linear or TV potential Just check that Truncated L 2 is also a metric Potts is another important example of a metric Quadratic (squared L 2 ) and truncated quadratic potentials are not metrics. Other very good approximation algorithms are available (e.g. TRW, Kolmogorov&Weinright 2006) Note: unlike Ishikawa, a-expansion and other methods (LBP, TRWS, etc.) apply to labels in

The University of Ontario a-expansion algorithm 1.Start with any initial solution 2.For each label “a” in any (e.g. random) order 1.Compute optimal a-expansion move (s-t graph cuts) 2.Decline the move if there is no energy decrease 3.Stop when no expansion move would decrease energy

The University of Ontario a-expansion moves initial solution -expansion In each a-expansion a given label “a” grabs space from other labels For each move we choose expansion that gives the largest decrease in the energy: binary optimization problem

The University of Ontario Multi-way graph cuts stereo vision original pair of “stereo” images depth map ground truth BVZ 1998 KZ 2002 right image left image

The University of Ontario a-expansions vs. ICM basic general alternative Iterated Conditional Modes (ICM) Example : consider pair-wise energy Unlike a-expansion, optimizes only over a single pixel at a time => local (pixel-wise) optimization - Consider current labeling L and any given pixel - Treat label L p as the only optimization variable x keeping other labels for fixed. - This reduces our energy to - Select optimal by enumeration and set new label. - Iterate over all pixels until no pixel operation reduces the energy. ICM algorithm

The University of Ontario a-expansions vs. simulated annealing - Consider current labeling L and any given pixel - Treat label L p as the only optimization variable x keeping other labels for fixed. - This reduces our energy to - Select optimal by enumeration and set new label. - Iterate over all pixels until no pixel operation reduces the energy. ICM algorithm - incorporates the following randomization strategy into ICM: at each pixel p label is updated randomly according to probabilities x … n Pr(x) ~... NOTE 1 - lower energy E p (x) gives x more chances to be selected randomization of ICM simulated annealing (SA) NOTE 2 - higher temperature parameter T means more randomness - lower temperature parameter T reduces to ICM (optimal x is always selected) Unlike a-expansion, optimizes only over a single pixel at a time => local (pixel-wise) optimization Typical SA starts with high T and gradually reduces T to zero. Szeliski: appendix B.5.1 [Geman&Geman 1982]

The University of Ontario a-expansions vs. simulated annealing initial labeling one pixel local move (ICM or SA) large move ( a-b swap) large move (a-expansion) a- expansion and ICM/SA are all greedy iterative methods but, they converge to different kinds of local minima -ICM/SA solution can not be improved by changing a label of any one pixel to any given label a -a- expansion solution can not be improved by changing any subset of pixels to any given label a small vs. large moves

The University of Ontario normalized correlation, start for annealing, 24.7% err a-expansions (BVZ 89,01) 90 seconds, 5.8% err a-expansions vs. simulated annealing

The University of Ontario simulated annealing, 19 hours, 20.3% err a-expansions (BVZ 89,01) 90 seconds, 5.8% err a-expansions vs. simulated annealing NOTE 1: ICM and SA are general methods applicable to arbitrary non-metric and high-order energies NOTE 2: now-days there are other general methods based on graph cuts, message passing, relaxations, etc. [BVZ,2001] a-expansion

The University of Ontario Other applications Graph-cut textures (Kwatra, Schodl, Essa, Bobick 2003) similar to “image-quilting” (Efros & Freeman, 2001) A B C D E F G H I J A B G D C F H I J E

The University of Ontario Other applications Graph-cut textures (Kwatra, Schodl, Essa, Bobick 2003) Google (maps stitching)

The University of Ontario Other applications Multi-object Extraction Obvious generalization of binary object extraction technique (Boykov, Jolly, Funkalea 2004)

The University of Ontario Other applications Image compositing (Agarwala et al. 2004, see Szeliski Sec 9.3.2)

The University of Ontario Block-coordinate descent alternating a-expansion (for segmentation L ) and fitting colors I i Color model fitting (multi-label version of Chan-Vese) Potts model

The University of Ontario Block-coordinate descent alternating a-expansion (for segmentation L ) and fitting colors I i Stereo via piece-wise constant plane fitting [Birchfield &Tomasi 1999] Models T = parameters of affine transformations T(p)=a p + b 2x2 2x1 Potts model

The University of Ontario Block-coordinate descent alternating a-expansion (for segmentation L ) and fitting colors I i Piece-wise smooth local plane fitting [Olsson et al. 2013] truncated angle-differences non-metric interactions need other optimization

The University of Ontario Block-coordinate descent alternating a-expansion (for segmentation L ) and fitting colors I i Signboard segmentation [Milevsky 2013] Labels = planes in RGBXY space C(p) = a x + b Potts model 3x2 3x1

The University of Ontario Signboard segmentation [Milevsky 2013] 3x2 3x1 Goal: detection of characters, then text line fitting and translation

The University of Ontario Learning pair-wise potentials Structure detection (Kumar and Hebert 2006, see Szeliski Sec 3.7) standard pair-wise potentialslearned pair-wise potentials

The University of Ontario Multi-label optimization n 80% of computer vision and bio-medical image analysis are ill-posed labeling problems requiring optimization of regularization energies n Most problems are NP hard n Optimization algorithms is area of active research Google, Microsoft, GE, Siemens, Adobe, etc. LP relaxations [Schlezinger, Kolmogorov, Komodakis, Savchinsky,…] Message passing, e.g. LBP, TRWS [Kolmogorov] Graph Cuts (a-expanson, a/b-swap, fusion, FTR, etc) Variational methods