Efficient Inference for Fully-Connected CRFs with Stationarity

Slides:



Advertisements
Similar presentations
Mean-Field Theory and Its Applications In Computer Vision1 1.
Advertisements

Vibhav Vineet, Jonathan Warrell, Paul Sturgess, Philip H.S. Torr
Mean-Field Theory and Its Applications In Computer Vision3 1.
1 Vibhav Vineet, Jonathan Warrell, Paul Sturgess, Philip H.S. Torr Improved Initialisation and Gaussian Mixture Pairwise Terms for Dense Random Fields.
O BJ C UT M. Pawan Kumar Philip Torr Andrew Zisserman UNIVERSITY OF OXFORD.
Mean-Field Theory and Its Applications In Computer Vision2
Tighter and Convex Maximum Margin Clustering Yu-Feng Li (LAMDA, Nanjing University, China) Ivor W. Tsang.
Introduction to Markov Random Fields and Graph Cuts Simon Prince
Joint Optimisation for Object Class Segmentation and Dense Stereo Reconstruction Ľubor Ladický, Paul Sturgess, Christopher Russell, Sunando Sengupta, Yalin.
Learning to Combine Bottom-Up and Top-Down Segmentation Anat Levin and Yair Weiss School of CS&Eng, The Hebrew University of Jerusalem, Israel.
Scene Labeling Using Beam Search Under Mutex Constraints ID: O-2B-6 Anirban Roy and Sinisa Todorovic Oregon State University 1.
Human Action Recognition across Datasets by Foreground-weighted Histogram Decomposition Waqas Sultani, Imran Saleemi CVPR 2014.
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
ImageNet Classification with Deep Convolutional Neural Networks
Exploiting Sparse Markov and Covariance Structure in Multiresolution Models Presenter: Zhe Chen ECE / CMR Tennessee Technological University October 22,
I Images as graphs Fully-connected graph – node for every pixel – link between every pair of pixels, p,q – similarity w ij for each link j w ij c Source:
CS 551 / CS 645 Antialiasing. What is a pixel? A pixel is not… –A box –A disk –A teeny tiny little light A pixel is a point –It has no dimension –It occupies.
Mixture of trees model: Face Detection, Pose Estimation and Landmark Localization Presenter: Zhang Li.
An Analysis of Convex Relaxations (PART I) Minimizing Higher Order Energy Functions (PART 2) Philip Torr Work in collaboration with: Pushmeet Kohli, Srikumar.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Accurate Binary Image Selection From Inaccurate User Input Kartic Subr, Sylvain Paris, Cyril Soler, Jan Kautz University College London, Adobe Research,
1 Fast Primal-Dual Strategies for MRF Optimization (Fast PD) Robot Perception Lab Taha Hamedani Aug 2014.
Contextual Classification with Functional Max-Margin Markov Networks Dan MunozDrew Bagnell Nicolas VandapelMartial Hebert.
Models for Scene Understanding – Global Energy models and a Style-Parameterized boosting algorithm (StyP-Boost) Jonathan Warrell, 1 Simon Prince, 2 Philip.
A Constraint Generation Approach to Learning Stable Linear Dynamical Systems Sajid M. Siddiqi Byron Boots Geoffrey J. Gordon Carnegie Mellon University.
Robust Higher Order Potentials For Enforcing Label Consistency
Schedule Introduction Models: small cliques and special potentials Tea break Inference: Relaxation techniques:
An Analysis of Convex Relaxations M. Pawan Kumar Vladimir Kolmogorov Philip Torr for MAP Estimation.
Improved Moves for Truncated Convex Models M. Pawan Kumar Philip Torr.
MRF Labeling With Graph Cut CMPUT 615 Nilanjan Ray.
The Layout Consistent Random Field for Recognizing and Segmenting Partially Occluded Objects By John Winn & Jamie Shotton CVPR 2006 presented by Tomasz.
Graph Cut based Inference with Co-occurrence Statistics Ľubor Ladický, Chris Russell, Pushmeet Kohli, Philip Torr.
A Constraint Generation Approach to Learning Stable Linear Dynamical Systems Sajid M. Siddiqi Byron Boots Geoffrey J. Gordon Carnegie Mellon University.
Relaxations and Moves for MAP Estimation in MRFs M. Pawan Kumar STANFORDSTANFORD Vladimir KolmogorovPhilip TorrDaphne Koller.
Measuring Uncertainty in Graph Cut Solutions Pushmeet Kohli Philip H.S. Torr Department of Computing Oxford Brookes University.
What, Where & How Many? Combining Object Detectors and CRFs
Radial Basis Function Networks
Graph-based Segmentation
Introduction to Image Processing Grass Sky Tree ? ? Review.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
CS774. Markov Random Field : Theory and Application Lecture 08 Kyomin Jung KAIST Sep
Conditional Topic Random Fields Jun Zhu and Eric P. Xing ICML 2010 Presentation and Discussion by Eric Wang January 12, 2011.
CSC2535: Computation in Neural Networks Lecture 11: Conditional Random Fields Geoffrey Hinton.
Multiplicative Bounds for Metric Labeling M. Pawan Kumar École Centrale Paris Joint work with Phil Torr, Daphne Koller.
Texture We would like to thank Amnon Drory for this deck הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Texture We would like to thank Amnon Drory for this deck הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Markov Random Fields Probabilistic Models for Images
Associative Hierarchical CRFs for Object Class Image Segmentation Ľubor Ladický 1 1 Oxford Brookes University 2 Microsoft Research Cambridge Based on the.
Andreas Papadopoulos - [DEXA 2015] Clustering Attributed Multi-graphs with Information Ranking 26th International.
Associative Hierarchical CRFs for Object Class Image Segmentation
A Dynamic Conditional Random Field Model for Object Segmentation in Image Sequences Duke University Machine Learning Group Presented by Qiuhua Liu March.
CS654: Digital Image Analysis Lecture 28: Advanced topics in Image Segmentation Image courtesy: IEEE, IJCV.
Mean field approximation for CRF inference
TextonBoost: Joint Appearance, Shape and Context Modeling for Multi-Class Object Recognition and Segmentation J. Shotton ; University of Cambridge J. Jinn,
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Tightening LP Relaxations for MAP using Message-Passing David Sontag Joint work with Talya Meltzer, Amir Globerson, Tommi Jaakkola, and Yair Weiss.
M ACHINE L EARNING : F OUNDATIONS C OURSE TAU – 2012A P ROF. Y ISHAY M ANSOUR o TextonBoost : Joint Appearance, Shape and Context Modeling for Multi-Class.
The Chinese University of Hong Kong
IEEE 2015 Conference on Computer Vision and Pattern Recognition Active Learning for Structured Probabilistic Models with Histogram Approximation Qing SunAnkit.
Image segmentation.
Gaussian Conditional Random Field Network for Semantic Segmentation
Holistic Scene Understanding Virginia Tech ECE /02/26 Stanislaw Antol.
Markov Random Fields Tomer Michaeli Graduate Course
Summary of “Efficient Deep Learning for Stereo Matching”
The Chinese University of Hong Kong
Efficient Graph Cut Optimization for Full CRFs with Quantized Edges
Learning to Combine Bottom-Up and Top-Down Segmentation
An Analysis of Convex Relaxations for MAP Estimation
Semantic segmentation
“Traditional” image segmentation
Presentation transcript:

Efficient Inference for Fully-Connected CRFs with Stationarity Yimeng Zhang, Tsuhan Chen CVPR 2012

Summary Explore object-class segmentation with fully-connected CRF models Only restriction on pairwise terms is `spatial stationarity’ (i.e. depend on relative locations) Show how efficient inference can be achieved by Using a QP formulation Using FFT to calculate gradients in complexity (linear in) O(NlogN)

Fully-connected CRF model General pairwise CRF model: Image I Class labeling, X: Label set, L: V = set of pixels, N_i = neighbourhood of pixel i, Z(I) = partition function, psi = potential functions

Fully-connected CRF model General pairwise CRF model: In fully-connected CRF, for all i, N_i = V

Unary Potential Unary potential generates a score for each object class per pixel (TextonBoost)

Pairwise Potential Pairwise potential measures compatibility of the labels at each pair of pixels Combines spatial and colour contrast factors

Pairwise Potential Colour contrast: Spatial term:

Pairwise Potential Learning the spatial term

MAP inference using QP relaxation Introduce a binary indicator variable for each pixel and label MAP inference expressed as a quadratic integer program, and relaxed to give the QP

MAP inference using QP relaxation QP relaxation has been proved to be tight in all cases (Ravikumar ICML 2006 [24]) Moreover, it is convex whenever matrix of edge-weights is negative-definite Additive bound for non-convex case QP requires O(KN) variables, LP requires (K^2E)

MAP inference using QP relaxation Gradient Derive fixed-point update by forming Lagrangian and setting its derivative to 0

Illustration of QP updates

Efficiently evaluating the gradient Required summation Would be a convolution without the color term With color term is requires 5D-filtering Can be approximated by clustering into C color clusters, => C convolutions across

Efficiently evaluating the gradient Hence, for the case x_i = x_j, we need to evaluate Instead, evaluate for C clusters (C = 10 to 15) where Finally, interpolate

Update complexity FFTs of each spatial filters can be calculated in advance (K^2 filters) At each update, we require C FFTs calculating, O(CNlogN) K^2 convolutions are needed, each requiring a multiplication, O(K^2CN) Terms can be added in Fourier domain, => only KC inverse FFTs needed, O(KCNlogN) Run-time per iteration < 0.1s for 213x320 pixels (+ downsampling by factor of 5)

MSRC synthetic experiment Unary terms randomized Spatial distributions set to ground-truth

MSRC synthetic experiment Running times

Sowerby synthetic experiment

MSRC full experiment Use TextonBoost unary potentials Compare with several other CRFs with same unaries Grid only Grid + P^N (Kohli, CVPR 2008) Grid + P^N + Cooccurrence (Ladickỳ, ECCV 2010) Fully-connected + Gaussian spatial (Krähenbühl, NIPS 2011)

MSRC full experiment Qualitative comparison

MSRC full experiment Quantitative comparison Overall Per-class Timing: 2-8s per image