Distributed Message Passing for Large Scale Graphical Models Alexander Schwing Tamir Hazan Marc Pollefeys Raquel Urtasun CVPR2011.

Slides:



Advertisements
Similar presentations
Mean-Field Theory and Its Applications In Computer Vision1 1.
Advertisements

Bayesian Belief Propagation
HOPS: Efficient Region Labeling using Higher Order Proxy Neighborhoods Albert Y. C. Chen 1, Jason J. Corso 1, and Le Wang 2 1 Dept. of Computer Science.
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
Constrained Approximate Maximum Entropy Learning (CAMEL) Varun Ganapathi, David Vickrey, John Duchi, Daphne Koller Stanford University TexPoint fonts used.
Introduction to Markov Random Fields and Graph Cuts Simon Prince
Exact Inference in Bayes Nets
Junction Trees And Belief Propagation. Junction Trees: Motivation What if we want to compute all marginals, not just one? Doing variable elimination for.
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
CS774. Markov Random Field : Theory and Application Lecture 17 Kyomin Jung KAIST Nov
Convergent Message-Passing Algorithms for Inference over General Graphs with Convex Free Energies Tamir Hazan, Amnon Shashua School of Computer Science.
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Markov Networks.
Belief Propagation on Markov Random Fields Aggeliki Tsoli.
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
Practical Belief Propagation in Wireless Sensor Networks Bracha Hod Based on a joint work with: Danny Dolev, Tal Anker and Danny Bickson The Hebrew University.
Convergent and Correct Message Passing Algorithms Nicholas Ruozzi and Sekhar Tatikonda Yale University TexPoint fonts used in EMF. Read the TexPoint manual.
2010/5/171 Overview of graph cuts. 2010/5/172 Outline Introduction S-t Graph cuts Extension to multi-label problems Compare simulated annealing and alpha-
Message Passing Algorithms for Optimization
Belief Propagation, Junction Trees, and Factor Graphs
Understanding Belief Propagation and its Applications Dan Yuan June 2004.
Linear Solution to Scale and Rotation Invariant Object Matching Professor: 王聖智 教授 Student : 周 節.
Stereo Computation using Iterative Graph-Cuts
Computer vision: models, learning and inference
BiGraph BiGraph: Bipartite-oriented Distributed Graph Partitioning for Big Learning Jiaxin Shi Rong Chen, Jiaxin Shi, Binyu Zang, Haibing Guan Institute.
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University COT 5410 – Spring 2004.
Computer vision: models, learning and inference
Reconstructing Relief Surfaces George Vogiatzis, Philip Torr, Steven Seitz and Roberto Cipolla BMVC 2004.
Efficient Gathering of Correlated Data in Sensor Networks
Network Aware Resource Allocation in Distributed Clouds.
CS774. Markov Random Field : Theory and Application Lecture 13 Kyomin Jung KAIST Oct
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Optimization Flow Control—I: Basic Algorithm and Convergence Present : Li-der.
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London.
Continuous Variables Write message update equation as an expectation: Proposal distribution W t (x t ) for each node Samples define a random discretization.
Fast and accurate energy minimization for static or time-varying Markov Random Fields (MRFs) Nikos Komodakis (Ecole Centrale Paris) Nikos Paragios (Ecole.
CS774. Markov Random Field : Theory and Application Lecture 02
Tell Me What You See and I will Show You Where It Is Jia Xu 1 Alexander G. Schwing 2 Raquel Urtasun 2,3 1 University of Wisconsin-Madison 2 University.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Presenter : Kuang-Jui Hsu Date : 2011/3/24(Thur.).
Probabilistic Graphical Models seminar 15/16 ( ) Haim Kaplan Tel Aviv University.
Using Combinatorial Optimization within Max-Product Belief Propagation
Machine Learning Queens College Lecture 7: Clustering.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Domain decomposition in parallel computing Ashok Srinivasan Florida State University.
Data Structures and Algorithms in Parallel Computing Lecture 7.
1 CS612 Algorithms for Electronic Design Automation CS 612 – Lecture 8 Lecture 8 Network Flow Based Modeling Mustafa Ozdal Computer Engineering Department,
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
Pattern Recognition and Machine Learning
Tightening LP Relaxations for MAP using Message-Passing David Sontag Joint work with Talya Meltzer, Amir Globerson, Tommi Jaakkola, and Yair Weiss.
Daphne Koller Overview Maximum a posteriori (MAP) Probabilistic Graphical Models Inference.
Markov Random Fields in Vision
Perfect recall: Every decision node observes all earlier decision nodes and their parents (along a “temporal” order) Sum-max-sum rule (dynamical programming):
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
Introduction of BP & TRW-S
Multiplicative updates for L1-regularized regression
Markov Networks.
Markov Random Fields Presented by: Vladan Radosavljevic.
Graduate School of Information Sciences, Tohoku University
Expectation-Maximization & Belief Propagation
Parallel and Distributed Graph Cuts by Dual Decomposition
Junction Trees 3 Undirected Graphical Models
Topics in Algorithms 2005 Max Cuts
Markov Networks.
Presentation transcript:

Distributed Message Passing for Large Scale Graphical Models Alexander Schwing Tamir Hazan Marc Pollefeys Raquel Urtasun CVPR2011

Outline Introduction Related work Message passing algorithm Distributed convex belief propagation Experiment evaluation Conclusion

Introduction Vision problems → discrete labeling problems in an undirected graphical model (Ex : MRF) – Belief propagation (BP) – Graph cut Depending on the potentials and structure of the graph The main underlying limitations to real-world problems are memory and computation.

Introduction A new algorithm – distribute and parallelize the computation and memory requirements – conserving the convergence and optimality guarantees Computation can be done in parallel by partitioning the graph and imposing agreement between the beliefs in the boundaries. – Graph-based optimization program → local optimization problems (one per machine). – Messages between machines : Lagrange multipliers Stereo reconstruction from high-resolution image Handle large problems (more than 200 labels in images larger than 10 MPixel)

Related work Provable convergence while still being computationally tractable. – parallelizes convex belief propagation – conserves its convergence and optimality guarantees Strandmark and Kahl [24] – splitting the model across multiple machines GraphLab – assumes that all the data is stored in shared-memory

Related work Split the message passing task at hand into several local optimization problems that are solved in parallel. To ensure convergence we force the local tasks to communicate occasionally. At the local level we parallelize the message passing algorithm using a greedy vertex coloring

Message passing algorithm The joint distribution factors into a product of non-negative functions – defines a hypergraph whose nodes represent the n random variables and the subsets of variables x correspond to its hyperedges. Hypergraph – Bipartite graph : factor graph [11] one set of nodes corresponding to the original nodes of the hypergraph : variable nodes the other set consisting of its hyperedges : factor nodes N(i) : all factor nodes that are neighbors of variable node i

Message passing algorithm Maximum a posteriori (MAP) assignment Reformulate the MAP problem as integer linear program.

Message passing algorithm

Distributed convex belief propagation Partition the vertices of the graphical model to disjoint subgraphs – each computer solves independently a variational program with respect to its subgraph. The distributed solutions are then integrated through message- passing between the subgraphs – preserving the consistency of the graphical model. Properties : – If (5) is strictly concave then the algorithm converges for all ε >= 0, and converges to the global optimum when ε > 0.

Distributed convex belief propagation

Lagrange multipliers : – : the marginalization constraints within each computer – : the consistency constraints between the different computers

Distributed convex belief propagation

Experiment evaluation Stereo reconstruction – nine 2.4 GHz x64 Quad-Core computers with 24 GB memory each, connected via a standard local area network libDAI [17] and GraphLAB [16]

Experiment evaluation

relative duality gap

Conclusion Large scale graphical models by dividing the computation and memory requirements into multiple machines. Convergence and optimality guarantees are preserved. Main benefit : the use of multiple computers.