Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
Graphical Models BRML Chapter 4 1. the zoo of graphical models Markov networks Belief networks Chain graphs (Belief and Markov ) Factor graphs =>they.
Approximation, Chance and Networks Lecture Notes BISS 2005, Bertinoro March Alessandro Panconesi University La Sapienza of Rome.
Gibbs sampler - simple properties It’s not hard to show that this MC chain is aperiodic. Often is reversible distribution. If in addition the chain is.
Introduction to Markov Random Fields and Graph Cuts Simon Prince
Exact Inference in Bayes Nets
Randomized Algorithms Kyomin Jung KAIST Applied Algorithm Lab Jan 12, WSAC
CS774. Markov Random Field : Theory and Application Lecture 17 Kyomin Jung KAIST Nov
Convergent Message-Passing Algorithms for Inference over General Graphs with Convex Free Energies Tamir Hazan, Amnon Shashua School of Computer Science.
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Probabilistic Inference Lecture 1
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
Belief Propagation on Markov Random Fields Aggeliki Tsoli.
Graduate School of Information Sciences, Tohoku University
CS774. Markov Random Field : Theory and Application Lecture 04 Kyomin Jung KAIST Sep
Approximate Counting via Correlation Decay Pinyan Lu Microsoft Research.
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
CS774. Markov Random Field : Theory and Application Lecture 06 Kyomin Jung KAIST Sep
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
Distributed Message Passing for Large Scale Graphical Models Alexander Schwing Tamir Hazan Marc Pollefeys Raquel Urtasun CVPR2011.
Convergent and Correct Message Passing Algorithms Nicholas Ruozzi and Sekhar Tatikonda Yale University TexPoint fonts used in EMF. Read the TexPoint manual.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 8 May 4, 2005
Recent Development on Elimination Ordering Group 1.
MRF Labeling With Graph Cut CMPUT 615 Nilanjan Ray.
Announcements Readings for today:
Understanding Belief Propagation and its Applications Dan Yuan June 2004.
Belief Propagation Kai Ju Liu March 9, Statistical Problems Medicine Finance Internet Computer vision.
Bayesian Networks Alan Ritter.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Finding a maximum independent set in a sparse random graph Uriel Feige and Eran Ofek.
©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.
Image Analysis and Markov Random Fields (MRFs) Quanren Xiong.
Some Surprises in the Theory of Generalized Belief Propagation Jonathan Yedidia Mitsubishi Electric Research Labs (MERL) Collaborators: Bill Freeman (MIT)
CS774. Markov Random Field : Theory and Application Lecture 08 Kyomin Jung KAIST Sep
Graph Coalition Structure Generation Maria Polukarov University of Southampton Joint work with Tom Voice and Nick Jennings HUJI, 25 th September 2011.
Presenter : Kuang-Jui Hsu Date : 2011/5/23(Tues.).
CS774. Markov Random Field : Theory and Application Lecture 13 Kyomin Jung KAIST Oct
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
Message-Passing for Wireless Scheduling: an Experimental Study Paolo Giaccone (Politecnico di Torino) Devavrat Shah (MIT) ICCCN 2010 – Zurich August 2.
CS774. Markov Random Field : Theory and Application Lecture 21 Kyomin Jung KAIST Nov
Edge-disjoint induced subgraphs with given minimum degree Raphael Yuster 2012.
Expanders via Random Spanning Trees R 許榮財 R 黃佳婷 R 黃怡嘉.
Markov Random Fields Probabilistic Models for Images
Combinatorial Optimization Problems in Computational Biology Ion Mandoiu CSE Department.
CS774. Markov Random Field : Theory and Application Lecture 02
An Introduction to Variational Methods for Graphical Models
Probabilistic Graphical Models seminar 15/16 ( ) Haim Kaplan Tel Aviv University.
Lecture 2: Statistical learning primer for biologists
Belief Propagation and its Generalizations Shane Oldenburger.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
CS774. Markov Random Field : Theory and Application Lecture 15 Kyomin Jung KAIST Oct
Pattern Recognition and Machine Learning
Spatial decay of correlations and efficient methods for computing partition functions. David Gamarnik Joint work with Antar Bandyopadhyay (U of Chalmers),
NOTE: To change the image on this slide, select the picture and delete it. Then click the Pictures icon in the placeholder to insert your own image. Fast.
Bayesian Belief Propagation for Image Understanding David Rosenberg.
Markov Random Fields in Vision
A Binary Linear Programming Formulation of the Graph Edit Distance Presented by Shihao Ji Duke University Machine Learning Group July 17, 2006 Authors:
Ch 6. Markov Random Fields 6.1 ~ 6.3 Adaptive Cooperative Systems, Martin Beckerman, Summarized by H.-W. Lim Biointelligence Laboratory, Seoul National.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
Markov Random Fields with Efficient Approximations
Path Coupling And Approximate Counting
Exact Inference Continued
Bucket Renormalization for Approximate Inference
≠ Particle-based Variational Inference for Continuous Systems
Graduate School of Information Sciences, Tohoku University
Expectation-Maximization & Belief Propagation
Locality In Distributed Graph Algorithms
Presentation transcript:

Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research) Devavrat Shah (MIT) 서울대학교 July 30 th 2009

Graphical Model A probabilistic model for which a graph denotes the conditional independence structure between random variables.  Bayesian network (directed graph)  Markov Random Field (undirected graph) Recently successful in machine learning

Graphical Model A probabilistic model for which a graph denotes the conditional independence structure between random variables.  Bayesian network (directed graph)

Graphical Model A probabilistic model for which a graph denotes the conditional independence structure between random variables.  Markov Random Field (undirected graph)

Markov Random Field (MRF) Developed from Ising model in statistical physics. Applications  computer vision, error correct coding, speech recognition, gene finding etc. Many heuristics for inference problems in MRF are devised.  Theoretical guarantee for the correctness of those algorithms are not known much. Our goal : designing simple algorithms for inference with provable error bound by utilizing structures of the MRF.

Outline Problem Statement and an Example Relevant Work Our Algorithms for Approximate Inference  Efficient algorithms based on local updates  When the MRF is defined on a graph with polynomial growth, our algorithm achieves approximation within arbitrary accuracy Applications  Image denoising  Image segmentation Conclusion

Markov Random Field (MRF) A collection of random variables, defined on a graph G. The probability distribution of at vertex is dependent only on its neighbors : Graph G

Markov Random Field (MRF) A collection of random variables, defined on a graph G. The probability distribution of at vertex is dependent only on its neighbors : Graph G

Markov Random Field (MRF) A collection of random variables, defined on a graph G. The probability distribution of at vertex is dependent only on its neighbors : Graph G

pair-wise MRF if for some and Pair-wise MRF Z is called the partition function.

Computing Maximum A Posteriori MAP(Maximum A Posteriori) assignment  Most likely assignment (mode of the distribution)  NP-hard even for simple graphs like grid. Our goal  For a given, compute approximation of MAP : such that

Example : Image denoising We want to restore a binary (-1/+1) image Y of size with noise added. Consider Y as an element of Use an MRF model to restore the original image. The underlying graph is a grid graph of size

Example : Image denoising Utilizes two properties of the original image Similar to Y It is smooth, i.e. number of edges with different color is small Define the following MRF, where MAP assignment : original image

Computing partition function Z Equivalent to computing marginal probability approximation of log Z is useful for many applications including statistical physics, computer vision. Our goal: compute such that

Relevant Work Belief Propagation (BP)  BP and its variants like Tree-Reweighted algorithm have been very successful when G does not have many small cycles. Ex) good when G is locally tree-like, and the MRF has correlation decay [Jordan, Tatikonda ‘99].  When G has lots of small cycles, their correctness are not known.  Pearl [‘88], Weiss [‘00], Yididia and Freeman [‘02], Wainwright, Jaakkola and Willsky [‘03]

Relevant Work Markov Chain Monte Carlo  Computing approximation of log Z  key is to prove rapid mixing property which is non-trivial.  Jerrum and Sinclair [‘89], Dyer, Frieze and Kannan [‘91] Recent development  Weitz [‘06] using self-avoiding walk tree approach  Deterministic computation for Z for graphs with degree <6  Cannot be applied to graphs with higher degree.

Our approach Computing approximation of MAP and log-partition function for general graphs are NP-hard. Many real applications of MRF model are defined on polynomially growing graphs. We utilize structural properties of the polynomially growing graphs to obtain approximation algorithms.

Polynomially growing graph ball of radius r around v w.r.t. the shortest path distance of G. G

Polynomially growing graph G

G (A sequence of) graph is polynomially growing if there is constants s.t. for all

Outline of our algorithm : MAP Begin with a random assignment. Choose an arbitrary order of vertices With the given vertex as a center, choose a ball of radius r, where r is chosen from a geometric distribution. Compute a MAP inside the ball while fixing the assignment outside the ball. Update by the computed MAP inside the ball. Output We show is an approximation of MAP.

Our MAP Algorithm

for

Our MAP Algorithm

for

Property of the geometric distribution For any Hence, for any edge e, Pr[ e is on the boundary of B(v,r)] Pr[e is inside the ball B(v,r)]

Proof for MAP Algorithm Consider an imaginary boundary of the algorithm as follows

Proof for MAP Algorithm Consider an imaginary boundary of the algorithm as follows

Proof for MAP Algorithm For any edge e of the graph G Pr[ e belongs to the boundary of the algorithm] Polynomial growth  Size of each ball is small  computation is efficient

Proof of approximation If we restrict to a region R, it is a MAP assignment in R with some fixed assignment outside R. Also, restricted to the region R is a MAP assignment in R with another fixed assignment outside R. region

Proof of approximation We show the following Lemma : if the total differences of the potential functions for two MRFs and on R is small, the difference between the probabilities induced by the MAP assignments for and on R is small. region

Proof of approximation By this lemma and the fact that for any edge e of G, Pr[ e belongs to the boundary of the algorithm] we obtain that the sum of the differences of the probabilities for all regions induced by and is small. region

Theorem [Jung, Shah] For the computation of MAP, our algorithm achieves approximate solution in expectation and it runs in time

Outline of Our Algorithm : log-partition function Obtain a random graph decomposition by removing some edges. Compute the log-partition function inside each connected component, while replacing the potential functions of the removed boundary edges of the component by a constant. Summand the computed values and output it. we show that the output is an approximation of the log-partition function.

Graph decomposition

for Graph decomposition

for Graph decomposition

for Graph decomposition

for Graph decomposition

Proof of approximation bounds Ex, for the upper bound, where R is regions and B is the set of boundary edges.

Theorem [Jung, Shah] For the computation of log Z, our algorithm outputs approximate upper bound and lower bound of log Z in expectation, and it runs in time

Application to Image Processing In computer vision, the underlying graph is a grid Relevant Problems  Image Denoising  Image segmentation/ reconstruction Detect a specific object in an image Ex) face recognition, medical image process

We require the ratio of specific part of an object is close to a fixed ratio Ex) Face segmentation Fix ratios of eye, nose, mouth, etc. For the computation of MAP with fixed ratio, we provide an algorithm that outputs approximate solution in time, where k is the number of objects*. MRF with fixed ratio * Joint work with Kohli

Future work Adaptation of existing algorithms to computations in each component Learning underlying Markov Random Field Understanding limitations of inference algorithms Thank you