Markov Networks: Theory and Applications Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Linear Time Methods for Propagating Beliefs Min Convolution, Distance Transforms and Box Sums Daniel Huttenlocher Computer Science Department December,
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
Exact Inference in Bayes Nets
HMM II: Parameter Estimation. Reminder: Hidden Markov Model Markov Chain transition probabilities: p(S i+1 = t|S i = s) = a st Emission probabilities:
Dynamic Bayesian Networks (DBNs)
Loopy Belief Propagation a summary. What is inference? Given: –Observabled variables Y –Hidden variables X –Some model of P(X,Y) We want to make some.
Exploiting Sparse Markov and Covariance Structure in Multiresolution Models Presenter: Zhe Chen ECE / CMR Tennessee Technological University October 22,
Learning to estimate human pose with data driven belief propagation Gang Hua, Ming-Hsuan Yang, Ying Wu CVPR 05.
Modeling the Shape of People from 3D Range Scans
3D Human Body Pose Estimation from Monocular Video Moin Nabi Computer Vision Group Institute for Research in Fundamental Sciences (IPM)
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
Introduction of Probabilistic Reasoning and Bayesian Networks
3 March, 2003University of Glasgow1 Statistical-Mechanical Approach to Probabilistic Inference --- Cluster Variation Method and Generalized Loopy Belief.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
Graphical models: approximate inference and learning CA6b, lecture 5.
Belief Propagation on Markov Random Fields Aggeliki Tsoli.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Bayesian network inference
… Hidden Markov Models Markov assumption: Transition model:
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
Hidden Markov Model 11/28/07. Bayes Rule The posterior distribution Select k with the largest posterior distribution. Minimizes the average misclassification.
Global Approximate Inference Eran Segal Weizmann Institute.
Genome evolution: a sequence-centric approach Lecture 5: Undirected models and variational inference.
Learning Low-Level Vision William T. Freeman Egon C. Pasztor Owen T. Carmichael.
Conditional Random Fields
Understanding Belief Propagation and its Applications Dan Yuan June 2004.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
. Applications and Summary. . Presented By Dan Geiger Journal Club of the Pharmacogenetics Group Meeting Technion.
. Expressive Graphical Models in Variational Approximations: Chain-Graphs and Hidden Variables Tal El-Hay & Nir Friedman School of Computer Science & Engineering.
Approximate Inference 2: Monte Carlo Markov Chain
Super-Resolution of Remotely-Sensed Images Using a Learning-Based Approach Isabelle Bégin and Frank P. Ferrie Abstract Super-resolution addresses the problem.
Reconstructing Relief Surfaces George Vogiatzis, Philip Torr, Steven Seitz and Roberto Cipolla BMVC 2004.
Background Subtraction based on Cooccurrence of Image Variations Seki, Wada, Fujiwara & Sumi Presented by: Alon Pakash & Gilad Karni.
Markov Random Fields Probabilistic Models for Images
28 February, 2003University of Glasgow1 Cluster Variation Method and Probabilistic Image Processing -- Loopy Belief Propagation -- Kazuyuki Tanaka Graduate.
Physics Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 12th Bayesian network and belief propagation in statistical inference Kazuyuki Tanaka.
Daphne Koller Message Passing Belief Propagation Algorithm Probabilistic Graphical Models Inference.
Topic Models Presented by Iulian Pruteanu Friday, July 28 th, 2006.
1 Mean Field and Variational Methods finishing off Graphical Models – Carlos Guestrin Carnegie Mellon University November 5 th, 2008 Readings: K&F:
Belief Propagation and its Generalizations Shane Oldenburger.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Maximum Entropy Model, Bayesian Networks, HMM, Markov Random Fields, (Hidden/Segmental) Conditional Random Fields.
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
Contextual models for object detection using boosted random fields by Antonio Torralba, Kevin P. Murphy and William T. Freeman.
Eric Xing © Eric CMU, Machine Learning Algorithms and Theory of Approximate Inference Eric Xing Lecture 15, August 15, 2010 Reading:
Efficient Belief Propagation for Image Restoration Qi Zhao Mar.22,2006.
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Daphne Koller Overview Conditional Probability Queries Probabilistic Graphical Models Inference.
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
Distributed cooperation and coordination using the Max-Sum algorithm
Bayesian Belief Propagation for Image Understanding David Rosenberg.
Graduate School of Information Sciences, Tohoku University
Learning Deep Generative Models by Ruslan Salakhutdinov
Statistical-Mechanical Approach to Probabilistic Image Processing -- Loopy Belief Propagation and Advanced Mean-Field Method -- Kazuyuki Tanaka and Noriko.
Multi-scale Visual Tracking by Sequential Belief Propagation
Markov Networks.
Integration and Graphical Models
Markov Random Fields Presented by: Vladan Radosavljevic.
Graduate School of Information Sciences, Tohoku University
Expectation-Maximization & Belief Propagation
Graduate School of Information Sciences, Tohoku University
Graduate School of Information Sciences, Tohoku University
Markov Networks.
Graduate School of Information Sciences, Tohoku University
Kazuyuki Tanaka Graduate School of Information Sciences
Presentation transcript:

Markov Networks: Theory and Applications Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL

2 Warm up … Given a directed graphical model Prove x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 p(x 2 |x 1 ) p(x 4 |x 2,x 3 ) p(x 6 |x 3 ) p(x 1 ) p(x 3 ) p(x 5 |x 4,x 6 )

3 Wake up ! Given a HMM: p(Z i |X i ) and p(X t |X t-1 ) Prove: x1x1 x2x2 xkxk x t-1 xtxt z1z1 z2z2 zkzk z t-1 ztzt …

4 Interesting problems (i) ? Super-resolution –Given a low-res image, can you recover a high-res image? –To make it feasible, you may be given a set of training pairs of low-res and high-res images.

5 Interesting problems (ii) Capturing body motion from video –Given a video sequence, you need to estimate the articulated body motion from the video.

6 Outline Markov network representation Three issues –Inferring hidden variables –Calculating evidence –Learning Markov networks Inference –Exact inference –Approximate inference Applications –Super-resolution –Capturing body articulation

7 Representations Undirected graphical models –G(V, E), v  V represents a random variable x, e  E indicate the relation of two random variables; –Each r.v. x i has a prior  i (x i ); –Each undirected link (between x i and x j ) is associated with a potential function  ij (x i, x j ); –v  V, N(v) means the neighborhood of v; We can combine undirected and directed graphical models.

8 Markov Network Undirected links Directed links

9 Three issues Inferring hidden variables –Given a Markov network  –To infer the hidden variable of the evidence p(X|Z,  ) Calculating evidence –Given a Markov network  –To calculate the likelihood of the evidence p(Z|  ) Learning Markov networks –Given a set of training data {Z 1,…,Z n } –To find the best model

10 To be covered The inference problem is the key among the three –Why? –Let’s discuss … We will focus on inference in this lecture

11 An example Given the above model, with –p(z i |x i ),  i (x i ) –  12 (x 1,x 2 ),  23 (x 2,x 3 ),  34 (x 3,x 4 ),  45 (x 4,x 5 ),  46 (x 4,x 6 ) To solve –p(x 3 |z 1,z 2,z 3,z 4,z 5,z 6 ) x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 z1z1 z2z2 z3z3 z4z4 z5z5 z6z6

12 An example “message” x 2  x 3 “message” x 4  x 3

13 What do we learn from the Ex.? x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 z1z1 z2z2 z3z3 z4z4 z5z5 z6z6 M 12 (x 2 ) M 23 (x 3 ) M 43 (x 3 ) M 54 (x 4 ) M 64 (x 4 )

14 What do we learn from the Ex.?

15 Sum-Product Algorithm Message passing xkxk xixi M ki (x i ) zkzk  ik (x i, x k )

16 Properties Properties of “belief propagation” –When the network is not loopy, the algorithm is exact; –When the network is loopy, the inference can be done by iterating the message passing procedures; –In this case, the inference is not exact, –and it is not clear how good is the approximation.

17 Application: Super-resolution Bill Freeman (ICCV’99) –X: high-frequency data –Z: mid-frequency data –“scene”: the high-frequency components –“evidence”: the mid-freq of the low-res image input –The “scene” is decomposed by a network of scene patches; and each scene patch is associated with an evidence patch –Learning p(z|x) from training samples scene patch evidence patch

18 Results (Freeman) + = Low-res input Iteration of B.P. to obtain high-frequency data Recovered high-res image actual high-res actual high-freq

19 Results (Freeman) Training data

20 Variational Inference We want to infer It is difficult, because of the networked structure. We perform probabilistic variational analysis The idea is to find an optimal approximation of, such that the Kullback-Leibler (KL) divergence of these two distribution is minimized:

21 Mean Field Theory When we choose a full factorization variation: We end up with a very interesting result: a set of fixed point equations: This is very similar to the Mean Field theory in statistical physics.

22 Computational Paradigm Three factors affect the posterior of a node: –Local prior –Neighborhood prior –Local likelihood

23 Application: body tracking Capturing body articulation (Wu et al.) –X: the motion of body parts –Z: image features (e.g., edges) –The motion of body parts are constrained and correlated –Specifying  (x i,x j ) –modeling p(z|x) explicitly

24 Dynamic Markov Network

25 Mean Field Iterations We are still investigating the convergence property of mean field algorithms But our empirical study shows that it converges very fast, generally, less than five iterations.

26 Application: body tracking

27 Application: body tracking

28 Summary Markov networks contain undirected edges in the graph to model the non-casual correlation Inference is the key of analyzing Markov networks –Exact inference –Approximate inference Two inference techniques were introduced here: –Belief propagation –Variational inference Two applications –Super-resolution –Capturing articulated body motion