MCMC for PGMs: The Gibbs Chain

Slides:



Advertisements
Similar presentations
Markov models and HMMs. Probabilistic Inference P(X,H,E) P(X|E=e) = P(X,E=e) / P(E=e) with P(X,E=e) = sum h P(X, H=h, E=e) P(X,E=e) = sum h,x P(X=x, H=h,
Advertisements

Motivating Markov Chain Monte Carlo for Multiple Target Tracking
Markov Chain Sampling Methods for Dirichlet Process Mixture Models R.M. Neal Summarized by Joon Shik Kim (Thu) Computational Models of Intelligence.
A*-tree: A Structure for Storage and Modeling of Uncertain Multidimensional Arrays Presented by: ZHANG Xiaofei March 2, 2011.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
State Estimation and Kalman Filtering CS B659 Spring 2013 Kris Hauser.
Introduction of Markov Chain Monte Carlo Jeongkyun Lee.
Gibbs Sampling Qianji Zheng Oct. 5th, 2010.
Bayesian Networks. Contents Semantics and factorization Reasoning Patterns Flow of Probabilistic Influence.
Stochastic approximate inference Kay H. Brodersen Computational Neuroeconomics Group Department of Economics University of Zurich Machine Learning and.
BAYESIAN INFERENCE Sampling techniques
Beam Sampling for the Infinite Hidden Markov Model Van Gael, et al. ICML 2008 Presented by Daniel Johnson.
Approximate Counting via Correlation Decay Pinyan Lu Microsoft Research.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Discrete random variables Probability mass function Distribution function (Secs )
Conditional Random Fields
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference.
. Approximate Inference Slides by Nir Friedman. When can we hope to approximate? Two situations: u Highly stochastic distributions “Far” evidence is discarded.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Sampling Combinatorial Space Using Biased Random Walks Jordan Erenrich, Wei Wei and Bart Selman Dept. of Computer Science Cornell University.
Approximate Inference 2: Monte Carlo Markov Chain
1 Approximate Inference 2: Importance Sampling. (Unnormalized) Importance Sampling.
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
An Introduction to Markov Chain Monte Carlo Teg Grenager July 1, 2004.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
Lecture #9: Introduction to Markov Chain Monte Carlo, part 3
Inference of Non-Overlapping Camera Network Topology by Measuring Statistical Dependence Date :
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
Daphne Koller Overview Conditional Probability Queries Probabilistic Graphical Models Inference.
1 Structure Learning (The Good), The Bad, The Ugly Inference Graphical Models – Carlos Guestrin Carnegie Mellon University October 13 th, 2008 Readings:
Reasoning Patterns Bayesian Networks Representation Probabilistic
Daphne Koller Sampling Methods Metropolis- Hastings Algorithm Probabilistic Graphical Models Inference.
Tea – Time - Talks Every Friday 3.30 pm ICS 432. We Need Speakers (you)! Please volunteer. Philosophy: a TTT (tea-time-talk) should approximately take.
How many iterations in the Gibbs sampler? Adrian E. Raftery and Steven Lewis (September, 1991) Duke University Machine Learning Group Presented by Iulian.
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
Maximum Expected Utility
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 12
Conditional Random Fields
Markov Networks.
Markov Chain Monte Carlo
Synthesis of MCMC and Belief Propagation
General Gibbs Distribution
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 12
Instructors: Fei Fang (This Lecture) and Dave Touretzky
General Gibbs Distribution
Preliminaries: Distributions
Bayesian Inference for Mixture Language Models
CS 188: Artificial Intelligence
Markov Chain Monte Carlo
General Gibbs Distribution
Lecture 15 Sampling.
Simple Sampling Sampling Methods Inference Probabilistic Graphical
I-equivalence Bayesian Networks Representation Probabilistic Graphical
Junghoo “John” Cho UCLA
Readings: K&F: 5.1, 5.2, 5.3, 5.4, 5.5, 5.6, 5.7 Markov networks, Factor graphs, and an unified view Start approximate inference If we are lucky… Graphical.
Inference III: Approximate Inference
Conditional Random Fields
Approximate Inference by Sampling
Probabilistic Influence & d-separation
Reasoning Patterns Bayesian Networks Representation Probabilistic
Preliminaries: Factors
Conditional Random Fields
Approximate Inference: Particle-Based Methods
Flow of Probabilistic Influence
Markov Networks.
Discrete-time markov chain (continuation)
Generalized Belief Propagation
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 12
Presentation transcript:

MCMC for PGMs: The Gibbs Chain Inference Probabilistic Graphical Models Sampling Methods MCMC for PGMs: The Gibbs Chain

Gibbs Chain Target distribution P(X1,…,Xn) Markov chain state space: complete assignments x to X = {X1,…,Xn} Transition model given starting state x: For i=1,…,n Sample xi ~ P(Xi | x-i) Set x’ = x

Example Intelligence Difficulty Grade Letter SAT s0 s1 i0 0.95 0.05 i1 0.6 0.4 i0 i1 0.7 0.3 Intelligence Difficulty Grade Letter SAT g1 g2 g3 i0,d0 0.3 0.4 i0,d1 0.05 0.25 0.7 i1,d0 0.9 0.08 0.02 i1,d1 0.5 0.2 s0 s1 i0 0.95 0.05 i1 0.2 0.8 l0 l1 g1 0.1 0.9 g2 0.4 0.6 g3 0.99 0.01

Computational Cost For i=1,…,n Sample xi ~ P(Xi | x-i)

Another Example A D B C

Computational Cost Revisited For i=1,…,n Sample xi ~ P(Xi | x-i)

Gibbs Chain and Regularity X1 X2 Y Prob 0.25 1 X1 X2 Y XOR If all factors are positive, Gibbs chain is regular However, mixing can still be very slow

Summary Converts the hard problem of inference to a sequence of “easy” sampling steps Pros: Probably the simplest Markov chain for PGMs Computationally efficient to sample Cons: Often slow to mix, esp. when probabilities are peaked Only applies if we can sample from product of factors

END END END