General Gibbs Distribution

Slides:



Advertisements
Similar presentations
1 Undirected Graphical Models Graphical Models – Carlos Guestrin Carnegie Mellon University October 29 th, 2008 Readings: K&F: 4.1, 4.2, 4.3, 4.4,
Advertisements

Markov Networks Alan Ritter.
Graphical Models BRML Chapter 4 1. the zoo of graphical models Markov networks Belief networks Chain graphs (Belief and Markov ) Factor graphs =>they.
Bayes Networks Markov Networks Noah Berlow. Bayesian -> Markov (Section 4.5.1) Given B, How can we turn into Markov Network? The general idea: – Convert.
Bayesian Networks. Contents Semantics and factorization Reasoning Patterns Flow of Probabilistic Influence.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
Bayesian Network Representation Continued
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Bayesian Networks Alan Ritter.
Approximate Inference 2: Monte Carlo Markov Chain
Lectures 2 – Oct 3, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday 12:00-1:20 Johnson Hall.
Generalizing Variable Elimination in Bayesian Networks 서울 시립대학원 전자 전기 컴퓨터 공학과 G 박민규.
Bayesian Network By Zhang Liliang. Key Point Today Intro to Bayesian Network Usage of Bayesian Network Reasoning BN: D-separation.
Slides for “Data Mining” by I. H. Witten and E. Frank.
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2008 Readings: K&F: 3.1, 3.2, –  Carlos.
Daphne Koller Markov Networks General Gibbs Distribution Probabilistic Graphical Models Representation.
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
1 BN Semantics 2 – Representation Theorem The revenge of d-separation Graphical Models – Carlos Guestrin Carnegie Mellon University September 17.
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
Reasoning Patterns Bayesian Networks Representation Probabilistic
Daphne Koller Introduction Motivation and Overview Probabilistic Graphical Models.
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2006 Readings: K&F: 3.1, 3.2, 3.3.
Daphne Koller Independencies Bayesian Networks Probabilistic Graphical Models Representation.
Context-Specific CPDs
The set  of all independence statements defined by (3
Conditional Random Fields
Bayesian Networks (Directed Acyclic Graphical Models)
Markov Networks.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Dependency Models – abstraction of Probability distributions
Independence in Markov Networks
Markov Networks Independencies Representation Probabilistic Graphical
General Gibbs Distribution
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17
Preliminaries: Distributions
Luger: Artificial Intelligence, 5th edition
Bayesian Networks Independencies Representation Probabilistic
Markov Networks.
Independence in Markov Networks
Pairwise Markov Networks
General Gibbs Distribution
Markov Random Fields Presented by: Vladan Radosavljevic.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11
I-equivalence Bayesian Networks Representation Probabilistic Graphical
Readings: K&F: 5.1, 5.2, 5.3, 5.4, 5.5, 5.6, 5.7 Markov networks, Factor graphs, and an unified view Start approximate inference If we are lucky… Graphical.
MCMC for PGMs: The Gibbs Chain
Conditional Random Fields
Probabilistic Influence & d-separation
Reasoning Patterns Bayesian Networks Representation Probabilistic
Factorization & Independence
Factorization & Independence
I-maps and perfect maps
Conditional Random Fields
BN Semantics 3 – Now it’s personal! Parameter Learning 1
Markov Networks Independencies Representation Probabilistic Graphical
I-maps and perfect maps
Junction Trees 3 Undirected Graphical Models
Tree-structured CPDs Local Structure Representation Probabilistic
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Crypto Encryption Intro to public key.
Independence in Markov Networks
Markov Networks Independencies Representation Probabilistic Graphical
Flow of Probabilistic Influence
Preliminaries: Independence
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Markov Networks.
Mean Field and Variational Methods Loopy Belief Propagation
Generalized Belief Propagation
BN Semantics 2 – The revenge of d-separation
Presentation transcript:

General Gibbs Distribution Representation Probabilistic Graphical Models Markov Networks General Gibbs Distribution

Consider a fully connected pairwise Markov network over X1,…,Xn where each Xi has d values. How many parameters does the network have? O(dn) O(nd) O(n2d2) O(nd)

Gibbs Distribution Parameters: a1 b1 c1 0.25 c2 0.35 b2 0.08 0.16 a2 0.05 0.07 a3 0.15 0.21 0.09 0.18 Parameters:

Gibbs Distribution

Markov Network Representation P factorizes over H

Separation in Undirected Graph H A trail between X and Y is active given Z X and Y are separated in H given Z if

Independence Assumptions in H The independencies implied by H I(H) = We say that H is an I-map (independence map) of P if Define I(G)

Factorization  Independence Theorem: If P factorizes over H then H is an I-map for P

Independence  Factorization Hammersley-Clifford Theorem: If H is an I-map for P, and P is positive, then P factorizes over H

Which parameterization of P factorizes over the graph H? D B C All of the above

Graph Structure & Factorization Factorization not unique, but same independencies

Summary Gibbs distribution represents distribution as a product of factors Associated Markov network connects every pair of nodes that are in the same factor Can read independencies that must hold in P from Markov network separation Markov network structure doesn’t fully specify the factorization of P

END END END

The Chain Rule for Bayesian Nets Intelligence Difficulty Grade Letter SAT 0.3 0.08 0.25 0.4 g2 0.02 0.9 i1,d0 0.7 0.05 i0,d1 0.5 g1 g3 0.2 i1,d1 i0,d0 l1 l0 0.99 0.1 0.01 0.6 0.95 s0 s1 0.8 i1 i0 d1 d0 P(D,I,G,S,L) = P(D) P(I) P(G | I,D) P(L | G) P(S | I)

Suppose q is at a local minimum of a function Suppose q is at a local minimum of a function. What will one iteration of gradient descent do? Leave q unchanged. Change q in a random direction. Move q towards the global minimum of J(q). Decrease q.

Fig. A corresponds to a=0.01, Fig. B to a=0.1, Fig. C to a=1.