Markov Networks.

Slides:



Advertisements
Similar presentations
Linear Time Methods for Propagating Beliefs Min Convolution, Distance Transforms and Box Sums Daniel Huttenlocher Computer Science Department December,
Advertisements

Markov Networks Alan Ritter.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Graphical Models BRML Chapter 4 1. the zoo of graphical models Markov networks Belief networks Chain graphs (Belief and Markov ) Factor graphs =>they.
Exact Inference in Bayes Nets
Junction Trees And Belief Propagation. Junction Trees: Motivation What if we want to compute all marginals, not just one? Doing variable elimination for.
Undirected Probabilistic Graphical Models (Markov Nets) (Slides from Sam Roweis)
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
Overview of Inference Algorithms for Bayesian Networks Wei Sun, PhD Assistant Research Professor SEOR Dept. & C4I Center George Mason University, 2009.
Markov Networks.
BAYESIAN INFERENCE Sampling techniques
Exact Inference (Last Class) variable elimination  polytrees (directed graph with at most one undirected path between any two vertices; subset of DAGs)
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Lecture 5: Learning models using EM
CS 561, Session 29 1 Belief networks Conditional independence Syntax and semantics Exact inference Approximate inference.
CS 188: Artificial Intelligence Spring 2007 Lecture 14: Bayes Nets III 3/1/2007 Srini Narayanan – ICSI and UC Berkeley.
Announcements Homework 8 is out Final Contest (Optional)
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Approximate Inference 2: Monte Carlo Markov Chain
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by B.-H. Kim Biointelligence Laboratory, Seoul National.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Markov Random Fields Probabilistic Models for Images
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Bayes’ Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Exact Inference (Last Class) Variable elimination  polytrees (directed graph with at most one undirected path between any two vertices; subset of DAGs)
1 Mean Field and Variational Methods finishing off Graphical Models – Carlos Guestrin Carnegie Mellon University November 5 th, 2008 Readings: K&F:
Lecture 2: Statistical learning primer for biologists
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
CS 188: Artificial Intelligence Bayes Nets: Approximate Inference Instructor: Stuart Russell--- University of California, Berkeley.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Inference Algorithms for Bayes Networks
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Pattern Recognition and Machine Learning
Daphne Koller Overview Conditional Probability Queries Probabilistic Graphical Models Inference.
Bayesian Belief Propagation for Image Understanding David Rosenberg.
CS 541: Artificial Intelligence Lecture VII: Inference in Bayesian Networks.
10 October, 2007 University of Glasgow 1 EM Algorithm with Markov Chain Monte Carlo Method for Bayesian Image Analysis Kazuyuki Tanaka Graduate School.
CS 4/527: Artificial Intelligence
The set  of all independence statements defined by (3
Prof. Adriana Kovashka University of Pittsburgh April 4, 2017
CAP 5636 – Advanced Artificial Intelligence
CSCI 5822 Probabilistic Models of Human and Machine Learning
Learning Markov Networks
Independence in Markov Networks
Instructors: Fei Fang (This Lecture) and Dave Touretzky
Lecture 5 Unsupervised Learning in fully Observed Directed and Undirected Graphical Models.
Markov Networks.
CS 188: Artificial Intelligence
Independence in Markov Networks
Class #19 – Tuesday, November 3
Lecture 15 Sampling.
CS 188: Artificial Intelligence Fall 2008
Expectation-Maximization & Belief Propagation
Class #16 – Tuesday, October 26
Readings: K&F: 5.1, 5.2, 5.3, 5.4, 5.5, 5.6, 5.7 Markov networks, Factor graphs, and an unified view Start approximate inference If we are lucky… Graphical.
Approximate Inference by Sampling
Markov Networks Independencies Representation Probabilistic Graphical
Junction Trees 3 Undirected Graphical Models
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Independence in Markov Networks
Markov Networks.
Mean Field and Variational Methods Loopy Belief Propagation
Sequential Learning with Dependency Nets
Generalized Belief Propagation
Presentation transcript:

Markov Networks

Markov Networks A B C D Undirected graphical models Potential functions defined over cliques

Markov Networks A B C D Undirected graphical models Potential functions defined over cliques Weight of Feature i Feature i

Hammersley-Clifford Theorem If Distribution is strictly positive (P(x) > 0) And Graph encodes conditional independences Then Distribution is product of potentials over cliques of graph Inverse is also true. (“Markov network = Gibbs distribution”)

Markov Nets vs. Bayes Nets Property Markov Nets Bayes Nets Form Prod. potentials Potentials Arbitrary Cond. probabilities Cycles Allowed Forbidden Partition func. Z = ? Z = 1 Indep. check Graph separation D-separation Indep. props. Some Inference MCMC, BP, etc. Convert to Markov

Inference in Markov Networks Goal: compute marginals & conditionals of Exact inference is #P-complete Conditioning on Markov blanket is easy: Gibbs sampling exploits this

Markov Chain Monte Carlo Gibbs Sampler 1. Start with an initial assignment to nodes 2. One node at a time, sample node given others 3. Repeat 4. Use samples to compute P(X) Convergence: Burn-in + Mixing time Many modes  Multiple chains

Other Inference Methods Belief propagation (sum-product) Mean field / Variational approximations

MAP Inference Iterated conditional modes Simulated annealing Graph cuts Belief propagation (max-product)

Learning Weights Maximize likelihood (or posterior) Convex optimization: gradient ascent, quasi-Newton methods, etc. Requires inference at each step (slow!) Feature count according to data Feature count according to model

Pseudo-Likelihood Likelihood of each variable given its Markov blanket in the data Does not require inference at each step Very fast numerical optimization Ignores non-local dependences

Learning Structure Feature search Evaluation 1. Start with atomic features 2. Form conjunctions of pairs of features 3. Select best and add to feature set 4. Repeat until no improvement Evaluation Likelihood, K-L divergence Approximation: Previous weights don’t change