Gibbs Sampling Qianji Zheng Oct. 5th, 2010.

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
State Estimation and Kalman Filtering CS B659 Spring 2013 Kris Hauser.
Markov Chain Monte Carlo Convergence Diagnostics: A Comparative Review By Mary Kathryn Cowles and Bradley P. Carlin Presented by Yuting Qi 12/01/2006.
Bayesian Estimation in MARK
Introduction of Markov Chain Monte Carlo Jeongkyun Lee.
Ch 11. Sampling Models Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by I.-H. Lee Biointelligence Laboratory, Seoul National.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Graduate School of Information Sciences, Tohoku University
Suggested readings Historical notes Markov chains MCMC details
Stochastic approximate inference Kay H. Brodersen Computational Neuroeconomics Group Department of Economics University of Zurich Machine Learning and.
BAYESIAN INFERENCE Sampling techniques
Industrial Engineering College of Engineering Bayesian Kernel Methods for Binary Classification and Online Learning Problems Theodore Trafalis Workshop.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
The University of Texas at Austin, CS 395T, Spring 2008, Prof. William H. Press IMPRS Summer School 2009, Prof. William H. Press 1 4th IMPRS Astronomy.
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
CS 561, Session 29 1 Belief networks Conditional independence Syntax and semantics Exact inference Approximate inference.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Bayesian Analysis for Extreme Events Pao-Shin Chu and Xin Zhao Department of Meteorology School of Ocean & Earth Science & Technology University of Hawaii-
Analyzing iterated learning Tom Griffiths Brown University Mike Kalish University of Louisiana.
Image Analysis and Markov Random Fields (MRFs) Quanren Xiong.
Bayes Factor Based on Han and Carlin (2001, JASA).
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
1 Gil McVean Tuesday 24 th February 2009 Markov Chain Monte Carlo.
Monte Carlo Methods1 T Special Course In Information Science II Tomas Ukkonen
Simulation of the matrix Bingham-von Mises- Fisher distribution, with applications to multivariate and relational data Discussion led by Chunping Wang.
Model-based Bayesian Reinforcement Learning in Partially Observable Domains by Pascal Poupart and Nikos Vlassis (2008 International Symposium on Artificial.
Markov Random Fields Probabilistic Models for Images
Improved Cross Entropy Method For Estimation Presented by: Alex & Yanna.
Gibbs sampling for motif finding Yves Moreau. 2 Overview Markov Chain Monte Carlo Gibbs sampling Motif finding in cis-regulatory DNA Biclustering microarray.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Learning to Detect Events with Markov-Modulated Poisson Processes Ihler, Hutchins and Smyth (2007)
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
An Introduction to Markov Chain Monte Carlo Teg Grenager July 1, 2004.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
Lecture #9: Introduction to Markov Chain Monte Carlo, part 3
The Markov Chain Monte Carlo Method Isabelle Stanton May 8, 2008 Theory Lunch.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
A shared random effects transition model for longitudinal count data with informative missingness Jinhui Li Joint work with Yingnian Wu, Xiaowei Yang.
CS774. Markov Random Field : Theory and Application Lecture 15 Kyomin Jung KAIST Oct
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
An Iterative Monte Carlo Method for Nonconjugate Bayesian Analysis B. P. Carlin and A. E. Gelfand Statistics and Computing 1991 A Generic Approach to Posterior.
HW7: Evolutionarily conserved segments ENCODE region 009 (beta-globin locus) Multiple alignment of human, dog, and mouse 2 states: neutral (fast-evolving),
Efficiency Measurement William Greene Stern School of Business New York University.
How many iterations in the Gibbs sampler? Adrian E. Raftery and Steven Lewis (September, 1991) Duke University Machine Learning Group Presented by Iulian.
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Introduction to Sampling based inference and MCMC
MCMC Output & Metropolis-Hastings Algorithm Part I
Markov Chain Monte Carlo methods --the final project of stat 6213
Advanced Statistical Computing Fall 2016
Introducing Bayesian Approaches to Twin Data Analysis
Jun Liu Department of Statistics Stanford University
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
Markov Networks.
Bayesian Network Reasoning with Gibbs Sampling
Multidimensional Integration Part I
Instructors: Fei Fang (This Lecture) and Dave Touretzky
Ch13 Empirical Methods.
Robust Full Bayesian Learning for Neural Networks
Markov Networks.
Presentation transcript:

Gibbs Sampling Qianji Zheng Oct. 5th, 2010

Outline Motivation & Basic Idea Algorithm Example Applications Why Gibbs Works

Gibbs Sampling: Motivation Gibbs sampling is a particular form of Markov chain Monte Carlo (MCMC) algorithm for approximating joint and marginal distribution by sampling from conditional distributions.

Gibbs Sampling: Basic Idea If the joint distribution is not known explicitly or is difficult to sample from directly, but the conditional distribution is known or easy to sample from. Even if the joint distribution is known, the computational burden needed to calculate it may be huge. Gibbs Sampling algorithm could generate a sequence of samples from conditional individual distributions, which constitutes a Markov chain, to approximate the joint distribution.

Characteristics of Gibbs Sampling Algorithm A particular Markov Chain Monte Carlo (MCMC) algorithm Sample from conditional distribution while other parameters are fixed Update a single parameter at a time

Gibbs Sampling Algorithm Let be the conditional distribution of the element given all the other parameters minus the , then Gibbs Sampling for an m-component variable is given by the transition from to generated as: Given an arbitrary initial value

Gibbs Sampling Algorithm Contd… Steps 1 through m can be iterated J times to get , j = 1, 2, … , J. 2. The joint and marginal distributions of generated converge at an exponential rate to joint and marginal distribution of , as . 3. Then the joint and marginal distributions of can be approximated by the empirical distributions of M simulated values (j=L+1,…, L+M). The mean of the marginal distribution of may be approximated by

Gibbs Sampling Algorithm In BN

Example

Example Example refer to Gibbs Sampling for Approximate Inference in Bayesian Networks http://www-users.cselabs.umn.edu/classes/Spring-2010/csci5512/notes/gibbs.pdf

Gibbs Sampling: Applications Gibbs Sampling algorithm has been widely used on a broad class of areas, e.g. , Bayesian networks, statistical inference, bioinformatics, econometrics. The power of Gibbs Sampling is: 1. Approximate joint and marginal distribution 2. Estimate unknown parameters 3. Compute an integral (e.g. mean, median, etc)

Why Gibbs Works The Gibbs sampling can simulate the target distribution by constructing a Gibbs sequence which converges to a stationary distribution that is independent of the starting value. The stationary distribution is the target distribution.

Online Resources Reference Gibbs Sampling for Approximate Inference in Bayesian Networks http://www-users.cselabs.umn.edu/classes/Spring-2010/csci5512/notes/gibbs.pdf Markov Chain Monte Carlo and Gibbs Sampling http://membres-timc.imag.fr/Olivier.Francois/mcmc_gibbs_sampling.pdf Markov Chains, the Gibbs Sampler and Data Augmentation http://athens.src.uchicago.edu/jenni/econ350/Salvador/h4.pdf Reference Kim, C. J. and Nelson, C. R. (1999), State-Space Models with Regime Switching, Cambridge, Massachusetts: MIT Press.