Advanced Statistical Computing Fall 2016

Slides:



Advertisements
Similar presentations
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Advertisements

Slice Sampling Radford M. Neal The Annals of Statistics (Vol. 31, No. 3, 2003)
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
Markov Chain Monte Carlo Convergence Diagnostics: A Comparative Review By Mary Kathryn Cowles and Bradley P. Carlin Presented by Yuting Qi 12/01/2006.
Bayesian Estimation in MARK
Ch 11. Sampling Models Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by I.-H. Lee Biointelligence Laboratory, Seoul National.
Gibbs Sampling Qianji Zheng Oct. 5th, 2010.
Computer Vision Lab. SNU Young Ki Baik An Introduction to MCMC for Machine Learning (Markov Chain Monte Carlo)
Variational Inference for Dirichlet Process Mixture Daniel Klein and Soravit Beer Changpinyo October 11, 2011 Applied Bayesian Nonparametrics Special Topics.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
1 Bayesian Methods with Monte Carlo Markov Chains II Henry Horng-Shing Lu Institute of Statistics National Chiao Tung University
Suggested readings Historical notes Markov chains MCMC details
Beam Sampling for the Infinite Hidden Markov Model Van Gael, et al. ICML 2008 Presented by Daniel Johnson.
Stochastic Collapsed Variational Bayesian Inference for Latent Dirichlet Allocation James Foulds 1, Levi Boyles 1, Christopher DuBois 2 Padhraic Smyth.
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
1 Engineering Computation Part 6. 2 Probability density function.
Particle filters (continued…). Recall Particle filters –Track state sequence x i given the measurements ( y 0, y 1, …., y i ) –Non-linear dynamics –Non-linear.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Particle Filters for Mobile Robot Localization 11/24/2006 Aliakbar Gorji Roborics Instructor: Dr. Shiri Amirkabir University of Technology.
Bayesian Analysis for Extreme Events Pao-Shin Chu and Xin Zhao Department of Meteorology School of Ocean & Earth Science & Technology University of Hawaii-
Incomplete Graphical Models Nan Hu. Outline Motivation K-means clustering Coordinate Descending algorithm Density estimation EM on unconditional mixture.
Bayes Factor Based on Han and Carlin (2001, JASA).
Material Model Parameter Identification via Markov Chain Monte Carlo Christian Knipprath 1 Alexandros A. Skordos – ACCIS,
Bayesian parameter estimation in cosmology with Population Monte Carlo By Darell Moodley (UKZN) Supervisor: Prof. K Moodley (UKZN) SKA Postgraduate conference,
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Adaptive CSMA under the SINR Model: Fast convergence using the Bethe Approximation Krishna Jagannathan IIT Madras (Joint work with) Peruru Subrahmanya.
Particle Filters for Shape Correspondence Presenter: Jingting Zeng.
1 Gil McVean Tuesday 24 th February 2009 Markov Chain Monte Carlo.
Computer Science, Software Engineering & Robotics Workshop, FGCU, April 27-28, 2012 Fault Prediction with Particle Filters by David Hatfield mentors: Dr.
Monte Carlo Methods1 T Special Course In Information Science II Tomas Ukkonen
Simulation of the matrix Bingham-von Mises- Fisher distribution, with applications to multivariate and relational data Discussion led by Chunping Wang.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
Suppressing Random Walks in Markov Chain Monte Carlo Using Ordered Overrelaxation Radford M. Neal 발표자 : 장 정 호.
Maximum a posteriori sequence estimation using Monte Carlo particle filters S. J. Godsill, A. Doucet, and M. West Annals of the Institute of Statistical.
CS Statistical Machine learning Lecture 24
MCMC (Part II) By Marc Sobel. Monte Carlo Exploration  Suppose we want to optimize a complicated distribution f(*). We assume ‘f’ is known up to a multiplicative.
IE 300, Fall 2012 Richard Sowers IESE. 8/30/2012 Goals: Rules of Probability Counting Equally likely Some examples.
Lecture 2: Statistical learning primer for biologists
Lecture #9: Introduction to Markov Chain Monte Carlo, part 3
Multi-target Detection in Sensor Networks Xiaoling Wang ECE691, Fall 2003.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
How can we maintain an error bound? Settle for a “per-step” bound What’s the probability of a mistake at each step? Not cumulative, but Equal footing with.
An Iterative Monte Carlo Method for Nonconjugate Bayesian Analysis B. P. Carlin and A. E. Gelfand Statistics and Computing 1991 A Generic Approach to Posterior.
Gil McVean, Department of Statistics Thursday February 12 th 2009 Monte Carlo simulation.
Introduction: Metropolis-Hasting Sampler Purpose--To draw samples from a probability distribution There are three steps 1Propose a move from x to y 2Accept.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
Introduction to Sampling based inference and MCMC
(joint work with Ai-ru Cheng, Ron Gallant, Beom Lee)
Why Stochastic Hydrology ?
Markov Chain Monte Carlo methods --the final project of stat 6213
Introduction to Monte Carlo Method
Bayesian Generalized Product Partition Model
Advanced Statistical Computing Fall 2016
Simple Instances of Swendson-Wang & RJMCMC
Jun Liu Department of Statistics Stanford University
Omiros Papaspiliopoulos and Gareth O. Roberts
Latent Variables, Mixture Models and EM
Markov Chain Monte Carlo
Expectation-Maximization
Kernel Stick-Breaking Process
Igor V. Cadez, Padhraic Smyth, Geoff J. Mclachlan, Christine and E
Location-Scale Normal Model
More about Posterior Distributions
Robust Full Bayesian Learning for Neural Networks
FDA – A Scalable Evolutionary Algorithm for the Optimization of Additively Decomposed Functions BISCuit EDA Seminar
Applied Statistics and Probability for Engineers
Presentation transcript:

Advanced Statistical Computing Fall 2016 Steve Qin

Outline Slice sampler Reversible jump Parallel tempering Collapsing, predictive updating Sequential Monte Carlo Convergence checking

Fundamental theorem of simulation Simulating X ~ f(x) is equivalent to simulating (X,U) ~ Uniform{(x,u): 0 < u < f(x)}. f is the marginal density of the joint distribution.

Slice sampler 2D slice sampler At iteration t, simulate 1. 2. with Neal (1997), Damien et al. (1999)

Examples Simple slice sampler. density function for x > 0. Truncated normal distribution.

The general slice sampler Suppose Introduce auxiliary variables ωi, such that f is the marginal distribution of the joint dist

Slice sampler algorithm At iteration t+1, simulate 1. 2. … k. k+1. where

Examples Truncated exponential E(β,a,b) Sample and set B(β,a,b) Sample Y from the above distribution, and set

Examples Standard normal introduce latent variable Y > 0, Gamma

Scale mixture of uniforms Normal If then student-t

Related algorithms Auxiliary variable algorithm Swendson and Wang Edwards and Sokal (1988) Swendson and Wang Swendson and Wang (1987)

Reversible jump Motivation: variable dimension models a “model where one of the things you do not know is the number of things you do not know.” –Peter Green. Bayesian model comparison and model selection.

An example Mixture modeling Order of an AR (p) model Number of components Order of an AR (p) model

Green’s algorithm At iteration t, if x(t) = (m,θm(t)), Select model Mn with probability πmn Generate umn ~ φmn(u) Set (θn ,vnm) ~ Tmn(θm(t), ,umn) Take θn(t) = θn with probability

Remarks The clever idea of RJMCMC is to supplement “smaller” space with artificial space. Dimension matching transform Tmn is flexible, but quite difficult to create and optimize, a serious drawback of the method. Methodologically brilliant but difficult to implement.

Convergence check Trace plot Autocorrelation plot Gelman and Rubin convergence measure r.