Using MCMC Separating MCMC from Bayesian Inference? Line fitting revisited A toy equaliser problem Some lessons A problem in film restoration/retouching.

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

1 -Classification: Internal Uncertainty in petroleum reservoirs.
Unsupervised Learning Clustering K-Means. Recall: Key Components of Intelligent Agents Representation Language: Graph, Bayes Nets, Linear functions Inference.
Mixture Models and the EM Algorithm
Bayesian inference of normal distribution
Biologically Inspired Computation Really finishing off EC.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
Least squares CS1114
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Segment 4 Sampling Distributions - or - Statistics is really just a guessing game George Howard.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The FIR Adaptive Filter The LMS Adaptive Filter Stability and Convergence.
CHAPTER 2 D IRECT M ETHODS FOR S TOCHASTIC S EARCH Organization of chapter in ISSO –Introductory material –Random search methods Attributes of random search.
Visual Recognition Tutorial
EE-148 Expectation Maximization Markus Weber 5/11/99.
Making rating curves - the Bayesian approach. Rating curves – what is wanted? A best estimate of the relationship between stage and discharge at a given.
Hidden Markov Model 11/28/07. Bayes Rule The posterior distribution Select k with the largest posterior distribution. Minimizes the average misclassification.
Environmental Data Analysis with MatLab Lecture 5: Linear Models.
The University of Texas at Austin, CS 395T, Spring 2008, Prof. William H. Press IMPRS Summer School 2009, Prof. William H. Press 1 4th IMPRS Astronomy.
Lecture 5: Learning models using EM
Discrete Event Simulation How to generate RV according to a specified distribution? geometric Poisson etc. Example of a DEVS: repair problem.
Particle Filters for Mobile Robot Localization 11/24/2006 Aliakbar Gorji Roborics Instructor: Dr. Shiri Amirkabir University of Technology.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Learning Bayesian Networks
G. Cowan Lectures on Statistical Data Analysis Lecture 14 page 1 Statistical Data Analysis: Lecture 14 1Probability, Bayes’ theorem 2Random variables and.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1cs426-winter-2008 Notes  SIGGRAPH crunch time - my apologies :-)
Lecture 10: Robust fitting CS4670: Computer Vision Noah Snavely.
CHAPTER 15 S IMULATION - B ASED O PTIMIZATION II : S TOCHASTIC G RADIENT AND S AMPLE P ATH M ETHODS Organization of chapter in ISSO –Introduction to gradient.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Introduction to Monte Carlo Methods D.J.C. Mackay.
Bayes Factor Based on Han and Carlin (2001, JASA).
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Priors, Normal Models, Computing Posteriors
2 nd Order CFA Byrne Chapter 5. 2 nd Order Models The idea of a 2 nd order model (sometimes called a bi-factor model) is: – You have some latent variables.
Numerical Methods Applications of Loops: The power of MATLAB Mathematics + Coding 1.
Choosing Sample Size for Knowledge Tracing Models DERRICK COETZEE.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Bayesian Inversion of Stokes Profiles A.Asensio Ramos (IAC) M. J. Martínez González (LERMA) J. A. Rubiño Martín (IAC) Beaulieu Workshop ( Beaulieu sur.
1 The Scientist Game Chris Slaughter, DrPH (courtesy of Scott Emerson) Dept of Biostatistics Vanderbilt University © 2002, 2003, 2006, 2008 Scott S. Emerson,
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
A unifying framework for hybrid data-assimilation schemes Peter Jan van Leeuwen Data Assimilation Research Center (DARC) National Centre for Earth Observation.
STAR Sti, main features V. Perevoztchikov Brookhaven National Laboratory,USA.
2005MEE Software Engineering Lecture 11 – Optimisation Techniques.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Bayesian Reasoning: Tempering & Sampling A/Prof Geraint F. Lewis Rm 560:
STAR Kalman Track Fit V. Perevoztchikov Brookhaven National Laboratory,USA.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
Statistics What is the probability that 7 heads will be observed in 10 tosses of a fair coin? This is a ________ problem. Have probabilities on a fundamental.
Week 5-6 MondayTuesdayWednesdayThursdayFriday Testing III No reading Group meetings Testing IVSection ZFR due ZFR demos Progress report due Readings out.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Plan for Today: Chapter 1: Where Do Data Come From? Chapter 2: Samples, Good and Bad Chapter 3: What Do Samples Tell US? Chapter 4: Sample Surveys in the.
Unpacking each and every strategy! THE MATHEMATICIAN’S TOOLBOX.
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
SIR method continued. SIR: sample-importance resampling Find maximum likelihood (best likelihood × prior), Y Randomly sample pairs of r and N 1973 For.
How many iterations in the Gibbs sampler? Adrian E. Raftery and Steven Lewis (September, 1991) Duke University Machine Learning Group Presented by Iulian.
Markov Chain Monte Carlo in R
Fast search for Dirichlet process mixture models
Introducing Bayesian Approaches to Twin Data Analysis
Classification of unlabeled data:
Lecture 7: Image alignment
Latent Variables, Mixture Models and EM
Probabilistic Models for Linear Regression
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
PSY 626: Bayesian Statistics for Psychological Science
Learning From Observed Data
Presentation transcript:

Using MCMC Separating MCMC from Bayesian Inference? Line fitting revisited A toy equaliser problem Some lessons A problem in film restoration/retouching

MCMC is (just) a tool Choose a ModelIdentify Parameters Articulate Probabilities [Bayesian Inference] Try to see if you can integrate out nuisances Derive the Posterior Solve Deterministically Direct, CG, Steepest Descent etc Use MCMC Manipulate Random Samples To get one answer (if you want) Gives you one answer Need better model If solution not ok

Good, Bad Can always design single parameter-at-a-time schemes. So iterations can be very low complexity Simple iterations = long convergence Gives you a picture of alternate answers Do you really need alternate answers? Will always allow you to get to “best” solution Iterations can be high complexity? Convergence can be rapid (e.g. CG) for well defined problems Gives you just one answer Can give local minimum for non-linear problems MCMC Others

Ugly To solve your problem you need a good model MCMC is not really going to help you if you have the wrong model MCMC suited to BIG problems: but what is BIG really? E.g. Exhaustive search for motion estimation is possible in real time (TV rates) in hardware: why bother with other things? (an exaggeration … but interesting nevertheless)

Line Fitting (again) Needs Latex Observed Data Actual Line Initial Guess

Typical Results Nice Convergence because we can draw samples directly See Matlab demo

Typical Results c m var_e

Watch out All random number generators are not created equal (See NR) Harder problems require longer runs (of course) Sometimes hard to get all bugs out because its all a random search anyway

Blind (?) Equalisation Signal2 nd Order All pole System Noise Rec’d Signal Identify the system coefficients AND recover the original signal Comms, Deblurring, Overshoot Cancellation

Equaliser Problem Now more latex

Direct numerical sampling Number line interpretation P(1) = 0.3, p(2) = 0.25, P(3) = 0.2, p(4) = points evaluated

Gibbs sampler 1 (equaliser1.m) Typical Estimated System Actual System X 20 ! 300 iterations Back to Latex

Gibbs Sampler II (equaliser2.m) Using Filter Bank (system choices) 30 filters

Samples from filter bank

Samples of signal parameter

System Estimate

Equalised signal

Lessons Gibbs sampler takes big problems and breaks them into lots of small ones Spotting the functional form of a known p.d.f. is a useful skill. Books help. If all else fails, can always sample directly MCMC does not necessarily solve your problem. Good priors, better models are still important Deterministic/Stochastic Hybrid mix is v. useful