Exercise 1: Example of latent variable - occupancy inference The Hobbiton council has recently expanded the hobbit-lands further west. However, it turned.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

SADC Course in Statistics Importance of the normal distribution (Session 09)
A Tutorial on Learning with Bayesian Networks
Bayesian Inference Chris Mathys Wellcome Trust Centre for Neuroimaging UCL SPM Course London, May 12, 2014 Thanks to Jean Daunizeau and Jérémie Mattout.
Getting the most out of insect-related data. A major issue for pollinator studies is to find out what affects the number of various insects. Example from.
Welcome to Amsterdam!. Bayesian Modeling for Cognitive Science: A WinBUGS Workshop.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
Bayesian Estimation in MARK
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Maximum Likelihood And Expectation Maximization Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Bayesian statistics 2 More on priors plus model choice.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
Markov Networks.
Bayesian statistics – MCMC techniques
Maximum Likelihood. Likelihood The likelihood is the probability of the data given the model.
Making rating curves - the Bayesian approach. Rating curves – what is wanted? A best estimate of the relationship between stage and discharge at a given.
Introduction  Bayesian methods are becoming very important in the cognitive sciences  Bayesian statistics is a framework for doing inference, in a principled.
Bayesian estimation Bayes’s theorem: prior, likelihood, posterior
Lecture 5: Learning models using EM
4-1 Statistical Inference The field of statistical inference consists of those methods used to make decisions or draw conclusions about a population.
CS 188: Artificial Intelligence Spring 2007 Lecture 14: Bayes Nets III 3/1/2007 Srini Narayanan – ICSI and UC Berkeley.
Visual Recognition Tutorial
. Approximate Inference Slides by Nir Friedman. When can we hope to approximate? Two situations: u Highly stochastic distributions “Far” evidence is discarded.
Computer vision: models, learning and inference
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Conditional & Joint Probability A brief digression back to joint probability: i.e. both events O and H occur Again, we can express joint probability in.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Mean Field Inference in Dependency Networks: An Empirical Study Daniel Lowd and Arash Shamaei University of Oregon.
Model Inference and Averaging
Bayesian Inference Ekaterina Lomakina TNU seminar: Bayesian inference 1 March 2013.
Exam I review Understanding the meaning of the terminology we use. Quick calculations that indicate understanding of the basis of methods. Many of the.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Introduction to Bayesian statistics Yves Moreau. Overview The Cox-Jaynes axioms Bayes’ rule Probabilistic models Maximum likelihood Maximum a posteriori.
Applied Bayesian Inference, KSU, April 29, 2012 § ❷ / §❷ An Introduction to Bayesian inference Robert J. Tempelman 1.
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Yung-Kyun Noh and Joo-kyung Kim Biointelligence.
Fast Simulators for Assessment and Propagation of Model Uncertainty* Jim Berger, M.J. Bayarri, German Molina June 20, 2001 SAMO 2001, Madrid *Project of.
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Bayesian statistics Probabilities for everything.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Slides for “Data Mining” by I. H. Witten and E. Frank.
CHAPTER 5 Probability Theory (continued) Introduction to Bayesian Networks.
Lecture 2: Statistical learning primer for biologists
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Trond Reitan (Division of statistics and insurance mathematics, Department of Mathematics, University of Oslo) Statistical modelling and latent variables.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
STAT03 - Descriptive statistics (cont.) - variability 1 Descriptive statistics (cont.) - variability Lecturer: Smilen Dimitrov Applied statistics for testing.
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Joo-kyung Kim Biointelligence Laboratory,
1 Getting started with WinBUGS Mei LU Graduate Research Assistant Dept. of Epidemiology, MD Anderson Cancer Center Some material was taken from James and.
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
Bayesian Estimation and Confidence Intervals Lecture XXII.
CS 2750: Machine Learning Directed Graphical Models
MCMC Output & Metropolis-Hastings Algorithm Part I
Latent Variables, Mixture Models and EM
Markov Networks.
More about Posterior Distributions
Learning Markov Networks
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Lecture 5 Unsupervised Learning in fully Observed Directed and Undirected Graphical Models.
CS 188: Artificial Intelligence Fall 2008
#21 Marginalize vs. Condition Uninteresting Fitted Parameters
Parametric Methods Berlin Chen, 2005 References:
Learning From Observed Data
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions July chonbuk national university.
Markov Networks.
Mathematical Foundations of BME Reza Shadmehr
Presentation transcript:

Exercise 1: Example of latent variable - occupancy inference The Hobbiton council has recently expanded the hobbit-lands further west. However, it turned out the land is infested with dragons. Of the 10kmx10km areas studied so far, 70% of them contained dragons. A standardized procedure for doing transects was developed so that each transect has the same detection probability, given occupancy. The Hobbiton biology department has found that the detection probability for dragon-infested areas is about 50% per transect. With no dragon, there is of course, no detection. ? Here be dragons No dragons ? Dragon Hobbit

Exercise 1: Dragon occupancy What’s the (marginal) probability that you’ll detect a dragon on any one transect? Show, using Bayes theorem, that the probability for having a dragon in the area given detection is 100%. Find the probability for there being a dragon in the area (occupancy) given that you didn’t detect anything. Could you expect this probability to drop from it’s previous value, even without knowing the specific occupancy rate and detection rate? Occupied Detection (and occupied) Occupied Detection (and occupied) Occupied or Model: Dragon occupancy (L)  Dragon detection (D) Since field biologists are getting scarce, the council has now reduced the number of transects to one per area.

Exercise 1: Extra Since Hobbiton field biologists are now getting *really* scarce, the transect procedure has been changed. You are now no longer require to poke the dragon with a stick before confirming detection. Visual detection from afar is now allowed. It is assumed this doesn’t change the detection rate given occupancy. As all biologists know, dragons and wyverns can be confused with each other when seen afar. It’s assumed this means the false positive rate (the probability of detection given no dragon occupancy) is 2%. What’s the overall detection rate now? Also, what’s the probability of occupancy given detection and given no detection now? Dragon Wyvern

Exercise 2: Salamander data using binomial model Detection data from a salamander species will be examined using a purely binomial model and using the occupancy model. There are A=39 areas. R-code for data can be found describing the data plus some likelihood-code: or Study the data using the binomial model (the issue of occupancy is ignored). a)Study the histogram of detection counts, k i. Are there signs of zero-inflation? b)Plot the likelihood function or it’s logarithm in it’s ordered form and try to do ML estimation graphically. c)Do the same for the unordered form (with the binomial coefficient). Discuss the differences and sameness of these plots. Will be inference be different? d)The ML estimate for p is k/n, where k=  k i, n=  n i. What’s the log-likelihood?

Exercise 3: Salamander data using occupancy Study the data using the occupancy model. a)Fetch the code for the likelihood (ordered version) at or write it yourself. b)Plot the likelihood surface ( the function now has two inputs,  and p ). You can use “contour” in R. Estimate the parameters graphically. c)Use numeric optimization to calculate the ML estimates for  and p. d)Find the likelihood value for these estimates. e)Use the likelihood ratio-test to test whether we need to reject the binomial model to the advantage of the occupancy model or not.

Exercise 4: Weta occupancy Weta is the name given to about 70 insect species endemic to New Zealand. Again, occupancy modelling is called for. You can use the WinBUGS-code for the salamander case to use as a template for these exercise. See and The data can be found in a WinBUGS digestable format in PS: The Weta data consists of detection counts rather than single detection indicators. You will thus have to change the likelihood-representation in the code (from Bernoulli on single detections to binomial for detection counts). a)Pre-exercise: Redo the analysis on the salamanders. b)Examine the data using the binomial model. Does the Markov chain seem to converge? Is it efficient? Look at the posterior distribution of the detection rate. Compare to the ML approach. c)Examine the data using the occupancy model. Again, comment on convergence and efficiency. Study the chains and the posterior distribution and compare to the ML approach. d)Extra: Redo the salamander data analysis and compare to the ML estimates in exercise 3.

Exercise 5: Weta occupancy with an explanation variable An expanded version with an explanation variable (goat browsing) is found here: Fetch the data with the “browsed” explanation variable. a)Change the model specification in WinBUGS so that there different occupancy rates and detection rates for the browsed and the unbrowsed areas. b)Study the convergence and efficiency. c)Compare the posterior of  and p for browsed and unbrowsed areas. Could the model be simplified while still having some dependency on the explanation variable?

Exercise 6: Weta occupancy model testing A template file for model comparison can be found here: a)Do Bayesian model comparison on binomial vs occupancy (easy), occupancy with or without explanation variable (hard) or all models (very hard). b)Which model would you recommend? c)Extra: Do frequentist model comparison and compare.