Statistical Distributions

Slides:



Advertisements
Similar presentations
Request Dispatching for Cheap Energy Prices in Cloud Data Centers
Advertisements

SpringerLink Training Kit
Luminosity measurements at Hadron Colliders
From Word Embeddings To Document Distances
Choosing a Dental Plan Student Name
Virtual Environments and Computer Graphics
Chương 1: CÁC PHƯƠNG THỨC GIAO DỊCH TRÊN THỊ TRƯỜNG THẾ GIỚI
THỰC TIỄN KINH DOANH TRONG CỘNG ĐỒNG KINH TẾ ASEAN –
D. Phát triển thương hiệu
NHỮNG VẤN ĐỀ NỔI BẬT CỦA NỀN KINH TẾ VIỆT NAM GIAI ĐOẠN
Điều trị chống huyết khối trong tai biến mạch máu não
BÖnh Parkinson PGS.TS.BS NGUYỄN TRỌNG HƯNG BỆNH VIỆN LÃO KHOA TRUNG ƯƠNG TRƯỜNG ĐẠI HỌC Y HÀ NỘI Bác Ninh 2013.
Nasal Cannula X particulate mask
Evolving Architecture for Beyond the Standard Model
HF NOISE FILTERS PERFORMANCE
Electronics for Pedestrians – Passive Components –
Parameterization of Tabulated BRDFs Ian Mallett (me), Cem Yuksel
L-Systems and Affine Transformations
CMSC423: Bioinformatic Algorithms, Databases and Tools
Some aspect concerning the LMDZ dynamical core and its use
Bayesian Confidence Limits and Intervals
实习总结 (Internship Summary)
Current State of Japanese Economy under Negative Interest Rate and Proposed Remedies Naoyuki Yoshino Dean Asian Development Bank Institute Professor Emeritus,
Front End Electronics for SOI Monolithic Pixel Sensor
Face Recognition Monday, February 1, 2016.
Solving Rubik's Cube By: Etai Nativ.
CS284 Paper Presentation Arpad Kovacs
انتقال حرارت 2 خانم خسرویار.
Summer Student Program First results
Theoretical Results on Neutrinos
HERMESでのHard Exclusive生成過程による 核子内クォーク全角運動量についての研究
Wavelet Coherence & Cross-Wavelet Transform
yaSpMV: Yet Another SpMV Framework on GPUs
Creating Synthetic Microdata for Higher Educational Use in Japan: Reproduction of Distribution Type based on the Descriptive Statistics Kiyomi Shirakawa.
MOCLA02 Design of a Compact L-­band Transverse Deflecting Cavity with Arbitrary Polarizations for the SACLA Injector Sep. 14th, 2015 H. Maesaka, T. Asaka,
Hui Wang†*, Canturk Isci‡, Lavanya Subramanian*,
Fuel cell development program for electric vehicle
Overview of TST-2 Experiment
Optomechanics with atoms
داده کاوی سئوالات نمونه
Inter-system biases estimation in multi-GNSS relative positioning with GPS and Galileo Cecile Deprez and Rene Warnant University of Liege, Belgium  
ლექცია 4 - ფული და ინფლაცია
10. predavanje Novac i financijski sustav
Wissenschaftliche Aussprache zur Dissertation
FLUORECENCE MICROSCOPY SUPERRESOLUTION BLINK MICROSCOPY ON THE BASIS OF ENGINEERED DARK STATES* *Christian Steinhauer, Carsten Forthmann, Jan Vogelsang,
Particle acceleration during the gamma-ray flares of the Crab Nebular
Interpretations of the Derivative Gottfried Wilhelm Leibniz
Advisor: Chiuyuan Chen Student: Shao-Chun Lin
Widow Rockfish Assessment
SiW-ECAL Beam Test 2015 Kick-Off meeting
On Robust Neighbor Discovery in Mobile Wireless Networks
Chapter 6 并发:死锁和饥饿 Operating Systems: Internals and Design Principles
You NEED your book!!! Frequency Distribution
Y V =0 a V =V0 x b b V =0 z
Fairness-oriented Scheduling Support for Multicore Systems
Climate-Energy-Policy Interaction
Hui Wang†*, Canturk Isci‡, Lavanya Subramanian*,
Ch48 Statistics by Chtan FYHSKulai
The ABCD matrix for parabolic reflectors and its application to astigmatism free four-mirror cavities.
Measure Twice and Cut Once: Robust Dynamic Voltage Scaling for FPGAs
Online Learning: An Introduction
Factor Based Index of Systemic Stress (FISS)
What is Chemistry? Chemistry is: the study of matter & the changes it undergoes Composition Structure Properties Energy changes.
THE BERRY PHASE OF A BOGOLIUBOV QUASIPARTICLE IN AN ABRIKOSOV VORTEX*
Quantum-classical transition in optical twin beams and experimental applications to quantum metrology Ivano Ruo-Berchera Frascati.
The Toroidal Sporadic Source: Understanding Temporal Variations
FW 3.4: More Circle Practice
ارائه یک روش حل مبتنی بر استراتژی های تکاملی گروه بندی برای حل مسئله بسته بندی اقلام در ظروف
Decision Procedures Christoph M. Wintersteiger 9/11/2017 3:14 PM
Limits on Anomalous WWγ and WWZ Couplings from DØ
Presentation transcript:

Statistical Distributions

Why do we care about probability? Foundation of theory of statistics. Description of uncertainty associated with random variables. Measurement error Process error Needed to understand: Estimating model parameters. Model validation Prediction

Types of random variables Discrete random variables take only integer values (e.g. counts, memberships in categories). They are represented by probability mass functions. Continuous random variables are represented by probability density functions.

Terminology This notation is equivalent. You will see it all in this course. These are general stand-ins for distributions.

Probability Mass Functions For a discrete random variable, X, the probability that x takes on a value x is a discrete density function, f(x) also known as probability mass or distribution function. 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 0.2 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 Event (z) Probability S=support is the domain of function [z]

Probability Density Functions: Continuous variables A probability density function [z] gives the probability that a random variable Z takes on values within a range. [z]≥0 Pr(a ≤ z ≤ b) 𝑎 𝑏 𝑧 𝑑𝑧 −∞ ∞ 𝑧 𝑑𝑧 =1 Non-negative…

Probability Mass Functions: First Moment The expectation of a random variable z is the weighted value of the possible values that z can take, each value weighted by the probability that z assumes it. For discrete variables: Analogous to “center of gravity”. First moment. -1 0 1 2 p(-1)=0.10 p(0)=0.25 p(1)=0.3 p(2)=0.35

Probability Mass Functions: Second Central Moment The variance of a random variable reflects the spread of Z values around the expected value. For discrete variables: Second central moment of a distribution.

Continuous distributions

Cumulative distribution and quantile function Z takes on a value less than or equal to u

Probability Distributions & Stochasticity To build stochastic models for ecological data, we need a toolbox of [z]s. This toolbox contains probability functions and probability density functions that describe the way in which different types of data arise. The [z]’s link our deterministic model with the data in a way that reveals uncertainty.

A toolbox of [z]’s Discrete Bernoulli: binary outcome of one trial Binomial: Outcome of multiple trials. Poisson: Counts. Negative binomial: Overdispersed counts. Multinomial: Multiple categorical outcomes. Continuous Normal. Lognormal. Exponential Gamma Beta Others (Uniform, Dirilecht, Wishart).

Binomial distribution: Number of successes in n trials (Discrete events can take one of two values) p = 0.5 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 0.2 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 Event (xz Probability E[z] = np Variance =np(1-p) n = number of trials p = prob. of success Example: Probability of survival derived from population data

Binomial distribution: Number of successes in n trials (Discrete events can take one of two values) What are the data? What parameters do we want to estimate? What terms control the uncertainty? E[x] = np Variance =np(1-p) n = number of trials p = prob. of success 0 ≤ p ≤ 1 Data are count (z==y). Prob (p),

Binomial distribution

Bernoulli=Binomial with n=1 trial

Multinomial distribution: Number of successes in n trials with > 2 possible outcomes for “success”

Poisson Distribution: Counts (or getting hit in the head by a horse) 500 0.5 1 2 3 4 5 6 7 POISSON 100 200 300 400 Count 0.0 0.1 0.2 0.3 0.4 Proportion per Bar Number of Seedlings/quadrat y= number of units per sampling effort λ=average number of units **Alt param= λ=rt

Poisson distribution

Exercise What are the data? What parameters do we want to estimate? What terms control the uncertainty? Data=z’s, we want to estimate the lambda/////

Clustering in space or time (overdispersion) Poisson process E[z]=Variance[z] Negative binomial? Poisson process E[z]<Variance[z] Overdispersed Clumped or patchy

Negative binomial: Table 4.2 & 4.3 in H&M Bycatch Data E[Z]=0.279 Variance[Z]=1.56 Suggests temporal or spatial aggregation in the data

Negative Binomial: Counts 100 0.2 90 80 70 60 Count 50 0.1 Proportion per Bar 40 N=total number of trials. R=total number of successes P=Probability of success 30 20 10 0.0 10 20 30 40 50 Number of Seeds NEGBIN

Negative Binomial: Counts 100 0.2 90 80 70 60 Count 50 0.1 Proportion per Bar 40 30 20 10 0.0 10 20 30 40 50 Number of Seeds NEGBIN

Negative binomial

Normal Distribution E[x] = μ Variance = σ2 Normal PDF with mean = 0 y 0.2 0.4 0.6 0.8 1 -5 -4 -3 -2 -1 2 3 4 5 Prob(x) Var = 0.25 Var = 0.5 Var = 1 Var = 2 Var = 5 Var = 10 Reuls

Normal Distribution How do the data arise? Often, the process we model represents the sum of continuous variables (ex., growth increment in trees increases on soil C content). The fact that sums of things tend to be normally distributed is what stands behind the central limit theorem, which shows that the mean of random variables will be normally distributed even when the variables themselves follow other distributions. With large sample sizes, many other distributions (Poisson, Gamma etc) can be approximated by a normal

Normal Distribution What are the data? What parameters do we want to estimate? What terms control the uncertainty?

A JAGS aside Note that in JAGS, the normal distribution is parameterized in terms of precision, τ (tau). Precision is the inverse of variance, so a when we use a small precision as a prior, it’s equivalent to a large variance (i.e., the uncertainty for that parameter is large).

Lognormal: One tail and no negative values 0.8 0.1 0.2 0.3 0.4 0.5 0.6 0.7 10 20 30 40 50 60 70 y is always positive x Product of processes.

Lognormal: Radial growth data 1 2 3 4 HEMLOCK 50 100 150 Count 0.0 0.1 0.2 Proportion per Bar REDCEDAR 10 20 30 40 Growth (cm/yr) Red cedar Hemlock (Data from Date Creek, British Columbia)

Exponential 1 2 3 4 5 6 10 20 30 40 50 60 70 80 0.0 0.1 0.2 0.3 0.4 Proportion per Bar Count Data for weather events. Rainfall, probability of extreme events. Exponential is a special case of the gamma.. Variable

Exponential: Growth data (negatives assumed 0) 1 2 3 4 5 6 7 8 Growth (mm/yr) 200 400 600 800 1000 1200 Count 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Proportion per Bar Beilschemedia pendula (Data from BCI, Panama)

Gamma: One tail and flexibility mean.gamma<-mean(sapling$Growth) var.gamma<-var(sapling$Growth) # now match moments to distribution parameters. #get prior shape parameters for prior gamma distribution of lambda: a=mean.gamma^2/var.gamma # shape parameter b=mean.gamma/var.gamma #rate parameter

Gamma: “raw” growth data 1 2 3 4 5 6 7 8 9 Growth (mm/yr) 200 400 600 800 1000 Count 10 20 30 50 100 150 200 Alseis blackiana Cordia bicolor Growth (mm/yr) (Data from BCI, Panama)

Beta distribution: proportions Survival, proportions.

Beta: Light interception by crown trees (Data from Luquillo, PR)

Others Dirilecht: multivariate version of beta Multivariate normal

Bolker 2007

Bolker 2007

The Method of Moments You can calculate the sample values of the moments of the distributions and match them up with the theoretical moments. Recall that: The MOM is a good way to get a first (but biased) estimate of the parameters of a distribution.

Why is MOM important? We often need to account for uncertainty in our models but we only have data on sample mean and variance. How can we make predictions when the mean and variance are not the parameters of the distribution (i.e., the normal distribution)? The MOM can give us initial estimates of distribution parameters by matching moments (shape parameters) with sample mean and variance.

Haul & capture data (ED) What distribution? What are the parameters? What are the moments of this distribution? Catch per haul

MOM: Negative binomial We can get these from the data…

MOM Gamma Model biomass production (μ) as a function of rainfall (x). Production cannot be negative, so you need a distribution for data that are continuous and strictly positive. Moreover, a plot of the data shows that the spread of the residuals increases with increasing production, casting doubt on the assumption that variance is constant How do you represent uncertainty in production? Why is normal not appropriate?

MOM: Gamma Match moments (mean and variance) to parameters of gamma distribution, scale (α) and rate (β).

Mixture models What do you do when your data don’t fit any known distribution? Add covariates Mixture models Discrete Continuous

Zero-inflated models Zero-inflated models are a common type of finite mixture models. Combine a standard discrete probability distribution (e.g. binomial, Poisson, or negative binomial), which typically include some probability of sampling zero counts even when some individuals are present with some additional process that can also lead to a zero count (e.g. complete absence of the species or trap failure). Extremely useful in ecology.

An example: Seed predation y =no seeds taken out of N available t1 t2 ( ) Assume each seed has equal probability (p) of being taken. Then: Normalization constant

Zero-inflated binomial

Discrete mixtures

Discrete mixture: Zero-inflated binomial

Continuous mixture Seed retention times Turdus albicollis Regurgitation Defecation Combination of a lognormal and a truncated normal distribution Retention time (min) Uriarte et al. 2011

Many other distributions….. rvm(100,30,2) rose.diag Zimmerman et al. 2007

Some intuition for likelihood and stochasticity 0.14 0.12 0.10 0.08 P(z|θ) 0.06 0.04 0.02 0.00 -10 -5 5 10 Observations of z What is the probability of obtaining the observations (z) conditional on the value of θ?

Some intuition for likelihood and stochasticity 0.14 0.12 0.10 “Unlikely” values for θ 0.08 P(z|θ) 0.06 0.04 0.02 0.00 -10 -5 5 10 Observations of z

Some intuition for likelihood and stochasticity 0.14 0.12 0.10 The most “likely” Values for θ 0.08 P(z|θ) 0.06 0.04 0.02 0.00 -10 -5 5 10 Observations of z