Stockholm, Sweden24-28 March 2014 Introduction to MAR(1) models for estimating community interactions This lecture follows largely from: Ives AR, Dennis.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Model specification (identification) We already know about the sample autocorrelation function (SAC): Properties: Not unbiased (since a ratio between two.
Nonlinear Regression Ecole Nationale Vétérinaire de Toulouse Didier Concordet ECVPT Workshop April 2011 Can be downloaded at
Empirical Maximum Likelihood and Stochastic Process Lecture VIII.
Université d’Ottawa / University of Ottawa 2001 Bio 4118 Applied Biostatistics L10.1 CorrelationCorrelation The underlying principle of correlation analysis.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
SYSTEMS Identification
Nonlinear Regression Probability and Statistics Boris Gervits.
Prediction and model selection
Mixed models Various types of models and their relation
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation X = {
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation Given.
Modern Navigation Thomas Herring
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
1 MULTI VARIATE VARIABLE n-th OBJECT m-th VARIABLE.
Chapter 4-5: Analytical Solutions to OLS
Model Inference and Averaging
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Introduction to Linear Regression
Module 1: Statistical Issues in Micro simulation Paul Sousa.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Modern Navigation Thomas Herring
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
Count based PVA Incorporating Density Dependence, Demographic Stochasticity, Correlated Environments, Catastrophes and Bonanzas.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Implementing the MAR-1 Algorithm A conceptual walkthrough.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
Estimation Method of Moments (MM) Methods of Moment estimation is a general method where equations for estimating parameters are found by equating population.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Machine Learning 5. Parametric Methods.
ALISON BOWLING MAXIMUM LIKELIHOOD. GENERAL LINEAR MODEL.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
AUTOCORRELATION 1 Assumption B.5 states that the values of the disturbance term in the observations in the sample are generated independently of each other.
Computacion Inteligente Least-Square Methods for System Identification.
Econometrics III Evgeniya Anatolievna Kolomak, Professor.
Environmental Data Analysis with MatLab 2 nd Edition Lecture 22: Linear Approximations and Non Linear Least Squares.
Geology 5670/6670 Inverse Theory 6 Feb 2015 © A.R. Lowry 2015 Read for Mon 9 Feb: Menke Ch 5 (89-114) Last time: The Generalized Inverse; Damped LS The.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Estimating the effects of covariates and seasonal effects in state-space models Applied Time Series Analysis for Ecologists Stockholm, Sweden24-28 March.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Statistical Interpretation of Least Squares ASEN.
FISH 507 – Applied Time Series Analysis
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves.
The simple linear regression model and parameter estimation
(5) Notes on the Least Squares Estimate
Probability Theory and Parameter Estimation I
FISH 507 – Applied Time Series Analysis
Model Inference and Averaging
Ch3: Model Building through Regression
CH 5: Multivariate Methods
Predictive distributions
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Simple Linear Regression
Multivariate Linear Regression
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves.
Parametric Methods Berlin Chen, 2005 References:
Probabilistic Surrogate Models
Presentation transcript:

Stockholm, Sweden24-28 March 2014 Introduction to MAR(1) models for estimating community interactions This lecture follows largely from: Ives AR, Dennis B, Cottingham KL, Carpenter SR Estimating community stability and ecological interactions from time series data. Ecological Monographs 73: Applied Time Series Analysis for Ecologists

Topics for this afternoon MAR(1) models as sum of community interactions MAR(1) models as a general approximation to nonlinear systems Stability properties of MAR(1) models Estimating MAR(1) models

“The Lake Washington Story” Hampton et al. (2006)

How can we make sense of this? Hampton et al. (2006)Dashed lines are (+) interactions; solid lines are (-)

How can we make sense of this? Hampton et al. (2006)Dashed lines are (+) interactions; solid lines are (-)

Multivariate AR(1) process Assume p different interacting species, then for each sp. i : X j,t-1 is abundance of species j at time t-1 b ij is effect of species j on species i Note: when i = j, b ij is effect of species i on itself

Multivariate AR(1) process Can add effects of q different environmental covariates ( U ): c k,t is covariate k (e.g., temperature) at time t C ik is effect of covariate k on species i NOTE: covariate not time lagged

Multivariate AR(1) process Can write all of this more compactly as: x t is p x 1 vector of species abundance at time t u is p x 1 vector of intrinsic growth rates B is p x p matrix of density-dependent effects x t-1 is p x 1 vector of species abundance at time t-1 C is p x q matrix of covariate effects c t is q x 1 vector of covariate values at time t

Multivariate AR(1) process A note on density-dependence: Intra-specific effects Inter-specific effects (can be set to zero) Inter-specific effects (can be set to zero)

MAR(1) models & nonlinear systems Consider the model: where: R t is a q x 1 normal random variable with mean 0 f is a nonlinear function MAR(1) models can be seen as linear approximations to nonlinear, first-order stochastic processes

MAR(1) models & nonlinear systems If this process has a stationary distribution with mean = X ∞, then via a first-order Taylor expansion: Rearrange some terms to get: B u wtwt grouping & substituting yields our MAR(1) model!

MAR(1) models & nonlinear systems This defines explicitly the elements in B in terms of interaction strengths between species Change in log popn. growth rate of species i w.r.t. change in log abundance of species j

MAR(1) models & nonlinear systems Note:If the time step is sufficiently small, a linear model may suffice NtNt Time NtNt NonlinearLinear approx.

MAR(1) transition distribution Defn:conditional probability distribution for X t at any time t given initial condition X = x 0 Mean Variance Because all lie within unit circle, as t  ∞ and

Stationary distribution for MAR(1) Mean vector: Covariance matrix: Note: cannot solve this for V ∞ Solution: use the Vec() operator (analogous to the univariate case)

Stability properties of MAR(1) models Our interest is in stable systems (i.e. all eigenvalues of B lie within unit circle) Stability can be measured in several ways, but here are three that we’ll use: Stability measureMore stable when… Variance variance of the stationary distribution is low relative to that for the process error Return rate rapid approach to the stationary distribution (i.e. high return rate) Reactivity fewer departures from the mean of the stationary distribution (i.e. low reactivity)

Stability properties of MAR(1) models Fig. 2A - Ives et al. (2003) More stableLess stable Variance comparison Generally interested in the time spent away from equilibrium More time  greater variance

Stability properties of MAR(1) models Variance comparison 1) Smaller  smaller   more stable 2) To standardize across various systems Smaller  smaller det( B) 2/p  more stable Fig. 3 - Ives et al. (2003)

Stability properties of MAR(1) models In deterministic models, equilibrium is a point or stable limit cycle In stochastic models, equilibrium is the stationary distribution Rate of return to the stochastic equilibrium is the rate at which the transition distribution converges to the stationary distribution from an initial, known value The more rapid the convergence, the more stable the system Return rates (ie, resilience) Fig. 2B - Ives et al. (2003) More stableLess stable

Fig. 3 - Ives et al. (2003) Stability properties of MAR(1) models Return rates Rate of return increases as | b | gets smaller Rate of return to mean governed by dominant eigenvalue of B  max( B ) Rate of return of covariance defined by dominant eigenvalue of B  B  max( B  B )

Stability properties of MAR(1) models Fig. 2C - Ives et al. (2003) Low reactivityHigh reactivity A “reactive” system moves away from a stable equilibrium following a perturbation, even though the system will eventually return to equilibrium. Reactivity High reactivity occurs when species interactions greatly amplify the environmental forcing

Fig. 3 - Ives et al. (2003) Stability properties of MAR(1) models Reactivity Squared Euclidian distance for X t-1 from  ∞ Squared Euclidian distance for X t from  ∞ Change in squared Euclidian distance Expectation of this difference “worst-case” scenario

Stability properties of MAR(1) models Another form of reactivity (Neubert et al. 2009) Given our MAR(1) model: Suppose we perturb it by  ; B  is response of x t to x t-1 Magnitude of response is: And the reactivity is:

MAR(1) parameter estimation Conditional least squares (CLS) The CLS parameter estimates are those values that minimize the squared difference between 1)the observed population abundances at time t, and 2)those predicted by the MAR(1) model, conditional on the population abundances at time t-1.

MAR(1) parameter estimation Conditional least squares (CLS) Letandthen ifthen Goal: findwhere ^ ^

MAR(1) parameter estimation Conditional least squares (CLS) There are a couple of advantages to solving this via CLS: 1)It’s VERY fast, and 2)It’s easy to zero out elements of B to “test” for their significance

MAR(1) parameter estimation Conditional least squares (CLS) There are a couple of advantages to solving this via CLS: 1)It’s VERY fast, and 2)It’s easy to zero out elements of B to “test” for their significance

MAR(1) parameter estimation Maximum likelihood estimation (MLE)* Goal: find where Note: minimizing L can be computationally intensive, so often use CLS (same estimates if B unconstrained) *This is how MARSS does it

MAR(1) model selection Idea is to balance model “fit” with the number of parameters Two common methods: 1)Akaike’s Information Criterion (AIC); what MARSS uses 2)Bayesian Information Criterion (BIC) Note: “Penalty” larger with BIC, so it selects for “simpler” models Common formIves et al. (2003)

Model diagnostics Once selected, a MAR(1) model should be scrutinized Useful diagnostics include: 1)Are residuals temporally autocorrelated (via ACF)? 2)Are mean or variance of the residuals correlated with any variates or covariates (via X-Y plots)? 3)Are residuals normally distributed (via normal probability plots)? 4)What proportion of the variance is explained by the model (via conditional R 2 )? Note: conditional R 2 measures the proportion of change in log density from time t-1 to t

Parameter confidence intervals (CI) Can obtain CI’s from model in 3 ways: 1)Innovations bootstrapping (how Ives does it) Parameter estimation produces residuals Randomly resample the residuals to produce a new X Use new X to again estimate parameters Repeat “many” times Estimate CI’s from distribution of parameter estimates

Parameter confidence intervals (CI) 2)Parametric bootstrapping Generate parametric bootstrap data from the MLE model Use new data to again estimate parameters (bootstrap parameter estimates) Repeat “many” times

Parameter confidence intervals (CI) 3) Estimate CI’s from the (estimated) distribution of parameter estimates Estimate the Hessian matrix (there are various R functions to do this) Use that to compute the (estimated) CIs Large-n approximation CIs on variances might include negative numbers

MAR(1) with observation error This is simply our (now) familiar MARSS model:

Topics for the computer lab Fit community interaction model to freshwater plankton data from Ives et al. (2003) Add covariates to the above model Add an additional covariate (fish) observed with error Compute stability metrics