A toolbox for continuous time series analysis.. Why continuous time? a) If you can handle continuous time, you can handle discrete observations also,

Slides:



Advertisements
Similar presentations
UNLOCKING THE SECRETS HIDDEN IN YOUR DATA Part 3 Data Analysis.
Advertisements

From
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Layered continuous time processes in biology Combining causal statistical time series with fossil measurements. Tore Schweder and Trond Reitan CEES, University.
STAT 497 APPLIED TIME SERIES ANALYSIS
Modelling phenotypic evolution using layered stochastic differential equations (with applications for Coccolith data) How to model layers of continuous.
Bayesian statistics – MCMC techniques
Making rating curves - the Bayesian approach. Rating curves – what is wanted? A best estimate of the relationship between stage and discharge at a given.
EE322 Digital Communications
2-1 Sample Spaces and Events Conducting an experiment, in day-to-day repetitions of the measurement the results can differ slightly because of small.
Business Statistics - QBM117 Interval estimation for the slope and y-intercept Hypothesis tests for regression.
Review of Probability and Random Processes
Lecture II-2: Probability Review
1 10. Joint Moments and Joint Characteristic Functions Following section 6, in this section we shall introduce various parameters to compactly represent.
Elec471 Embedded Computer Systems Chapter 4, Probability and Statistics By Prof. Tim Johnson, PE Wentworth Institute of Technology Boston, MA Theory and.
Review of Probability.
Computational Geophysics and Data Analysis
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Random Process The concept of random variable was defined previously as mapping from the Sample Space S to the real line as shown below.
STATISTICS: BASICS Aswath Damodaran 1. 2 The role of statistics Aswath Damodaran 2  When you are given lots of data, and especially when that data is.
Modelling Phenotypic Evolution by Stochastic Differential Equations Combining statistical timeseries with fossil measurements. Tore Schweder and Trond.
© 2002 Prentice-Hall, Inc.Chap 14-1 Introduction to Multiple Regression Model.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
The Kalman Filter ECE 7251: Spring 2004 Lecture 17 2/16/04
Yaomin Jin Design of Experiments Morris Method.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Statistical modelling and latent variables. Constructing models based on insight and motivation.
Data analyses 2008 Lecture Last Lecture Basic statistics Testing Linear regression parameters Skill.
HY436: Mobile Computing and Wireless Networks Data sanitization Tutorial: November 7, 2005 Elias Raftopoulos Ploumidis Manolis Prof. Maria Papadopouli.
K. Ensor, STAT Spring 2005 Model selection/diagnostics Akaike’s Information Criterion (AIC) –A measure of fit plus a penalty term for the number.
2. Stationary Processes and Models
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Modern Navigation Thomas Herring MW 11:00-12:30 Room
Week 21 Stochastic Process - Introduction Stochastic processes are processes that proceed randomly in time. Rather than consider fixed random variables.
ارتباطات داده (883-40) فرآیندهای تصادفی نیمسال دوّم افشین همّت یار دانشکده مهندسی کامپیوتر 1.
Chapter 15 – CTRW Continuous Time Random Walks. Random Walks So far we have been looking at random walks with the following Langevin equation  is a.
K. Ensor, STAT Spring 2004 Memory characterization of a process How would the ACF behave for a process with no memory? What is a short memory series?
Principles of the Global Positioning System Lecture 12 Prof. Thomas Herring Room ;
Analyzing wireless sensor network data under suppression and failure in transmission Alan E. Gelfand Institute of Statistics and Decision Sciences Duke.
Data Analysis.
Estimation Method of Moments (MM) Methods of Moment estimation is a general method where equations for estimating parameters are found by equating population.
Short Introduction to Particle Filtering by Arthur Pece [ follows my Introduction to Kalman filtering ]
Lab for Remote Sensing Hydrology and Spatial Modeling Dept of Bioenvironmental Systems Engineering National Taiwan University 1/45 GEOSTATISTICS INTRODUCTION.
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
Joint Moments and Joint Characteristic Functions.
ECE-7000: Nonlinear Dynamical Systems 2. Linear tools and general considerations 2.1 Stationarity and sampling - In principle, the more a scientific measurement.
CHAPTER 2.3 PROBABILITY DISTRIBUTIONS. 2.3 GAUSSIAN OR NORMAL ERROR DISTRIBUTION  The Gaussian distribution is an approximation to the binomial distribution.
1 Review of Probability and Random Processes. 2 Importance of Random Processes Random variables and processes talk about quantities and signals which.
Introduction to stochastic processes
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
Chapter 2. Signals and Linear Systems
EC 827 Module 2 Forecasting a Single Variable from its own History.
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
Stochastic Process - Introduction
Covariance, stationarity & some useful operators
Diffusion over potential barriers with colored noise
Deep Feedforward Networks
Applied Discrete Mathematics Week 11: Relations
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Department of Civil and Environmental Engineering
Random Process The concept of random variable was defined previously as mapping from the Sample Space S to the real line as shown below.
EE513 Audio Signals and Systems
The Spectral Representation of Stationary Time Series
Lecturer Dr. Veronika Alhanaqtah
WHY WE FILTER TO IDENTIFY THE RELATIONSHIP
CH2 Time series.
Lecture 16. Classification (II): Practical Considerations
Nome Sobrenome. Time time time time time time..
Presentation transcript:

A toolbox for continuous time series analysis.

Why continuous time? a) If you can handle continuous time, you can handle discrete observations also, particularly if they are equidistant in time (but not restricted to that). b) Time is continuous... c) The program I made (layer_analyzer_complex) is a continuous time analysis toolbox. A process is a stochastic function of time: X(t)  t. Typically a process has a starting point in time, which is typically set to t=0. Stationary processes can however extend infinitely back in time.

Continuous time series vs discrete measurements While our state may be continuous in time, we’ll only have a finite set of measurements. These are often (but not always) made equidistantly in time. Continuous time linear processes sampled equidistanly in time results in discrete time linear processes. Observations are deemed as flawed and can go missing. Likelihood extracted using the Kalman filter.

Single state linear processes (Ornstein-Uhlenbeck) A single state linear process represents a process not affected by anything except noise. It can (and often will) have a dampening factor that makes the process approach a stationary expectancy and stay near that expectancy in a stationary fashion. If so, it is characterized by three parameters: 1.  - the expectation value 2.  - the size of the noise contribution 3.  t - the characteristic time. The characteristic time,  t, is the most important parameter here for describing the dynamics. It tells a) How fast the process approaches the expectancy. The distance will be 1/e of it’s original value after a time  t. b) How fast the process forgets previous states, i.e. how fast the auto-correlation is dampened.  (t)=e -t/  t. The time it takes for the auto-correlation to drop with a factor of 1/e is again  t. If it is more comfortable to look at halving times, that can be easily derived: X(0)=10,  =1,  t =1 Green line=expectancy (and solution without noise) X(0)=1, same process Green line=stationary expectancy, red line=stationary 95% credibility interval.

Ornstein-Uhlenbeck – Stochastic differential equation representation Mathematically, the OU process is represented by a linear stochastic differential equations: where a=1/  t. The first term represents an ordinary differential equation that pulls X(t) towards , reducing the distance by a factor 1/e for each time interval  t. The second term represents stochastic contributions, dB t, having a noise size . The size of the noise contribution is a bit complicated, in that it tells how much a process will increase in variance without the dampening. With a dampening it can be related to the stationary standard deviation as: X(0)=1,  =1,  t =1, sd(X(t))=0.5 Green line=stationary expectancy, red line=stationary 95% credibility interval.

Single measurement time series but multi state processes Even without having multiple time series, multiple states can be of value. Examples: a) The characteristics of an organism in deep time depend both on previous characteristics but also on changes in the optimal characteristics (environment). b) Position of a particle (depending both on previous position and velocity). c) Discharge (depending on previous discharge and area precipitation). The top layer process which is associated with the observations is said to track the underlying state. Each layer comes with a set of dynamical parameters,  t and . Multiple state processes give rise to different correlation on the observable top process structures than single state. So such hidden layers are detectable! SDE: Effects are a bit delayed in time (a factor  t 1 ) and the tracking can have a certain dampening effect.

Multiple time series Multiple time series can be handled simply by expanding the process state space. A zero-hypothesis can be made by letting the time series behave separately.

Correlated time series Correlated time series are time series where the stochastic contributions for process 1 and 2 are correlated. On average there will be no time delay between process 1 and 2. They represents processes affected by the same randomness (a third and fourth process containing white noise uncorrelated in time but correlated between each other).

Causal (regressed) time series One observed time series can respond to another. Ex: discharge vs precipitation, sedimentation vs temperature(?).

Correlated time series – the advanced version Time series can be correlated because they respond to some underlying process. This process doesn’t have to be white noise.

Input formats to ’layer_analyzer_complex’ The program takes a couple of text input formats: Units of years Units of seconds / / / / / / / / / / / / / Can be made using DAGUT/FINUT in the start system. Some use of ’emacs’ afterwards can be necessary to remove header info and other stuff the program doesn’t understand :00:00; :00:00; :00:00; :00:00; :00:00; :00:00; :00:00; :00:00; :00:00; :00:00;1.59 Units of seconds

Prior file The prior distribution is specified in a csv file by a couple of 95% credibility intervals: is_log;mu1;mu2;dt1;dt2;s1;s2;lin1;lin2;beta1;beta2;init1;init2;obs1;obs2 1;0.01;100;0.01;100.0;0.001;10;-0.01;0.01;-10;10;0.01;100;0.001;1.0 Log indicator. 0=no, 1=yes, 2=yes but the transformation has already been done Lower and upper limits for 95% credibility interval for . Same for  t, ,  t, , x 0,  Linear time dependency Initial state Observational noise

Standard one series usage: ~trr/prog/layer_analyzer_complex -i -l 1 Q oksfjord_year.txt vf_prior.txt Output: In addition, plots will appear showing the MCMC samples w= p= e+101 lik= e+66 prior= e+35 w*p= e+106 probsum=7.2119e w= p=1.6414e+104 lik= e+66 prior= e+37 w*p= e+106 probsum= e+107 probsum_0= e+104 lprobsum= probsum= e+60 -DIC/2 = ( DIC: , mean_D= , p_eff= , D_mean= ) Computing time=1.65s Total time: a T0= T1= T2= T3= T4=0 T5= T6= exp(mu_Q): mean= median= % cred=( ) dt_Q_1: mean= median= % cred=( ) sigma_Q_1: mean= median= % cred=( ) obs_sd_Q_origscale: mean= median= % cred=( ) exp(mu_Q) - spacing between independent samples: Specifies a new input file with associated model structure #layers other model options go here Axis label Name of file Prior file #mcmc samples burnin spacing #temp- ering Parameter inference Model likelihood: log(f(D|M))

Model testing, silent mode Test 1 layer vs 2 layer: ~trr/prog/layer_analyzer_complex –s -i -l 1 Q oksfjord_year.txt vf_prior.txt ~trr/prog/layer_analyzer_complex –s -i -l 2 Q oksfjord_year.txt vf_prior.txt Output 1: Output 2: lprobsum= exp(mu_Q): mean= median= % cred=( ) dt_Q_1: mean= median= % cred=( ) sigma_Q_1: mean= median= % cred=( ) dt_Q_2: mean= median= % cred=( ) sigma_Q_2: mean= median= % cred=( ) obs_sd_Q_origscale: mean= median= % cred=( ) exp(mu_Q) - spacing between independent samples: dt_Q_1 - spacing between independent samples: sigma_Q_1 - spacing between independent samples: dt_Q_2 - spacing between independent samples: sigma_Q_2 - spacing between independent samples: obs_sd_Q_origscale - spacing between independent samples: Silent mode. Only show model likelihood and parameter inference lprobsum= probsum= e+60 -DIC/2 = ( DIC: , mean_D= , p_eff= , D_mean= ) Computing time=1.65s Total time: a T0= T1= T2= T3= T4=0 T5= T6= exp(mu_Q): mean= median= % cred=( ) dt_Q_1: mean= median= % cred=( ) sigma_Q_1: mean= median= % cred=( ) obs_sd_Q_origscale: mean= median= % cred=( ) exp(mu_Q) - spacing between independent samples: dt_Q_1 - spacing between independent samples: sigma_Q_1 - spacing between independent samples: obs_sd_Q_origscale - spacing between independent samples: Conclusion: 1 layer slightly better than 2. #layers

Model testing, silent mode (2) OU vs random walk: ~trr/prog/layer_analyzer_complex –s -i -l 1 Q oksfjord_year.txt vf_prior.txt ~trr/prog/layer_analyzer_complex –s -i -l 1 -np Q oksfjord_year.txt vf_prior.txt Output 1: Output 2: lprobsum= probsum= e+61 -DIC/2 = ( DIC: , mean_D= , p_eff= , D_mean= ) Computing time=1.36s Total time: a T0= T1= T2= T3= T4=0 T5= T6= exp(mu_Q): mean= median= % cred=( nan) sigma_Q_1: mean= median= % cred=( ) obs_sd_Q_origscale: mean= median= % cred=( ) exp(mu_Q) - spacing between independent samples: sigma_Q_1 - spacing between independent samples: obs_sd_Q_origscale - spacing between independent samples: Silent mode. Only show model likelihood and parameter inference lprobsum= probsum= e+60 -DIC/2 = ( DIC: , mean_D= , p_eff= , D_mean= ) Computing time=1.65s Total time: a T0= T1= T2= T3= T4=0 T5= T6= exp(mu_Q): mean= median= % cred=( ) dt_Q_1: mean= median= % cred=( ) sigma_Q_1: mean= median= % cred=( ) obs_sd_Q_origscale: mean= median= % cred=( ) exp(mu_Q) - spacing between independent samples: dt_Q_1 - spacing between independent samples: sigma_Q_1 - spacing between independent samples: obs_sd_Q_origscale - spacing between independent samples: Conclusion: No pull better than OU. No pull

Tying several series together Example: sediment vs temperature # Only from /prog/layer_analyzer_complex -i -l 1 sediment NewCore3-sediment-after1950.txt sediment_prior.txt -i -l temp Year_average_temperature.txt temp_prior.txt lprobsum= /prog/layer_analyzer_complex -i -l 1 sediment NewCore3-sediment-after1950.txt sediment_prior.txt -i -l temp Year_average_temperature.txt temp_prior.txt -f lprobsum= /prog/layer_analyzer_complex -i -l 1 sediment NewCore3-sediment-after1950.txt sediment_prior.txt -i -l temp Year_average_temperature.txt temp_prior.txt -C lprobsum= Conclusion: temperature -> sediment. Bayes factor of exp(4)