Applications of Neural Networks in Time-Series Analysis Adam Maus Computer Science Department Mentor: Doctor Sprott Physics Department.

Slides:



Advertisements
Similar presentations
Time-Series Analysis J. C. (Clint) Sprott
Advertisements

Polynomial Curve Fitting BITS C464/BITS F464 Navneet Goyal Department of Computer Science, BITS-Pilani, Pilani Campus, India.
Deepayan ChakrabartiCIKM F4: Large Scale Automated Forecasting Using Fractals -Deepayan Chakrabarti -Christos Faloutsos.
From
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Strange Attractors From Art to Science
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Machine Learning Neural Networks
1 Introduction to Bio-Inspired Models During the last three decades, several efficient machine learning tools have been inspired in biology and nature:
Lecture 14 – Neural Networks
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Artificial Intelligence Statistical learning methods Chapter 20, AIMA (only ANNs & SVMs)
Neural Networks. R & G Chapter Feed-Forward Neural Networks otherwise known as The Multi-layer Perceptron or The Back-Propagation Neural Network.
LAB 3 AIRBAG DEPLOYMENT SENSOR PREDICTION NETWORK Warning This lab could save someone’s life!
Support Vector Regression David R. Musicant and O.L. Mangasarian International Symposium on Mathematical Programming Thursday, August 10, 2000
Artificial Neural Networks
Radial Basis Function (RBF) Networks
Radial Basis Function G.Anuradha.
Radial-Basis Function Networks
Radial Basis Function Networks
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
Traffic modeling and Prediction ----Linear Models
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Artificial Neural Networks
Chapter 9 Neural Network.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
1 FARIMA(p,d,q) Model and Application n FARIMA Models -- fractional autoregressive integrated moving average n Generating FARIMA Processes n Traffic Modeling.
Time Series Data Analysis - I Yaji Sripada. Dept. of Computing Science, University of Aberdeen2 In this lecture you learn What are Time Series? How to.
Strange Attractors From Art to Science J. C. Sprott Department of Physics University of Wisconsin - Madison Presented to the Society for chaos theory in.
Ch 9.8: Chaos and Strange Attractors: The Lorenz Equations
GEOSTATISICAL ANALYSIS Course: Special Topics in Remote Sensing & GIS Mirza Muhammad Waqar Contact: EXT:2257.
Intro. ANN & Fuzzy Systems Lecture 26 Modeling (1): Time Series Prediction.
The Science of Complexity J. C. Sprott Department of Physics University of Wisconsin - Madison Presented to the First National Conference on Complexity.
A Physicist’s Brain J. C. Sprott Department of Physics University of Wisconsin - Madison Presented at the Chaos and Complex Systems Seminar In Madison,
Is Global Warming for Real? J. C. Sprott Department of Physics University of Wisconsin - Madison Presented at the Chaos and Complex Systems Seminar In.
Low-Dimensional Chaotic Signal Characterization Using Approximate Entropy Soundararajan Ezekiel Matthew Lang Computer Science Department Indiana University.
Introduction to Chaos by: Saeed Heidary 29 Feb 2013.
CCN COMPLEX COMPUTING NETWORKS1 This research has been supported in part by European Commission FP6 IYTE-Wireless Project (Contract No: )
Akram Bitar and Larry Manevitz Department of Computer Science
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
Analyzing wireless sensor network data under suppression and failure in transmission Alan E. Gelfand Institute of Statistics and Decision Sciences Duke.
Simple Models of Complex Chaotic Systems J. C. Sprott Department of Physics University of Wisconsin - Madison Presented at the AAPT Topical Conference.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by:
Gaussian Process and Prediction. (C) 2001 SNU CSE Artificial Intelligence Lab (SCAI)2 Outline Gaussian Process and Bayesian Regression  Bayesian regression.
CHEE825 Fall 2005J. McLellan1 Nonlinear Empirical Models.
Perceptrons Michael J. Watts
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Lecture 2 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 2/1 Dr.-Ing. Erwin Sitompul President University
L’Aquila 1 of 26 “Chance or Chaos?” Climate 2005, PIK, Jan 2005 Gabriele Curci, University of L’Aquila
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Semi-Supervised Clustering
Deep Feedforward Networks
Artificial Neural Networks
Deterministic Dynamics
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Nonlinear Structure in Regression Residuals
Lecture 26 Modeling (1): Time Series Prediction
Random walk initialization for training very deep feedforward networks
Dynamics of High-Dimensional Systems
Strange Attractors From Art to Science
Artificial Intelligence Methods
Using Artificial Neural Networks and Support Vector Regression to Model the Lyapunov Exponent Adam Maus.
CH2 Time series.
Akram Bitar and Larry Manevitz Department of Computer Science
Presentation transcript:

Applications of Neural Networks in Time-Series Analysis Adam Maus Computer Science Department Mentor: Doctor Sprott Physics Department

Outline Discrete Time Series Embedding Problem Methods Traditionally Used Lag Space for a System Overview of Artificial Neural Networks –Calculation of time lag sensitivities –Comparison of model to known systems Comparison to other methods Results and Discussions

Discrete Time Series Scalar chaotic time series –i.e. average daily temperatures Data given discrete intervals –i.e. seconds, days, etc. We assume a combination of past values in the time series predict the future [45, 34, 23, 56, 34, 25, …] Discrete Time Series (Wikimedia)

Time Lags Each discrete interval refers to a dimension or time lag [45, 34, 23, 56, 34, 25, …] Current Value 1 st Time Lag 2 nd Time Lag Current Value 2 nd Time Lag ….

Outline Discrete Time Series Embedding Problem Methods Traditionally Used Lag Space for a System Overview of Artificial Neural Networks –Calculation of time lag sensitivities –Comparison of model to known systems Comparison to other methods Results and Discussions

Embedding Problem The embedding dimension is related to the minimum number of variables required to construct the data Or Exactly how many time lags are required to reconstruct the system without any information being lost but without adding unnecessary information –i.e. a ball seen in 2d, 3d, and 3d + Time Problem: How do we choose the optimal embedding dimension that the model can use to unfold the data? X k

Outline Discrete Time Series Embedding Problem Methods Traditionally Used Lag Space for a System Overview of Artificial Neural Networks –Calculation of time lag sensitivities –Comparison of model to known systems Comparison to other methods Results and Discussions

ARMA Models Autoregressive Moving Average Models –Fits a polynomial to the data based on linear combinations of past values –Produces a linear function –Can create very complicated dynamics but has difficulty with nonlinear systems

Autocorrelation Function Finds correlations within data Much like the ARMA model, shows weak periodicity within nonlinear time series. No sense of the underlying dynamical system Logistic Map TheThe Nature of Mathematical Modeling (1999)

Correlation Dimension Introduced in 1983 by Grassberger and Procaccia to find the fractal dimension of a chaotic system One can determine the embedding dimension by calculating the correlation dimension in increasing dimensions until it ceases to change Good for large datasets with little noise Measuring the Strangeness of Strange Attractors (1983)

False Nearest Neighbors Introduced in 1992 by Kennel, Brown, and Abarbanel Calculation of false nearest neighbors in successively higher embedding dimensions As d is increased, the fraction of neighbors that are false drops to near zero Good for smaller datasets and rather robust to noise 1992 Paper on False Nearest Neighbors % of False Neighbors Dimension Goutte Map

Outline Discrete Time Series Embedding Problem Methods Traditionally Used Lag Space for a System Overview of Artificial Neural Networks –Calculation of time lag sensitivities –Comparison of model to known systems Comparison to other methods Results and Discussions

Lag Space Not necessarily the same dimensions as embedding space Goutte Map dynamics depend only on the second and fourth time lag Lag Space Estimation In Time Series Modelling (1997) Goutte Map Problem: How can we measure both the embedding dimension and lag space?

Outline Discrete Time Series Embedding Problem Methods Traditionally Used Lag Space for a System Overview of Artificial Neural Networks –Calculation of time lag sensitivities –Comparison of model to known systems Comparison to other methods Results and Discussions

Artificial Neural Networks Mathematical Models of Biological Neurons Used in Classification Problems –Handwriting Analysis Function Approximation –Forecasting Time Series –Studying Properties of Systems

Function Approximation Known as Universal Approximators The architecture of the neural network uses time-delayed data Structure of a Single-Layer Feed forward Neural Network x = [45, 34, 23, 56, 34, 25, …] X k-d n 1 Multilayer Feedforward Networks are Universal Approximators (1989)

Function Approximation Next Step Prediction –Takes d previous points and predicts the next step

Training 1.Initialize a matrix and b vector 2.Compare predictions to actual values 3.Change parameters accordingly 4.Repeat millions of times Mean Square Error c = length of time series Fitting the model to data (Wikimedia)

Convergence The global optimum is found at the lowest mean square error –Connection strengths can be any real number –Like finding the lowest point in a mountain range Numerous low points so we must devise ways to avoid these local optimum

Outline Discrete Time Series Embedding Problem Methods Traditionally Used Lag Space for a System Overview of Artificial Neural Networks –Calculation of time lag sensitivities –Comparison of model to known systems Comparison to other methods Results and Discussions

Time Lag Sensitivities We can train a neural network on data and study the model Find how much the output of the neural network varies when perturbing each time lag “Important” lags will have higher sensitivity to changes in values

Time Lag Sensitivities We estimate the sensitivity of each time lag in the neural network:

Expected Sensitivities For known systems we can estimate what the sensitivities should be After training neural networks on data from different maps the difference between actual and expected sensitivities is <1%

Outline Discrete Time Series Embedding Problem Methods Traditionally Used Lag Space for a System Overview of Artificial Neural Networks –Calculation of time lag sensitivities –Comparison of model to known systems Comparison to other methods Results and Discussions

Hénon Map Strange Attractor of Hénon Map A two-dimensional mapping with a strange attractor (1976) Embedding of 2 Sensitivities S(1) = S(2) = 0.3

Delayed Hénon Map Strange Attractor of Delayed Hénon Map High-dimensional Dynamics in the Delayed Hénon Map (2006) Embedding of d Sensitivities S(1) = S(4) =.1

Preface Map “The Volatile Wife” Embedding of 3 Strange Attractor of Preface Map Sensitivities S(1) = S(2) = 0.9 S(3) = 0.6 Chaos and Time-Series Analysis (2003) Images of a Complex World: The Art and Poetry of Chaos (2005)

Goutte Map Strange Attractor of Goutte Map Embedding of 4 Sensitivities Lag Space Estimation In Time Series Modelling (1997) S(2) = S(4) = 0.3

Outline Discrete Time Series Embedding Problem Methods Traditionally Used Lag Space for a System Overview of Artificial Neural Networks –Calculation of time lag sensitivities –Comparison of model to known systems Comparison to other methods Results and Discussions

Results from Other Methods Hénon Map Neural Network False Nearest Neighbors Correlation Dimension Optimal Embedding of 2

Results from Other Methods Delayed Hénon Map Neural Network False Nearest Neighbors Correlation Dimension Optimal Embedding of 4

Results from Other Methods Preface Map Neural Network False Nearest Neighbors Correlation Dimension Optimal Embedding of 3

Results from Other Methods Goutte Map Neural Network False Nearest Neighbors Correlation Dimension Optimal Embedding of 4

Comparison using Data Set Size Varied the length of the Hénon map time series by powers of 2 Compared methods to actual values using normalized RMS error E % 70 0 Where is predicted value for a test data set size is actual value for an ideal data set size j is one dimension of d that we are studying

Comparison using Noisy Data Sets Vary the noise in the system by adding Gaussian White Noise to a fixed length time series from the Hénon Map Compared methods to actual values using normalized RMS error Used noiseless case values for comparison of methods High NoiseLow Noise %

Outline Discrete Time Series Embedding Problem Methods Traditionally Used Lag Space for a System Overview of Artificial Neural Networks –Calculation of time lag sensitivities –Comparison of model to known systems Comparison to other methods Results and Discussions

Temperature in Madison

Precipitation in Madison

Summary Neural networks are models that can be used to predict the embedding dimension They can handle small datasets and accurately predict sensitivities for a given system They prove to be more robust to noise than other methods used They can be used to determine the lag space where methods cannot

Acknowledgments Doctor Sprott for guiding this project Doctor Young and Ed Hopkins at the State Climatology Office for the weather data and insightful discussions