Facultad de Ingeniería, Centro de Cálculo

Slides:



Advertisements
Similar presentations
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Advertisements

Kriging.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Chapter 10 Curve Fitting and Regression Analysis
Ch11 Curve Fitting Dr. Deshi Ye
A Short Introduction to Curve Fitting and Regression by Brad Morantz
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Yard. Doç. Dr. Tarkan Erdik Regression analysis - Week 12 1.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Regression Analysis. Unscheduled Maintenance Issue: l 36 flight squadrons l Each experiences unscheduled maintenance actions (UMAs) l UMAs costs $1000.
Spatial Analysis Longley et al., Ch 14,15. Transformations Buffering (Point, Line, Area) Point-in-polygon Polygon Overlay Spatial Interpolation –Theissen.
Applied Geostatistics
November 2, 2010Neural Networks Lecture 14: Radial Basis Functions 1 Cascade Correlation Weights to each new hidden node are trained to maximize the covariance.
Ordinary Kriging Process in ArcGIS
Classification and Prediction: Regression Analysis
Radial-Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Radial Basis Function Networks
Spatial Interpolation of monthly precipitation by Kriging method
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Image Preprocessing: Geometric Correction Image Preprocessing: Geometric Correction Jensen, 2003 John R. Jensen Department of Geography University of South.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Kalman Filter (Thu) Joon Shik Kim Computational Models of Intelligence.
Geographic Information Science
Quality control of daily data on example of Central European series of air temperature, relative humidity and precipitation P. Štěpánek (1), P. Zahradníček.
Chapter 13 Multiple Regression
Presented By, Shivvasangari Subramani. 1. Introduction 2. Problem Definition 3. Intuition 4. Experiments 5. Real Time Implementation 6. Future Plans 7.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Lecture 6: Point Interpolation
Predictive Application- Performance Modeling in a Computational Grid Environment (HPDC ‘99) Nirav Kapadia, José Fortes, Carla Brodley ECE, Purdue Presented.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
Computacion Inteligente Least-Square Methods for System Identification.
CPH Dr. Charnigo Chap. 11 Notes Figure 11.2 provides a diagram which shows, at a glance, what a neural network does. Inputs X 1, X 2,.., X P are.
Model Selection and the Bias–Variance Tradeoff All models described have a smoothing or complexity parameter that has to be considered: multiplier of the.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Kriging - Introduction Method invented in the 1950s by South African geologist Daniel Krige (1919-) for predicting distribution of minerals. Became very.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Chapter 4 Basic Estimation Techniques
Chapter 7. Classification and Prediction
Using satellite data and data fusion techniques
Linear Regression CSC 600: Data Mining Class 12.
Deep Feedforward Networks
Neural Networks Winter-Spring 2014
How Good is a Model? How much information does AIC give us?
Chapter 11: Simple Linear Regression
Investigating the role of individual neurons as outlier detectors
CSE 4705 Artificial Intelligence
CH 5: Multivariate Methods
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Structure learning with deep autoencoders
Roberto Battiti, Mauro Brunato
Dr. Unnikrishnan P.C. Professor, EEE
HYDROLOGICAL TIME SERIES GENERATION IN INDIAN CONTEXT
Spatial Analysis Longley et al..
Statistical Methods For Engineers
Unfolding Problem: A Machine Learning Approach
Neural Networks Advantages Criticism
Filtering and State Estimation: Basic Concepts
An Introduction To The Backpropagation Algorithm
Identification of Wiener models using support vector regression
Contact: Machine Learning – (Linear) Regression Wilson Mckerrow (Fenyo lab postdoc) Contact:
Unfolding with system identification
Lecture 16. Classification (II): Practical Considerations
Presentation transcript:

Facultad de Ingeniería, Centro de Cálculo Application of ANN to the prediction of missing daily precipitation records, and comparison against linear methodologies Carlos López-Vázquez Facultad de Ingeniería, Centro de Cálculo Montevideo URUGUAY

Organization of the presentation The missing value problem The experiment Data and test area description Methods tested Results Conclusions

The missing value problem Weather records typically have missing values Many applications require complete databases Well established linear methods for interpolate spatial observations exist Their performance is poor for daily rain records

The experiment We created missing values at random locations in our dataset We imputated them using all test methods We computed RMS, mean and some percentiles of the absolute deviation between the original and the imputated value We performed a number of simulations, and averaged the results

Data and test area description 20 years of daily records for 10 stations were available 30 % of the events have missing values More than 80% of the readings are of zero rain, evenly distributed in the year Annual averages ranges from 1600 to 500 mm/day; time correlation is low

Methods tested: linear All follow the general equation We restricted ourselves to weights w and bias b constant in time; its values depends on the method itself Some commonly used methods will be briefly described “Best among linear” methods can be obtained depending on the error measure used

Linear methods: Cressman being dij stands for the distance between station i and j It does not use historical information Only requires the station coordinates

Linear methods: Nearest neighbor Uses only data from the closest available station “closest” might be geometrical, or might be determined by an expert; we tested both Due to its simplicity it is widely used

Linear methods: Optimum interpolation It is the standard interpolation method in meteorology It is also known as kriging in geostatistics Requires the evaluation of the covariance matrix of the sample, and a model of the spatial correlation. It is designed in order to minimize the expected RMSE

Linear methods: “The Best ones” Given an objective measure of the error, it can be minimized by tuning the parameters w Least sum of squares, Least average of the absolute error, Least 95 percentile, etc. are examples. The optimum might led to different vectors w in each case. Calculating the parameters w might be a heavy nonlinear optimization problem

Non-linear methods: ANN We used ANN as interpolators, with 9 inputs and 1 output The training was performed with one third of the dataset using backpropagation and minimizing the RMSE Some different architectures were considered (one and two hidden layers; different number of neurons, etc.) as well as some transformations of the data

Results The results are poor, irrespective of the method; daily rain is a very difficult problem Both ANN gave only slightly better results than linear methods, but were more robust The training algorithm failed to escape from local minima; that leds to a significant CPU cost

Conclusions Linear methods with constant coefficients have limited possibilities Adding information of the day before improves to some extent the results ANN gave interesting results, but with too much CPU requirements Better training algorithms are required