Predicting Salinity in the Chesapeake Bay Using Neural Networks

Slides:



Advertisements
Similar presentations
Introduction to Training and Learning in Neural Networks n CS/PY 399 Lab Presentation # 4 n February 1, 2001 n Mount Union College.
Advertisements

CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Multiple Regression Analysis
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Artificial Neural Networks ML Paul Scheible.
October 28, 2010Neural Networks Lecture 13: Adaptive Networks 1 Adaptive Networks As you know, there is no equation that would tell you the ideal number.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Correlation & Regression
Chapter 13: Inference in Regression
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
STATISTICS: BASICS Aswath Damodaran 1. 2 The role of statistics Aswath Damodaran 2  When you are given lots of data, and especially when that data is.
A Shaft Sensorless Control for PMSM Using Direct Neural Network Adaptive Observer Authors: Guo Qingding Luo Ruifu Wang Limei IEEE IECON 22 nd International.
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Networks
Chapter 11 – Neural Networks COMP 540 4/17/2007 Derek Singer.
Appendix B: An Example of Back-propagation algorithm
You want to examine the linear dependency of the annual sales of produce stores on their size in square footage. Sample data for seven stores were obtained.
Neural Networks - Berrin Yanıkoğlu1 Applications and Examples From Mitchell Chp. 4.
Chapter 7 Neural Networks in Data Mining Automatic Model Building (Machine Learning) Artificial Intelligence.
Classification / Regression Neural Networks 2
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
1 GMDH and Neural Network Application for Modeling Vital Functions of Green Algae under Toxic Impact Oleksandra Bulgakova, Volodymyr Stepashko, Tetayna.
Multi-Layer Perceptron
Chapter 13 Multiple Regression
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
1 FORECASTING ENERGY PRODUCT PRICES M.E. Malliaris Loyola University Chicago S.G. Malliaris Massachusetts Institute of Technology.
VARIABILITY OF TOTAL ELECTRON CONTENT AT EUROPEAN LATITUDES A. Krankowski(1), L. W. Baran(1), W. Kosek (2), I. I. Shagimuratov(3), M. Kalarus (2) (1) Institute.
Speech Communication Lab, State University of New York at Binghamton Dimensionality Reduction Methods for HMM Phonetic Recognition Hongbing Hu, Stephen.
Scatter Diagrams scatter plot scatter diagram A scatter plot is a graph that may be used to represent the relationship between two variables. Also referred.
An Artificial Neural Network Approach to Surface Waviness Prediction in Surface Finishing Process by Chi Ngo ECE/ME 539 Class Project.
Chapter 10: Determining How Costs Behave 1 Horngren 13e.
Neural Networks Demystified by Louise Francis Francis Analytics and Actuarial Data Mining, Inc.
Multiple Regression Analysis Regression analysis with two or more independent variables. Leads to an improvement.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
BPA CSUB Prof. Yong Choi. Midwest Distribution 1. Create scatter plot Find out whether there is a linear relationship pattern or not Easy and simple using.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Bump Hunting The objective PRIM algorithm Beam search References: Feelders, A.J. (2002). Rule induction by bump hunting. In J. Meij (Ed.), Dealing with.
Introduction Many problems in Engineering, Management, Health Sciences and other Sciences involve exploring the relationships between two or more variables.
Intro. ANN & Fuzzy Systems Lecture 11. MLP (III): Back-Propagation.
Chapter 11 – Neural Nets © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
Chapter 11: Linear Regression E370, Spring From Simple Regression to Multiple Regression.
Neural Networks - Berrin Yanıkoğlu1 MLP & Backpropagation Issues.
Stats Methods at IC Lecture 3: Regression.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning Supervised Learning Classification and Regression
Multiple-Layer Networks and Backpropagation Algorithms
Fall 2004 Backpropagation CS478 - Machine Learning.
Regression Analysis Module 3.
Basic Estimation Techniques
Correlation and Regression
Classification / Regression Neural Networks 2
Basic Estimation Techniques
Regression Computer Print Out
Artificial Neural Network & Backpropagation Algorithm
Christophe Dubach, Timothy M. Jones and Michael F.P. O’Boyle
of the Artificial Neural Networks.
Chap. 7 Regularization for Deep Learning (7.8~7.12 )
STEM Fair Graphs.
network of simple neuron-like computing elements
Simple Linear Regression
Correlation and Regression
Neural Networks II Chen Gao Virginia Tech ECE-5424G / CS-5824
Neural networks (1) Traditional multi-layer perceptrons
Backpropagation David Kauchak CS159 – Fall 2019.
Neural Networks II Chen Gao Virginia Tech ECE-5424G / CS-5824
Learning Combinational Logic
Time Series Forecasting with Recurrent Neural Networks NN3 Competition Mahmoud Abou-Nasr Research & Advanced Engineering Ford Motor Company
Presentation transcript:

Predicting Salinity in the Chesapeake Bay Using Neural Networks by Bruce L. Golden Smith School of Business University of Maryland 10/93

Goal Construct multiple regression models and neural network models that accurately describe the dynamics of salinity in the Maryland portion of the Chesapeake Bay Other efforts use time series methods to predict surface and bottom salinity as part of a Bay water quality model

Source of Data Data collected by USEPA in five “regions” of the Chesapeake Bay Upper, middle, lower tributaries, and entire Bay 18 stations in the mainstem Bay 16 stations in tributaries

Source of Data -- continued Water samples collected at the bottom of the Bay (bottom data) and at various depths in the Bay (total data) Old data: 36,000 observations 1984-1989 New data: 7,000 observations 1989-1990

Source of Data -- continued Ten different regression models and ten different neural network models are built using the old data 5 regions x 2 depths Neural network models and regression models are compared using 20 data sets (old data and new data)

Regression Models Extensive screening phase for independent variables Four key independent variables ----------------------------------------------------------------- Day day of the year on which measurements were taken Depth depth at which measurements were taken Latitude latitude of sampling station Longitude longitude of sampling station

Regression Models -- continued Used stepwise regression in SPSS/PC Avoid highly correlated independent variables Keep models simple: don’t include variables that add little in predictive power

Regression Models Constructed 5 bottom-data models and 5 total-data models using old data Entire Bay model using 36,000 observations R2 = 0.649 Salinity = 199.839 – 1.151Day1 +1.161Day2 + 0.283Depth – 4.863Latitude – 1.543Longitude – 13.402Longitude1

Regression Models -- continued Six independent variables in each model All coefficients were significant Each model easily passed an F test No problems with multicollinearity R2 values ranged from 0.56 to 0.81

Neural Network Models Neural network configuration Salinity Level Hidden Nodes Station Depth Latitude Longitude Date Longitude x Depth Number

Neural Network Models --continued Neural network details Multilayer feedforward network Training by backpropagation Length of training session – 2000 iterations Training time on Sun 4/370 – 5 minutes Input value mapped to [-1, +1] Output (salinity) values mapped to [0, +1], same range as sigmoid function

Neural Network Models Neural Network parameters Bottom Data Region of the Bay ___________________________________________ Parameter Upper Middle Lower Tributaries Entire Learning rate .80 .60 .60 .20 .80 Momentum term .40 .70 .10 .10 .10 Hidden nodes .40 .30 .50 .30 .40 Slope .80 .80 .80 .80 .80 ________________________________________________________________

Neural Network Models -- continued Total Data Region of the Bay ___________________________________________ Parameter Upper Middle Lower Tributaries Entire Learning rate .20 .80 .80 .60 .20 Momentum term .10 .80 .40 .20 .10 Hidden nodes .30 .40 .40 .20 .30 Slope .80 .80 .80 .80 .80 ________________________________________________________________

Neural Network Models -- continued Training the neural network Region of the Bay ____________________________________________ Upper Middle Lower Tributaries Entire Bottom Data % in training set 20 20 20 20 10 # in training set 199 243 190 79 271 _________________________________________________________ Total Data % in training set 2 2 2 2 1 # in training set 250 330 280 78 363

Comparison of Models Regression models can use a different set of six independent variables in each region Neural network models are based on the same set of six variables in each region Computational results Range of Average Percent Absolute Errors 10 old data sets 10 new data sets Regression 9.60 – 16.46 9.19 – 20.15 Neural Network 9.54 – 16.18 7.70 – 19.37 _________________________________________________________

Comparison of Models -- continued Key points Neural network models have lower average PAE than the regression models in 18 out of 20 cases Worst errors of the neural network models are not as bad as those from regression Neural network models yield more errors in the 0-10% range than regression models

Conclusions Current combinations of training parameters work quite well for the neural network models Major advantage of the regression models is that they are easily explained Based on a small number of observations and six fixed variables, the neural network models predict salinity levels more accurately than do the regression models