Identifying Severe Weather Radar Characteristics

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Part I Introduction to Data Mining by Tan,
Molecular Biomedical Informatics 分子生醫資訊實驗室 Machine Learning and Bioinformatics 機器學習與生物資訊學 Machine Learning & Bioinformatics 1.
Comparison of Data Mining Algorithms on Bioinformatics Dataset Melissa K. Carroll Advisor: Sung-Hyuk Cha March 4, 2003.
Application of Stacked Generalization to a Protein Localization Prediction Task Melissa K. Carroll, M.S. and Sung-Hyuk Cha, Ph.D. Pace University, School.
Machine Learning Neural Networks
Decision Tree Rong Jin. Determine Milage Per Gallon.
Ensemble Learning: An Introduction
Three kinds of learning
Neural Networks Chapter Feed-Forward Neural Networks.
Data Mining: Discovering Information From Bio-Data Present by: Hongli Li & Nianya Liu University of Massachusetts Lowell.
Modeling Gene Interactions in Disease CS 686 Bioinformatics.
Lecture 4 Neural Networks ICS 273A UC Irvine Instructor: Max Welling Read chapter 4.
Rating Systems Vs Machine Learning on the context of sports George Kyriakides, Kyriacos Talattinis, George Stefanides Department of Applied Informatics,
Data mining and statistical learning - lecture 12 Neural networks (NN) and Multivariate Adaptive Regression Splines (MARS)  Different types of neural.
CS Instance Based Learning1 Instance Based Learning.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Ranga Rodrigo April 5, 2014 Most of the sides are from the Matlab tutorial. 1.
Predicting Income from Census Data using Multiple Classifiers Presented By: Arghya Kusum Das Arnab Ganguly Manohar Karki Saikat Basu Subhajit Sidhanta.
Artificial Neural Networks
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Some working definitions…. ‘Data Mining’ and ‘Knowledge Discovery in Databases’ (KDD) are used interchangeably Data mining = –the discovery of interesting,
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
M Machine Learning F# and Accord.net. Alena Dzenisenka Software architect at Luxoft Poland Member of F# Software Foundation Board of Trustees Researcher.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Jennifer Lewis Priestley Presentation of “Assessment of Evaluation Methods for Prediction and Classification of Consumer Risk in the Credit Industry” co-authored.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
An Artificial Neural Network Approach to Surface Waviness Prediction in Surface Finishing Process by Chi Ngo ECE/ME 539 Class Project.
In part from: Yizhou Sun 2008 An Introduction to WEKA Explorer.
Linear Models & Clustering Presented by Kwak, Nam-ju 1.
Chapter 11 – Neural Nets © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Ensemble Classifiers.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning: Ensemble Methods
Big data classification using neural network
Fall 2004 Backpropagation CS478 - Machine Learning.
Data Transformation: Normalization
Semi-Supervised Clustering
Deep Feedforward Networks
Artificial Neural Networks
An Empirical Comparison of Supervised Learning Algorithms
Other Classification Models: Neural Network
Learning with Perceptrons and Neural Networks
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Understanding Neural Networks using .NET
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Forecasting The Future of Movies
Data Mining Lecture 11.
ANN Design and Training
Neural Networks Advantages Criticism
CS539: Project 3 Zach Pardos.
network of simple neuron-like computing elements
CS 478 Homework CS Homework.
CSCI N317 Computation for Scientific Applications Unit Weka
Neural Networks ICS 273A UC Irvine Instructor: Max Welling
Machine Learning in Practice Lecture 23
Ensemble learning.
COSC 4335: Part2: Other Classification Techniques
Ensembles An ensemble is a set of classifiers whose combined results give the final decision. test feature vector classifier 1 classifier 2 classifier.
A task of induction to find patterns
A task of induction to find patterns
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Modeling IDS using hybrid intelligent systems
Presentation transcript:

Identifying Severe Weather Radar Characteristics Ron Holmes NWS State College PA

Strategy Identify differences between original training set and original test set Prepare training and test files from original training data set with known classifications. WEKA Histograms Identify Attributes Decision Tree Neural Net Develop and train a Neural Net for real-time Operations

Methods Identify differences between original training set and original test set to: Eliminate potential outliers in training set. Eliminate suspect or bad data due to measurement errors or misclassification in training set. Data Changes: Eliminated Lat and Lon of Centroid for all instances Changed all missing data to 0 Determine Outliers: Computed Min, Max, Mean, STD for each predictor in the original training and test files. Eliminated 4 instances in training file where a few predictors were over 2 STD’s from the mean of the test file for that predictor. Contest based on highest TSS of test set Objective to bring Mean and STD for each predictor in training file as close as possible to those in the test file. Without original radar data it would be difficult to try to correct for misclassifications in the training set Train system for highest test results (….not necessarily reality )

Methods Created separate Training and Test files from original data file: Used 70% for training and 30% for testing. Randomly stratified Due to uneven distribution of predictands kept the same ratio for both training and test sets. Storm Type Total number in original file Number in training file based on 70% of original file Number in testing file based on 30% of original file Non-Svr 526 368 158 Isold 222 156 66 Line 208 146 62 Pulse 400 280 120 Resulting skill scores reflected uneven distribution (as with original study) Higher skill in predicting Non-Svr and Pulse Lowest skill in predicting Lines

Methods Weka workbench (http://www.cs.waikato.ac.nz/ml/weka/) Variety of tools for attribute selection, decision trees, logistic regression, and neural network. Attribute selection Histogram Analysis Various algorithms to rank important predictors Chi Squared, Genetic Search, Gain Ratio, Support Vector Machine Goal: To reduce number of predictors for faster training. Performed sensitivity experiments with decision tree/neural net to identify which predictors maximize skill scores. Main predictors (11 out of original 20): Max Ref, Mean Ref, Max VIL, Aspect Ratio, MESH, LatRadius, LonRadius, LifePOSH, Shear, Rot120, Motion South

Good Predictors Good Predictors Good Predictors Good Predictors Non Pulse Isold Line

943 Correctly Classified : 69.5% 413 Incorrectly Classified : 30.5% Majority of Pulse Storms Majority of Line Storms identified by the shape High Buoyancy, High Shear Environment Suspected Miss-classification due to Low MaxVil High Buoyancy, Weak Shear Environment. Slow Moving Pulse Storms Majority of Non-Svr Storms

Non-Linear Relationships by Storm Type Max Reflectivity (X axis) Max VIL (Y axis) MESH (X axis) Max VIL (Yaxis) Low Lvl Shear (X axis) Mean Ref (Y axis)

Neural Network Standard Feed-Forward/Backpropagation ANN 2 Layers – adjustable number of nodes Logistic Sigmoid activation function – winner takes all method Adjustable: Iterations Learning rate Momentum Weight Decay Split original training file: 70% Training and 30% for Testing Instances randomly stratified and proportioned in training and test files according to number of instances of each predictand. Same method as was used for Decision Tree preparation Inputs to ANN were Normalized according to: Normalized Value = (Value-min)/(max-min) Where min and max are the minimum and maximum values of each predictor.

ANN Results using original file split into training and test sets Training TSS: 0.67 POD FAR CSI Non Svr 0.86 0.14 0.75 Isold 0.28 0.57 Line 0.56 0.25 0.46 Pulse 0.31 Testing TSS: 0.70 POD FAR CSI Non Svr 0.88 0.12 0.78 Isold 0.71 0.27 0.55 Line 0.51 0.25 0.43 Pulse 0.80 0.29 0.60 Best Network: Iterations: 600 Hidden Nodes Layer1: 20 Hidden Nodes Layer2: 10 Learn: 0.08 Momentum: 0.3 Training Confusion Matrix Forecast Class: 0 1 2 4 ------------------------------------------------------------------------------------------------------------------- obs NOT SVR: 0| 318 1 0 47 Acc: 86.3% (POD) obs ISOLD: 1| 2 117 17 19 Acc: 75.0% (POD) obs LINE: 2| 5 30 81 28 Acc: 56.1% (POD) obs PULSE: 4| 45 14 9 211 Acc: 75.2% (POD) Accuracy: 85.8% 71.7% 74.2% 68.8% TSS: 0.67 Testing Confusion Matrix -------------------------------------------------------------------------------------------------------------------- obs NOT SVR: 0| 139 0 0 18 Acc: 88.5% (POD) obs ISOLD: 1| 1 47 4 14 Acc: 71.2% (POD) obs LINE: 2| 4 17 32 9 Acc: 51.6% (POD) obs PULSE: 4| 15 1 7 96 Acc: 80.7% (POD) Accuracy: 87.4% 72.3% 74.4% 70.1% TSS: 0.70 Confusion Matrix From Entry on Contest Test file obs NOT SVR: 0| 325 4 28 212 Acc: 57.1% (POD) obs ISOLD: 1| 1 100 16 14 Acc: 76.3% (POD) obs LINE: 2| 1 35 29 30 Acc: 30.5% (POD) obs PULSE: 4| 18 32 20 204 Acc: 74.5% (POD) Accuracy: 94.2% 58.5% 31.2% 44.3% TSS: 0.45

ANN Results using Original Training file and Contest Test file Training TSS: 0.65 POD FAR CSI Non Svr 0.86 0.14 0.75 Isold 0.70 0.29 0.54 Line 0.50 0.31 0.40 Pulse 0.73 0.33 0.53 Testing TSS: 0.65 POD FAR CSI Non Svr 0.86 0.06 0.81 Isold 0.64 0.34 0.48 Line 0.16 0.42 0.14 Pulse 0.82 0.41 0.52 Best Network: Iterations: 2000 Hidden Nodes Layer1: 14 Hidden Nodes Layer2: 3 Learn: 0.04 Momentum: 0.3 Training Confusion Matrix Forecast Class: 0 1 2 4 ------------------------------------------------------------------------------------------------------------------- obs NOT SVR: 0| 454 2 0 68 Acc: 86.5% (POD) obs ISOLD: 1| 2 156 29 33 Acc: 70.6% (POD) obs LINE: 2| 6 47 104 49 Acc: 50.1% (POD) obs PULSE: 4| 69 16 17 295 Acc: 73.9% (POD) Accuracy: 85.1% 70.2% 68.8% 66.1% TSS: 0.65 Testing Confusion Matrix -------------------------------------------------------------------------------------------------------------------- obs NOT SVR: 0| 492 3 0 74 Acc: 86.5% (POD) obs ISOLD: 1| 1 85 10 35 Acc: 64.9% (POD) obs LINE: 2| 1 29 16 49 Acc: 16.8% (POD) obs PULSE: 4| 33 12 2 227 Acc: 82.8% (POD) Accuracy: 93.4% 65.9% 57.1% 59.0% Confusion Matrix of the Official Contest Test File obs NOT SVR: 0| 498 0 3 68 Acc: 87.5% (POD) obs ISOLD: 1| 2 82 14 33 Acc: 62.6% (POD) obs LINE: 2| 4 28 41 22 Acc: 43.2% (POD) obs PULSE: 4| 57 24 25 168 Acc: 61.3% (POD) Accuracy: 88.8% 61.2% 49.4% 57.7% TSS: 0.58

Conclusions Improving classifications and bad/missing data in main data set Group of meteorologists to determine storm type from original radar data Better quality control (ie…low MaxVIL values for severe storms) Fill in missing values Uneven distribution of predictands Better skill with large amount of instances (Non-Svr and Pulse) Poor skill with few training instances (Line Storms) Provide an equal distribution of training instances for each predictand Decision Tree about equal in skill to original study and Neural Net an improvement (based on Answer Key). Possible that Neural Net is better than Decision Tree due to non-linear relationships Before Entry: Neural Net implementation “appeared” to show more skill than original study for training and test files made from original data set. Contest entry: not that great Re-trained NN with original, full Training file and tested it with contest Test file. Improved skill over baseline Test set. TSS Training: 0.65 TSS Testing: 0.65