Neural Computation Final Project -Earthquake Prediction 0368-4149-01, Spring 2004-5 Alon Talmor Ido Yariv.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

CSC321: Introduction to Neural Networks and Machine Learning Lecture 24: Non-linear Support Vector Machines Geoffrey Hinton.
Classification Classification Examples
ECG Signal processing (2)
Support Vector Machines
Machine learning continued Image source:
Lecture 22: Evaluation April 24, 2010.
What is Statistical Modeling
Studying Relationships between Human Posture and Health Risk Factors during Sedentary Activities Tejas Srinivasan Mentors: Vladimir Pavlovic Saehoon Yi.
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Support Vector Machines
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Failure Prediction in Hardware Systems Douglas Turnbull Neil Alldrin CSE 221: Operating System Final Project Fall
Supervised classification performance (prediction) assessment Dr. Huiru Zheng Dr. Franscisco Azuaje School of Computing and Mathematics Faculty of Engineering.
Speaker Adaptation for Vowel Classification
Evaluation.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Classification of Music According to Genres Using Neural Networks, Genetic Algorithms and Fuzzy Systems.
Experimental Evaluation
05/06/2005CSIS © M. Gibbons On Evaluating Open Biometric Identification Systems Spring 2005 Michael Gibbons School of Computer Science & Information Systems.
Today Evaluation Measures Accuracy Significance Testing
ENDA MOLLOY, ELECTRONIC ENG. FINAL PRESENTATION, 31/03/09. Automated Image Analysis Techniques for Screening of Mammography Images.
Predicting Income from Census Data using Multiple Classifiers Presented By: Arghya Kusum Das Arnab Ganguly Manohar Karki Saikat Basu Subhajit Sidhanta.
Outline 1-D regression Least-squares Regression Non-iterative Least-squares Regression Basis Functions Overfitting Validation 2.
Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering National Institute of Technology.
Support Vector Machines Mei-Chen Yeh 04/20/2010. The Classification Problem Label instances, usually represented by feature vectors, into one of the predefined.
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Processing of large document collections Part 3 (Evaluation of text classifiers, term selection) Helena Ahonen-Myka Spring 2006.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
START OF DAY 5 Reading: Chap. 8. Support Vector Machine.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
EVALUATION OF THE RADAR PRECIPITATION MEASUREMENT ACCURACY USING RAIN GAUGE DATA Aurel Apostu Mariana Bogdan Coralia Dreve Silvia Radulescu.
Active learning Haidong Shi, Nanyi Zeng Nov,12,2008.
USE RECIPE INGREDIENTS TO PREDICT THE CATEGORY OF CUISINE Group 7 – MEI, Yan & HUANG, Chenyu.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
Blackbox classifiers for preoperative discrimination between malignant and benign ovarian tumors C. Lu 1, T. Van Gestel 1, J. A. K. Suykens 1, S. Van Huffel.
A Kernel Approach for Learning From Almost Orthogonal Pattern * CIS 525 Class Presentation Professor: Slobodan Vucetic Presenter: Yilian Qin * B. Scholkopf.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Eco 6380 Predictive Analytics For Economists Spring 2016 Professor Tom Fomby Department of Economics SMU.
Gist 2.3 John H. Phan MIBLab Summer Workshop June 28th, 2006.
1 Learning Bias & Clustering Louis Oliphant CS based on slides by Burr H. Settles.
Evaluating Classifiers. Reading for this topic: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
Using Asymmetric Distributions to Improve Text Classifier Probability Estimates Paul N. Bennett Computer Science Dept. Carnegie Mellon University SIGIR.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
CSSE463: Image Recognition Day 14
Evaluating Classifiers
CS 4/527: Artificial Intelligence
Machine Learning Week 1.
CSSE463: Image Recognition Day 14
Model Evaluation and Selection
The Naïve Bayes (NB) Classifier
Model generalization Brief summary of methods
CAMCOS Report Day December 9th, 2015 San Jose State University
Support Vector Machines 2
Machine Learning: Lecture 5
Evaluation David Kauchak CS 158 – Fall 2019.
Presentation transcript:

Neural Computation Final Project -Earthquake Prediction , Spring Alon Talmor Ido Yariv

Project Goal The task Create an architecture which can provide a tight bound on the variability of the data (when no earthquake is present) and produces increasing abnormalities (outlieres) when earthquake conditions are beginning to develop. An architecture which receives the output of the previous architecture as an input provides a probably score for the development of an earthquake in the next 24 hours.

Earthquake Prediction – the problem The physics that control earthquakes are at this time poorly understood Good Sensors are hard to deploy Earthquakes are very long term – it is hard to collect large amount of statistics

Our Approach Changes in the eather and Seismic Events are very long term – use long term information. Good features are extracted by seismologists – use them. Use a more visual presentation of the data in order to choose the location, features and classifiers to work with.

Hypothesis We Assume two main types of events: A periodic release of pressure - where the local region of the earth is in a “loop” and releases its pressure periodically. A sudden release of a large amount of pressure - this event happens after a period of local “silence”. The earth can not release the pressure because of some blocking, pressure is accumulated, and then in one instance a large amount of pressure is released. This usually causes a large earthquake.

Gathering Data Reconstruction of the International Seismological Centre’s Database

International Seismological Centre A non-governmental organization charged with the final collection, analysis and publication of standard earthquake information from all over the world The parametric data is available to the general public via various means Offers a web interface for limited DB queries Concurrent querying are not allowed, due to bandwidth restrictions (56k/sec)

International Seismological Centre (Cont’d) The DB holds two kinds of data: Phases and Summery Phase data - The parametric data collected at each sensor worldwide (Local magnitude, time, etc.) Summery data - The summery of phase data from all stations, such as the weighted average of the magnitude, the calculated location, time of event, and so on.

International Seismological Centre (Cont’d) In order to quickly gather the data, we’ve written a Python script The script queries the DB via the web interface, parses the results, and inserts them into a local database The local database is based on SQLite, which makes it easy to reuse the data in future research works

ISC’s website

ISC’s Query Page

ISC’s Results Page (Mag.+Phases)

ISC’s Bandwidth Problems

Method – Analyzing the data

Method – Selecting the ROI We assume events are influnced mostly by activity in close regions. Region Of Intereset of 3-4 geographical degrees ( Km) is chosen.

Method – Building the basic data Only MB magnitude and Geographic location are used. One Dimensional Data is extracted for the chosen ROI. Data chosen between

Considering the location - ROI dependant weight matrix

Building the basic data (Cont’d) DataSample(i) = sqrt(sum(LocationWeightMatrix(Event(j)) *Magnitude(Event(j))^2) | 0<j<number of events in current sampling period The sampling period used in this project is 3600 seconds = 1 hour.

Constructing the features Feature Extracted: FTT – spectral features Ceprstum –real(log(PowerFFT(Frame)))

Constructing the classification labels

Feature Dimension reduction Two method used here 1. PCA ( reducing from 20 to ~5 features) 2. The Human Eye

Choosing the model 5 classifiers were compared: Perceptron Expectation-Maximization – Gaussian Mixture Models Radial Basis function Artificial neural network Support Vector Machine

Classifying Earthquakes Using SVM

Method The data was divided (after randomly shuffling it) into two separate sets: A training set (75%) - Used for constructing the classification model A testing set (25%) - Used for testing the model The model was constructed using only the training set, and tested on the testing set

Method (Cont’d) SVM has several variables which need to be set prior to the model construction process, namely the kernel function and its associated parameters Since RBF performed well on the data, and it is known as a generally good first pick, we chose it as our kernel function Thus, there were two free parameters left for us to set

Method (Cont’d) Cross-validation on the training set was used in order to evaluate various pairs of the parameters For each pair, the training set was divided into 5 different folds A SVM model was trained on four folds, and tested on the fifth This process was repeated for each of the folds (leave one out)

Method (Cont’d) In order to test the model on a set it is not enough to measure the general accuracy of the model, that is, the number of vectors which the model had successfully labeled Since there are more negative labeled vectors than positive ones, a trivial classifier which would label any vector as being negative could be evaluated as a relatively good classifier (with a record of over 80% success) Thus, it is important to divide the resulting classification labels into four groups: True Positives, True Negatives, False Positives and False Negatives

Method (Cont’d) Once statistics of the four above groups was collected, a break-even function was used in order to compare the models: The parameters of the best model, that is, the model which had the best break-even result were saved, and a model with these parameters was trained on the whole training set p=True Positives/Sum Positives n=True Negatives/Sum Negatives

Results Taking the Sumatra area as a test case (Longitude = 94, Latitude = 8) and labeling any earthquake with a magnitude over 5.5 as positive, the following typical results were achieved: Overall result Percentag e Break-Even Percentag e True Negative Percentag e True Positives Percentag e Mirror Location Variable Main Location Variable ± ± ± ± ± ± ± ± ± ± ±477±

Results (Cont’d) Results show that SVM can classify the extracted features quite well, achieving a record of 83.3%/96.28% (True Positive/True Negative) This is quite surprising, considering the problem's nature Further research could calibrate the variables more accurately, achieving even better results

The GUI

The GUI (Cont’d)