Predicting Earthquakes By Lois Desplat. Why Predict Earthquakes?  To minimize the loss of life and property.  Unfortunately, current techniques do not.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

G53MLE | Machine Learning | Dr Guoping Qiu
Brief introduction on Logistic Regression
Decision Tree Approach in Data Mining
Data Mining Classification: Alternative Techniques
Support Vector Machines and Kernels Adapted from slides by Tim Oates Cognition, Robotics, and Learning (CORAL) Lab University of Maryland Baltimore County.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Supervised Learning Recap
What is Statistical Modeling
x – independent variable (input)
Data Mining Techniques Outline
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Statistical Methods Chichang Jou Tamkang University.
2-1 Sample Spaces and Events Conducting an experiment, in day-to-day repetitions of the measurement the results can differ slightly because of small.
Data mining and statistical learning - lecture 13 Separating hyperplane.
Data Mining: A Closer Look Chapter Data Mining Strategies (p35) Moh!
Support Vector Machines
Experimental Evaluation
Role and Place of Statistical Data Analysis and very simple applications Simplified diagram of a scientific research When you know the system: Estimation.
Robust Bayesian Classifier Presented by Chandrasekhar Jakkampudi.
Chapter 5 Data mining : A Closer Look.
Basic Data Mining Techniques
Intrusion Detection Jie Lin. Outline Introduction A Frame for Intrusion Detection System Intrusion Detection Techniques Ideas for Improving Intrusion.
Overview G. Jogesh Babu. Probability theory Probability is all about flip of a coin Conditional probability & Bayes theorem (Bayesian analysis) Expectation,
DATA MINING : CLASSIFICATION. Classification : Definition  Classification is a supervised learning.  Uses training sets which has correct answers (class.
Principles of Pattern Recognition
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
Chapter 8 Discriminant Analysis. 8.1 Introduction  Classification is an important issue in multivariate analysis and data mining.  Classification: classifies.
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Ensemble Classification Methods Rayid Ghani IR Seminar – 9/26/00.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Comparison of Bayesian Neural Networks with TMVA classifiers Richa Sharma, Vipin Bhatnagar Panjab University, Chandigarh India-CMS March, 2009 Meeting,
Ruizhu Yang 04/25/2014. Otari G V, Kulkarni R V. A Review of Application of Data Mining in Earthquake Prediction[J] Dzwinel W, Yuen D A, Boryczko.
LOGISTIC REGRESSION David Kauchak CS451 – Fall 2013.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
CLASSIFICATION: Ensemble Methods
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Optimal Bayes Classification
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
An Investigation of Commercial Data Mining Presented by Emily Davis Supervisor: John Ebden.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
November 30, PATTERN RECOGNITION. November 30, TEXTURE CLASSIFICATION PROJECT Characterize each texture so as to differentiate it from one.
Classification Problem GivenGiven Predict class label of a given queryPredict class label of a given query
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
Inference: Probabilities and Distributions Feb , 2012.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Supervised Machine Learning: Classification Techniques Chaleece Sandberg Chris Bradley Kyle Walsh.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Data Mining By: Johan Johansson. Mining Techniques Association Rules Association Rules Decision Trees Decision Trees Clustering Clustering Nearest Neighbor.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
BAYESIAN LEARNING. 2 Bayesian Classifiers Bayesian classifiers are statistical classifiers, and are based on Bayes theorem They can calculate the probability.
Overfitting, Bias/Variance tradeoff. 2 Content of the presentation Bias and variance definitions Parameters that influence bias and variance Bias and.
Ch 7. Computing with Population Coding Summarized by Kim, Kwonill Bayesian Brain: Probabilistic Approaches to Neural Coding P. Latham & A. Pouget.
Howard Community College
Data Transformation: Normalization
Chapter 7. Classification and Prediction
Chapter 12 Object Recognition
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Basic machine learning background with Python scikit-learn
PROBABILITY AND STATISTICS
ECE 5424: Introduction to Machine Learning
Data Mining Practical Machine Learning Tools and Techniques
of the Artificial Neural Networks.
Classification and Prediction
The Naïve Bayes (NB) Classifier
Parametric Methods Berlin Chen, 2005 References:
Using Clustering to Make Prediction Intervals For Neural Networks
Statistics Review (It’s not so scary).
Presentation transcript:

Predicting Earthquakes By Lois Desplat

Why Predict Earthquakes?  To minimize the loss of life and property.  Unfortunately, current techniques do not have a high enough accuracy to be able to accurately predict earthquakes.

Estimating earthquake probabilities  Scientists study the histories of large earthquakes in a specific area  The rate at which strain accumulates in the rock

Methods to earthquake prediction  Need to construct models based on: Partial differential equations Partial differential equations Finite automata Finite automata Supervised learning techniques: Supervised learning techniques: Decision TreeDecision Tree Bayesian ClassificationBayesian Classification Feed-Forward Neural NetworksFeed-Forward Neural Networks

Decision Tree  Tries to generate rules with high accuracy  ID3, …

Bayesian Classifiers  They are statistical classifiers  Only needs a small sample to find the means and variances of the variables necessary for classification  It can find the probability that a given sample belongs to a certain class (earthquake > 3.0)  Uses Bayes Theorem

Feed-Forward Neural Network  Network given a set of input and respective output to start learning  It connects each Perceptron and the algorithm tries to minimize the weigths between Perceptrons to the minimum so that the input give the right output

The Bagging Method  Combine the predictions of the past three algorithms  You get a much more accurate prediction  Give different learning samples to each algorithm

Some Problems  The Data can have a lot of extra information that adds noise i.e. We might not want small scale earthquakes that are really just aftershocks of big earthquakes  We only look at the data in 1 dimension, maybe if we plot the data in multiple dimensions, we might some patterns

Not Good Enough!  Authors claim that their bagging method has 92% accuracy.  Highly doubt accuracy of that number but even if true: We still cannot predict earthquakes with enough confidence We still cannot predict earthquakes with enough confidence

Solution  Do short-term predictions instead of long- term  Analyze the data in multiple dimensions over space, time and feature space.

Visualization of the Data Space

 Data Space uses Magnitude, Epicentral Coordinate, Depth and Time of occurrence  7D space uses: NS: Degree of spatial non-randomness at short distances NS: Degree of spatial non-randomness at short distances LS: Degree of spatial non-randomness at long distances LS: Degree of spatial non-randomness at long distances CD: Spatial correlation dimension CD: Spatial correlation dimension SR: Degree of spatial repetitiveness SR: Degree of spatial repetitiveness AZ: Average Depth AZ: Average Depth TI: Time Interval for the occurance of 100 events in the sample space. TI: Time Interval for the occurance of 100 events in the sample space. MR: Ratio of two events falling into different magnitude ranges MR: Ratio of two events falling into different magnitude ranges

Conclusion  This method is able to find precursor events just prior to an earthquake.  Unfortunately, it only works for short-term predictions and cannot predict years or months in advance.  Plenty of work can still be done!

References  “Predicting the Earthquake using Bagging Method in Data Mining”, S.Sathiyabama, K.Thyagarajah, D. Ayyamuthukumar  “A Bagging Method using Decision Trees in the Role of Base Classifiers”, Kristína Machová, František Barčák, Peter Bednár  “Cluster Analysis, Data-Mining, Multi-dimensional Visualization of Earthquakes over Space, Time and Feature Space”, Witold Dzwinel, David A. Yuen, Krzysztor Boryczko, Yehuda Ben-Zion, Shoichi Yoshioka, Takeo Ito  college/projects-00/neural- networks/Architecture/feedforward.html