Hybrid Intelligent Systems for Network Security Lane Thames Georgia Institute of Technology Savannah, GA

Slides:



Advertisements
Similar presentations
Bayesian network classification using spline-approximated KDE Y. Gurwicz, B. Lerner Journal of Pattern Recognition.
Advertisements

Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Naïve-Bayes Classifiers Business Intelligence for Managers.
Neural Networks Dr. Peter Phillips. The Human Brain (Recap of week 1)
Data Mining Classification: Alternative Techniques
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Kohonen Self Organising Maps Michael J. Watts
Introduction of Probabilistic Reasoning and Bayesian Networks
X0 xn w0 wn o Threshold units SOM.
Self Organizing Maps. This presentation is based on: SOM’s are invented by Teuvo Kohonen. They represent multidimensional.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Distinguishing Photographic Images and Photorealistic Computer Graphics Using Visual Vocabulary on Local Image Edges Rong Zhang,Rand-Ding Wang, and Tian-Tsong.
Neural Networks Chapter Feed-Forward Neural Networks.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Lecture 09 Clustering-based Learning
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
DEEDS Meeting Oct., 26th 2006 Dependable, Embedded Systems and Software Group Department of Computer Science Darmstadt University of Technology Summary.
 C. C. Hung, H. Ijaz, E. Jung, and B.-C. Kuo # School of Computing and Software Engineering Southern Polytechnic State University, Marietta, Georgia USA.
Using Bayesian Networks for Detecting Network Anomalies Lane Thames ECE 8833 Intelligent Systems.
DATA MINING : CLASSIFICATION. Classification : Definition  Classification is a supervised learning.  Uses training sets which has correct answers (class.
Layered Approach using Conditional Random Fields For Intrusion Detection.
Bayesian Networks. Male brain wiring Female brain wiring.
Hybrid Intelligent Systems for Network Security Lane Thames Georgia Institute of Technology Savannah, GA
Self-organizing Maps Kevin Pang. Goal Research SOMs Research SOMs Create an introductory tutorial on the algorithm Create an introductory tutorial on.
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
Chapter 9 Neural Network.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Bayesian Networks Martin Bachler MLA - VO
Analysing Microarray Data Using Bayesian Network Learning Name: Phirun Son Supervisor: Dr. Lin Liu.
Data Mining Knowledge on rough set theory SUSHIL KUMAR SAHU.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Machine Learning Neural Networks (3). Understanding Supervised and Unsupervised Learning.
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture notes 9 Bayesian Belief Networks.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Mining Logs Files for Data-Driven System Management Advisor.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
By Timofey Shulepov Clustering Algorithms. Clustering - main features  Clustering – a data mining technique  Def.: Classification of objects into sets.
Hybrid Intelligent Systems for Detecting Network Anomalies Lane Thames ECE 8833 Intelligent Systems.
CUNY Graduate Center December 15 Erdal Kose. Outlines Define SOMs Application Areas Structure Of SOMs (Basic Algorithm) Learning Algorithm Simulation.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
Text Categorization With Support Vector Machines: Learning With Many Relevant Features By Thornsten Joachims Presented By Meghneel Gore.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A self-organizing map for adaptive processing of structured.
An Introduction Student Name: Riaz Ahmad Program: MSIT( ) Subject: Data warehouse & Data Mining.
Dimensionality Reduction in Unsupervised Learning of Conditional Gaussian Networks Authors: Pegna, J.M., Lozano, J.A., Larragnaga, P., and Inza, I. In.
Data Mining and Decision Support
Using Bayesian Networks to Predict Plankton Production from Satellite Data By: Rob Curtis, Richard Fenn, Damon Oberholster Supervisors: Anet Potgieter,
Sporadic model building for efficiency enhancement of the hierarchical BOA Genetic Programming and Evolvable Machines (2008) 9: Martin Pelikan, Kumara.
Don’t Follow me : Spam Detection in Twitter January 12, 2011 In-seok An SNU Internet Database Lab. Alex Hai Wang The Pensylvania State University International.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
BAYESIAN LEARNING. 2 Bayesian Classifiers Bayesian classifiers are statistical classifiers, and are based on Bayes theorem They can calculate the probability.
Dependency Networks for Inference, Collaborative filtering, and Data Visualization Heckerman et al. Microsoft Research J. of Machine Learning Research.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Presented by Edith Ngai MPhil Term 3 Presentation
Data Mining, Neural Network and Genetic Programming
Molecular Classification of Cancer
Classification and Prediction
Pattern Recognition and Image Analysis
Advanced Artificial Intelligence Classification
Binghui Wang, Le Zhang, Neil Zhenqiang Gong
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Prepared by: Mahmoud Rafeek Al-Farra
Physics-guided machine learning for milling stability:
Using Bayesian Network in the Construction of a Bi-level Multi-classifier. A Case Study Using Intensive Care Unit Patients Data B. Sierra, N. Serrano,
Co-operative neural networks and ‘integrated’ classification
Presentation transcript:

Hybrid Intelligent Systems for Network Security Lane Thames Georgia Institute of Technology Savannah, GA

Presentation Overview Discuss the goals of this project Overview of Self Organizing Maps Overview of Bayesian Learning Networks Describe the details of the Hybrid System Review the Experimental Results Discuss Conclusions and Future Work Q&A

Internet Growth Internet Growth is Steadily Increasing Many different types of applications are now using the Internet as a communication channel

Data Source:

The life of a network security professional

Data Source:

Current Issues with Security Short time between disclosure of vulnerability and attack Huge Rule Base Huge Signature Databases Lag time between attack detection and signature creation Lag time between vulnerability discovery and patch deployment

Project Goals Develop an Intelligent System that works reliably with data that can be collected purely within a Computer Network Why? If security mechanisms are difficult to use, people will not use them. Using data from the network takes some of the burden off the end user

Hybrid Intelligent Systems A system was developed that made use of two types of Intelligence Algorithms: Self-Organizing Maps Self-Organizing Maps Bayesian Learning Networks Bayesian Learning Networks

Training and Testing Data Set KDD-CUP 99 Data Set The Data set used for the Third International Knowledge Discovery and Data Mining Tools Competition

Training and Testing Data Set 41 Total Features Categorized as: Basic TCP/IP features Basic TCP/IP features Content Features Content Features Time Based Traffic Features Time Based Traffic Features Host Based Traffic Features Host Based Traffic Features

Self Organizing Maps—SOM Pioneered by Dr. Teuvo Kohonen An algorithm that transforms high dimensional input data domains to elements of a low dimensional array of nodes

Self-Organizing Maps Input Data Vectors Parametric Vector associated with each element, i, of the grid

Self-Organizing Map A decoder function is defined on the basis of distance between the input vector and the parametric vector. The decoder function is used to map the image of the input vector onto the SOM grid. The decoder function is usually chosen to be either the Manhattan or Euclidean distance metric.

Self-Organizing Maps A Best Matching Unit, denoted as the index c, is chosen as the node on the SOM grid that is closest to the input vector

Self-Organizing Maps The dynamics of the SOM algorithm demand that the M i be shifted towards the order of X such that a set of values {M i } are obtained as the limit of convergence of the following:

Bayesian Learning Networks—BLN A BLN is a probabilistic model, and the network is built on the basis of a Directed Acyclic Graph (DAG) The directed edges of the graph represent relationships among the variables

Bayesian Learning Networks The Fundamental Equation: Bayes Theorem

Bayesian Learning Networks In Bayesian learning, we calculate the probability of an hypothesis and make predictions on that basis

Bayesian Learning Networks With BLN, we have conditional probabilities for each node given its parents The graph shows causal connections between the variables Prediction and abduction x1x1x1x1 x3x3x3x3 x2x2x2x2 x5x5x5x5 x4x4x4x4

Naïve Bayesian Learning Network The Naïve BLN is a special case of the general BLN It contains one root node which is called the class variable, C The leaf nodes are the attribute variables (X 1 … X i ) It is Naïve because it assumes the attributes are conditionally independent given the class C x3x3x3x3 x2x2x2x2 x1x1x1x1

The Naïve BLN Classifier Once the network is trained, it can be used to classify new examples where the attributes are given and the class variable is unobserved—abduction The Goal: Find the most probable class value given a set of attribute instantiations (X 1 … X i )

Hybrid System Details SOM Training Training Data Subset

Hybrid System Details Data BN Development Module Modified Data Trained SOM

Hybrid System Details BN Development Module Training Data Bayesian Training Structure File

Hybrid System Details Bayesian/SOMClassifier Classification File Test Data

Experimental Results 4 types of analyses were made with the dataset BLN analysis with network and host based data BLN analysis with network and host based data BLN analysis with network data BLN analysis with network data Hybrid analysis with network and host based data Hybrid analysis with network and host based data Hybrid analysis with network based data Hybrid analysis with network based data

Experimental Results BLN- Host/Network Based BLN- Network Based Hybrid- Host/Network Based Hybrid- Network Based Total Cases 65,50562,04765,50562,047 Correctly Classified 65,01959,73465,23861,631 % Correctly Classified 99.26%96.27%99.59%99.33% Number of Incorrectly Classified

Future and Current Work HoneyNet Project Resource Management System with Intelligent System Processing at the Core

Conclusions Intelligent System algorithms are very useful tools for applications in Network Security

Conclusions Questions remain to be answered: How will the system behave as the data becomes very noisy with respect to training data How will the system behave as the data becomes very noisy with respect to training data How will other intelligence algorithms compare in performance—training time, accuracy, robustness in noise How will other intelligence algorithms compare in performance—training time, accuracy, robustness in noise