UNIVERSITY OF JYVÄSKYLÄ Yevgeniy Ivanchenko Yevgeniy Ivanchenko University of Jyväskylä

Slides:



Advertisements
Similar presentations
UNIVERSITY OF JYVÄSKYLÄ Resource Discovery in P2P Networks Using Evolutionary Neural Networks Presentation for International Conference on Advances in.
Advertisements

Slides from: Doug Gray, David Poole
Improvement on LEACH Protocol of Wireless Sensor Network
UNIVERSITY OF JYVÄSKYLÄ Building NeuroSearch – Intelligent Evolutionary Search Algorithm For Peer-to-Peer Environment Master’s Thesis by Joni Töyrylä
Modeling and Analysis of Random Walk Search Algorithms in P2P Networks Nabhendra Bisnik, Alhussein Abouzeid ECSE, Rensselaer Polytechnic Institute.
Self Organization: Competitive Learning
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Kohonen Self Organising Maps Michael J. Watts
Unsupervised Learning with Artificial Neural Networks The ANN is given a set of patterns, P, from space, S, but little/no information about their classification,
Artificial neural networks:
Unsupervised Networks Closely related to clustering Do not require target outputs for each input vector in the training data Inputs are connected to a.
Kostas Kontogiannis E&CE
Generated Waypoint Efficiency: The efficiency considered here is defined as follows: As can be seen from the graph, for the obstruction radius values (200,
X0 xn w0 wn o Threshold units SOM.
Self Organizing Maps. This presentation is based on: SOM’s are invented by Teuvo Kohonen. They represent multidimensional.
Models and Security Requirements for IDS. Overview The system and attack model Security requirements for IDS –Sensitivity –Detection Analysis methodology.
1 Complexity of Network Synchronization Raeda Naamnieh.
Simple Neural Nets For Pattern Classification
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Effects of Applying Mobility Localization on Source Routing Algorithms for Mobile Ad Hoc Network Hridesh Rajan presented by Metin Tekkalmaz.
Radial Basis Functions
Un Supervised Learning & Self Organizing Maps Learning From Examples
Basic Data Mining Techniques Chapter Decision Trees.
Active Learning with Support Vector Machines
Evaluating Hypotheses
Switch to Top-down Top-down or move-to-nearest Partition documents into ‘k’ clusters Two variants “Hard” (0/1) assignment of documents to clusters “soft”
UNIVERSITY OF JYVÄSKYLÄ Resource Discovery Using NeuroSearch Presentation for the Agora Center InBCT-seminar Mikko Vapa, researcher InBCT 3.2.
UNIVERSITY OF JYVÄSKYLÄ Resource Discovery in Unstructured P2P Networks Distributed Systems Research Seminar on Mikko Vapa, research student.
Distributed Model-Based Learning PhD student: Zhang, Xiaofeng.
A Hybrid Self-Organizing Neural Gas Network James Graham and Janusz Starzyk School of EECS, Ohio University Stocker Center, Athens, OH USA IEEE World.
Radial Basis Function (RBF) Networks
Radial-Basis Function Networks
Pipelined Two Step Iterative Matching Algorithms for CIOQ Crossbar Switches Deng Pan and Yuanyuan Yang State University of New York, Stony Brook.
Copyright © Cengage Learning. All rights reserved. CHAPTER 11 ANALYSIS OF ALGORITHM EFFICIENCY ANALYSIS OF ALGORITHM EFFICIENCY.
 C. C. Hung, H. Ijaz, E. Jung, and B.-C. Kuo # School of Computing and Software Engineering Southern Polytechnic State University, Marietta, Georgia USA.
19.5 Numeric Integration and Differentiation
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
Artificial Neural Network Unsupervised Learning
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Content Addressable Network CAN. The CAN is essentially a distributed Internet-scale hash table that maps file names to their location in the network.
Optimization with Neural Networks Presented by: Mahmood Khademi Babak Bashiri Instructor: Dr. Bagheri Sharif University of Technology April 2007.
REECH ME: Regional Energy Efficient Cluster Heads based on Maximum Energy Routing Protocol Prepared by: Arslan Haider. 1.
UNSUPERVISED LEARNING NETWORKS
Dual-Region Location Management for Mobile Ad Hoc Networks Yinan Li, Ing-ray Chen, Ding-chau Wang Presented by Youyou Cao.
Clustering.
TreeSOM :Cluster analysis in the self- organizing map Neural Networks 19 (2006) Special Issue Reporter 張欽隆 D
381 Self Organization Map Learning without Examples.
CUNY Graduate Center December 15 Erdal Kose. Outlines Define SOMs Application Areas Structure Of SOMs (Basic Algorithm) Learning Algorithm Simulation.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A self-organizing map for adaptive processing of structured.
Model-based learning: Theory and an application to sequence learning P.O. Box 49, 1525, Budapest, Hungary Zoltán Somogyvári.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Adaptive FIR Neural Model for Centroid Learning in Self-Organizing.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
CS 6401 Intra-domain Routing Outline Introduction to Routing Distance Vector Algorithm.
Dr. Chen, Data Mining  A/W & Dr. Chen, Data Mining Chapter 3 Basic Data Mining Techniques Jason C. H. Chen, Ph.D. Professor of MIS School of Business.
Copyright © Cengage Learning. All rights reserved. 1 Functions and Limits.
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
Self-Organizing Network Model (SOM) Session 11
Data Mining, Neural Network and Genetic Programming
Convolutional Networks
by Saltanat Mashirova & Afshin Mahini
An Artificial Intelligence Based Fisheries Research On The Evaluation Of Gnathiid Parasitism In Goldblotch Grouper of ISKENDERUN BAY ORAL, M. GENÇ, M.A. ELMAS,
The Naïve Bayes (NB) Classifier
Artificial Neural Networks
Unsupervised Networks Closely related to clustering
Presentation transcript:

UNIVERSITY OF JYVÄSKYLÄ Yevgeniy Ivanchenko Yevgeniy Ivanchenko University of Jyväskylä

UNIVERSITY OF JYVÄSKYLÄ 2004 OBJECTIVES (I) Since nothing is known about decision mechanism of NeuroSearch we need to look inside the algorithm to understand its behavior. Since nothing is known about behavior of NeuroSearch algorithm in dynamic environment, we need to know its behavior under conditions that are approximated to real life situation.

UNIVERSITY OF JYVÄSKYLÄ 2004 OBJECTIVES (II) To understand behavior of NeuroSearch data analysis techniques were used. The Self-Organizing Maps (SOM) is well known tool to perform data mining task. Set of rules was obtained based on the analysis of NeuroSearch. The rules were tested in static environment. The question that arises here: Is it possible to use the algorithm, which utilized properties of static environment, in dynamic scenario?

UNIVERSITY OF JYVÄSKYLÄ 2004 OBJECTIVES (III) If we know the inner structure of decision mechanism of NeuroSearch we will be able to tell about contribution of every input to particular decision of the algorithm. This for example can be used to remove unnecessary input information. This also can help evaluate complexity and robustness of the algorithm.

UNIVERSITY OF JYVÄSKYLÄ 2004 SOM (I) SOM is neural network model that maps high dimensional space onto low-dimensional space (usually two dimensional). After using SOM algorithm similar vectors from the input space are located near each other in the output space. This can help investigate properties of obtained clusters and as a consequence causes that produced these clusters on the output map.

UNIVERSITY OF JYVÄSKYLÄ 2004 SOM (II) R1 R2 Usually SOM represents itself either hexagonal or rectangular grid of neurons. In the figure R1 and R2 denote different neighborhood size. During the training process size of neighborhood is slightly decreased to provide more accurate adjustment of the weights of the neurons.

UNIVERSITY OF JYVÄSKYLÄ 2004 SOM (III) BMU In the figure one can see that the neurons that are ‘covered’ by neighborhood kernel function move closer to the input vector. Best Matching Unit (BMU) is the closest neuron to the current input vector. The weights of the neurons are updated according to the kernel function and the distance to BMU.

UNIVERSITY OF JYVÄSKYLÄ 2004 DATA ANALYSIS (I) NeuroSearch can be considered as the main part of information model of the system. To build this system black box method was used: we are modeling external behavior of the system and at the same time we don’t know what are the causes of particular behavior of the system. To investigate decision mechanism of NeuroSearch analysis of input-output pairs was done using SOM.

UNIVERSITY OF JYVÄSKYLÄ 2004 DATA ANALYSIS (II) To perform the analysis we used Component plane & U-matrix with ‘hit’ distribution on it. Component plane visualizes values of all components of the vectors according to the output map. U-matrix is one of possible ways to visualize the output map. The ‘hits’ on the U-matrix correspond to the decisions of NeuroSearch. This approach allows us investigating not only contribution of each component to particular decision, but also the correlations between components.

UNIVERSITY OF JYVÄSKYLÄ 2004 From The figure shows U-matrix (the left side of the figure) & fragment of Component plane (the right side of the figure). It is easy to see variable From is responsible for stopping further forwarding of the queries where it is 1. Other variables have different values in the area where From is 1, for example variable toUnsearchedNeighbors has different values in this area. DATA ANALYSIS (III) toUnsearchedNeighbors U-matrix

UNIVERSITY OF JYVÄSKYLÄ 2004 DATA ANALYSIS (IV) After the analysis it was found that 4 variables (From, toVisited, Sent and currentVisited) are responsible for stopping further forwarding of the queries. Variables toUnsearchedNeighbors and Neighbors are correlated. Variables packetsNow and Hops are highly correlated. Variables fromNeighborAmount, packetsNow and Hops are correlated somehow. NeuroSearch mostly doesn’t send the queries further if Neighbors or toUnsearchedNeighbors is small.

UNIVERSITY OF JYVÄSKYLÄ 2004 DATA ANALYSIS (V) Further investigation of the algorithm is based on Hops because only this variable shows the state of the algorithm in particular time interval, in other words analyzing intervals of this variable we can monitor the queries through their path. The maximum length of the queries’ path is 7. Thus we have 7 different cases to analyze. Data for each case contains only samples with the currently investigating value of Hops variable. All samples where at least one of From, Sent, currentVisted or toVisited variables is equal to 1 were removed as well. It is because we already know behavior of the algorithm in these areas.

UNIVERSITY OF JYVÄSKYLÄ 2004 DATA ANALYSIS (VI) After investigation of the algorithm for the different values of Hops we have produced Rule Based Algorithm (RBA). RBA is based on rules that were extracted using analysis of U-matrix and corresponding component plane. General strategy of the algorithm is quite simple: A decision is mostly based on interconnection between Hops, Neighbors/toUnsearchedNeighbors and NeighborsOrder values. In the beginning the algorithm sends the queries to the most connected nodes. When number of hops in the query is increasing NeuroSearch slightly starts to forward the queries to low-connected nodes.

UNIVERSITY OF JYVÄSKYLÄ 2004 DATA ANALYSIS (VII) AlgorithmPacketsReplies BFS BFS NeuroSearch RBA The table shows efficiency of four algorithms. One can see that NeuroSearch and RBA have almost the same level of performance. This means that RBA adapted behavior of NeuroSearch and we can say that SOM suits well for analyzing of NeuroSearch. Both these algorithms have better performance compared to BFS2 and BFS3. Comparison between algorithms

UNIVERSITY OF JYVÄSKYLÄ 2004 DYNAMIC ENVIRONMENT (I) Since RBA is based on decision mechanism of NeuroSearch it is possible to evaluate behavior of NeuroSearch using RBA in dynamic environment. As a simulation environment P2P extension for NS-2 was built. The environment provides quite high dynamical changes. There are two different classes of probabilities that define dynamical changes in the network. The first class is defined randomly before starting the simulation. The second is defined by the formulas:

UNIVERSITY OF JYVÄSKYLÄ 2004 DYNAMIC ENVIRONMENT (II) To make qualitative evaluation of performance, RBA was compared to BFS2 and BFS3 in static and dynamic environments. Number of replies and amount of used packets in static environment are shown in the figures:

UNIVERSITY OF JYVÄSKYLÄ 2004 DYNAMIC ENVIRONMENT (III) Analyzing behavior of the algorithms in static environment one can see that mostly RBA locates more resources than BFS2 and significantly less than BFS3. In general RBA uses more packets than BFS2 and significantly less than BFS3. This situation satisfies us because RBA is based on NeuroSearch’s decision mechanism that is trained to locate only half of available resources. In some points RBA locates more resources than BFS3 algorithm and in the same time uses less packets. This means that if some resource isn’t common in the network, RBA and as a consequence NeuroSearch can find enough instances of this resource.

UNIVERSITY OF JYVÄSKYLÄ 2004 DYNAMIC ENVIRONMENT (IV) Number of replies and amount of used packets in dynamic environment are shown in the figures: Analyzing the figures one can see that performance of the algorithms didn’t suffer so much in the dynamic environment.

UNIVERSITY OF JYVÄSKYLÄ 2004 DYNAMIC ENVIRONMENT (V) Algorithm PacketsReplies Static dynamic staticdynamic BFS BFS RBA Total number of located resources and used packets in static and dynamic environment are shown in the table: The algorithms still can find enough resources in dynamic environment. There are two possible causes that can explain the fact that all investigated algorithms found a little bit fewer resources: 1) Some nodes in offline mode could contain queried resources. 2) Some nodes in offline mode could lie on possible path of the query.

UNIVERSITY OF JYVÄSKYLÄ 2004 DYNAMIC ENVIRONMENT (VI) The algorithms used less packets in dynamic environment than in static environment. BFS strategy is very sensitive to the size of the network, because BFS based algorithms used significantly less packets in dynamic environment where size of the network was smaller all the simulation time. RBA used approximately the same amount of packets in both environments. Therefore we can say that RBA is not strongly sensitive to the size of the network.

UNIVERSITY OF JYVÄSKYLÄ 2004 FUTURE WORK Developing the supervised approach to train NeuroSearch. Developing modification of the algorithm for ad hoc wireless P2P networks. Paying more detailed and deeper attention to the inner structure of the algorithm, using knowledge discovery methods. Investigating and utilizing properties of other P2P algorithms to answer to the question about adding these properties to NeuroSearch.

UNIVERSITY OF JYVÄSKYLÄ 2004 Thank you!