JingTao Yao Growing Hierarchical Self-Organizing Maps for Web Mining Joseph P. Herbert and JingTao Yao Department of Computer Science, University or Regina.

Slides:



Advertisements
Similar presentations
Memristor in Learning Neural Networks
Advertisements

Data Mining Classification: Alternative Techniques
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
1 Software Design Introduction  The chapter will address the following questions:  How do you factor a program into manageable program modules that can.
TEMPLATE DESIGN © Self Organized Neural Networks Applied to Animal Communication Abstract Background Objective The main.
Self Organization of a Massive Document Collection
Self Organization: Competitive Learning
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Kohonen Self Organising Maps Michael J. Watts
Artificial neural networks:
Adaptive Resonance Theory (ART) networks perform completely unsupervised learning. Their competitive learning algorithm is similar to the first (unsupervised)
X0 xn w0 wn o Threshold units SOM.
Xyleme A Dynamic Warehouse for XML Data of the Web.
Self-Organizing Hierarchical Neural Network
University of CreteCS4831 The use of Minimum Spanning Trees in microarray expression data Gkirtzou Ekaterini.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
1 Kunstmatige Intelligentie / RuG KI2 - 7 Clustering Algorithms Johan Everts.
Lecture 09 Clustering-based Learning
Chapter 3: Cluster Analysis  3.1 Basic Concepts of Clustering  3.2 Partitioning Methods  3.3 Hierarchical Methods The Principle Agglomerative.
Clustering Unsupervised learning Generating “classes”
 C. C. Hung, H. Ijaz, E. Jung, and B.-C. Kuo # School of Computing and Software Engineering Southern Polytechnic State University, Marietta, Georgia USA.
Machine Learning. Learning agent Any other agent.
CS623: Introduction to Computing with Neural Nets (lecture-20) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
KOHONEN SELF ORGANISING MAP SEMINAR BY M.V.MAHENDRAN., Reg no: III SEM, M.E., Control And Instrumentation Engg.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
Chapter 9 Neural Network.
Artificial Neural Network Unsupervised Learning
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
Classifying Attributes with Game- theoretic Rough Sets Nouman Azam and JingTao Yao Department of Computer Science University of Regina CANADA S4S 0A2
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
NEURAL NETWORKS FOR DATA MINING
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Soft Computing Lecture 14 Clustering and model ART.
Hierarchical Clustering of Gene Expression Data Author : Feng Luo, Kun Tang Latifur Khan Graduate : Chien-Ming Hsiao.
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
Understanding User’s Query Intent with Wikipedia G 여 승 후.
Clustering.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. Externally growing self-organizing maps and its application to database visualization and exploration.
By Timofey Shulepov Clustering Algorithms. Clustering - main features  Clustering – a data mining technique  Def.: Classification of objects into sets.
381 Self Organization Map Learning without Examples.
Computational Biology Clustering Parts taken from Introduction to Data Mining by Tan, Steinbach, Kumar Lecture Slides Week 9.
2015/12/251 Hierarchical Document Clustering Using Frequent Itemsets Benjamin C.M. Fung, Ke Wangy and Martin Ester Proceeding of International Conference.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Chapter 8: Adaptive Networks
Hazırlayan NEURAL NETWORKS Backpropagation Network PROF. DR. YUSUF OYSAL.
Self-Organizing Maps (SOM) (§ 5.5)
Citation-Based Retrieval for Scholarly Publications 指導教授:郭建明 學生:蘇文正 M
CHAPTER 14 Competitive Networks Ming-Feng Yeh.
BINARY TREES Objectives Define trees as data structures Define the terms associated with trees Discuss tree traversal algorithms Discuss a binary.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Growing Hierarchical Tree SOM: An unsupervised neural.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Soft Computing Lecture 15 Constructive learning algorithms. Network of Hamming.
Linear Models & Clustering Presented by Kwak, Nam-ju 1.
::Network Optimization:: Minimum Spanning Trees and Clustering Taufik Djatna, Dr.Eng. 1.
Mehdi Kargar Department of Computer Science and Engineering
Self-Organizing Network Model (SOM) Session 11
Data Mining, Neural Network and Genetic Programming
Self-Organizing Maps for Content-Based Image Database Retrieval
Self organizing networks
Lecture 22 Clustering (3).
Graphs Chapter 11 Objectives Upon completion you will be able to:
of the Artificial Neural Networks.
Capabilities of Threshold Neurons
Authors: Barry Smyth, Mark T. Keane, Padraig Cunningham
Artificial Neural Networks
Presentation transcript:

JingTao Yao Growing Hierarchical Self-Organizing Maps for Web Mining Joseph P. Herbert and JingTao Yao Department of Computer Science, University or Regina CANADA S4S 0A2

JingTao Yao 2 Growing Hierarchical Self-Organizing Maps for Web Mining Introduction Many information retrieval and machine learning techniques have not evolved to survive the Web environment. There are two major problems in applying some machine learning techniques for Web Mining: 1.The dynamic and ever-changing nature of Web data. 2.Dimensionality and sheer size of Web data.

JingTao Yao 3 Growing Hierarchical Self-Organizing Maps for Web Mining Introduction Three domains of application: Web content mining, Web usage mining, and Web structure mining. Self-Organizing Maps (SOM) have been used for: –Web page clustering –Document retrieval –Recommendation systems

JingTao Yao 4 Growing Hierarchical Self-Organizing Maps for Web Mining Growing Hierarchical SOMs Growing Hierarchical SOMs are a hybridization of Growing SOMs and Hierarchical SOMs: –Growing SOMs have a dynamic topology of neurons to help solve the dynamic nature of data on the Web. –Hierarchical SOMs are multi-level systems designed to minimize the high dimensionality problem of data. Together, the hybrid system provides a logical solution when considering the combined problem of dynamic, high-dimensional data sources.

JingTao Yao 5 Growing Hierarchical Self-Organizing Maps for Web Mining The Consistency Problem The growing hierarchical SOM model suffers from a new problem: –Maintaining consistency of hierarchical relationships between levels. –Training is done locally, without consideration of how changes effect other SOMs that have connection to the local focus. The Web Mining model for Self-Organizing Maps solve this problem through bidirectional update propagation.

JingTao Yao 6 Growing Hierarchical Self-Organizing Maps for Web Mining The Web Mining Model for Self-Organizing Maps C w* U Input Layer: Each vector is inserted into the SOM network for the first stage of competition. An iteration is complete once all input vectors have been presented. Hierarchy Layer: A suitable level within the hierarchy of SOMs is found by traversing the tree. The SOM whose collectively maximum similarity to the input is marked and passed to the next layer. Growth Layer: This layer determines whether or not neurons need to be added or subtracted from the current SOM. If error is above an upper bound threshold, neurons are added. If error is below a lower bound threshold, neurons are removed. Update Layer: This layer updates the winning neuron and the neighborhood associated with it. Bidirectional Update Propagation updates parent neurons and children feature maps that are associated with the winning neuron.

JingTao Yao 7 Growing Hierarchical Self-Organizing Maps for Web Mining Formal Definition A = {A 1, …, A t } –A set of hierarchy levels. A i = {W i,1, …, W i,m } –A set of individual SOMs. W i,j = {w 1, …, w n } –A SOM of n neurons. –Each neuron contains a storage unit s k and a weight vector v k.

JingTao Yao 8 Growing Hierarchical Self-Organizing Maps for Web Mining Three Basic Functions Three functions are introduced for actions on the system: Lev() –Returns the hierarchy level that a SOM currently resides on. Chd() –Returns a set of SOMs that have child relationship to a particular neuron. Par() –Returns the parent SOM of a particular neuron.

JingTao Yao 9 Growing Hierarchical Self-Organizing Maps for Web Mining Process Flow for Training Input Determine winning neuron on current level Propagate Updates Downwards Proceed to next Hierarchy Level with closest neuron Propagate Updates Upwards Is neuron and input similar enough ? NY Add / Subtract Neuron Update Neighborhood Update Winner Neuron Is map representing input enough ? YN Bidirectional Propagation 1.Input is inserted into network 2.Neuron that is most similar is selected. 3.Descend through hierarchy until similarity is maximal. 4.Determine whether correct number of neurons represent pattern. 5.Add / Subtract neurons accordingly. 6.Update neuron and neighbourhood. 7.Update children SOMs. 8.Update parent SOM

JingTao Yao 10 Growing Hierarchical Self-Organizing Maps for Web Mining Conceptual View At the top-most hierarchy level (A 1 ), only one feature map would exist. –This map contains the absolute highest conceptual view of the entire hierarchical structure. Additional SOMs on subsequent levels offer more precise pattern abstraction. –SOMs are denoted by the sequence of their parents. –W 3,6,4 denotes the feature map is the fourth map on the third level derived from the sixth map on the previous level.

JingTao Yao 11 Growing Hierarchical Self-Organizing Maps for Web Mining Learning of Features Once a winning neuron wi*has been identified (denoted by an asterisk), its weight vector vi* is updated according to a learning rate α. –The value α decays over time according to the current training iteration. –v i * (q) = v i * (q-1) + α(p k (q) – v i * (q-1)) The neighbourhood must also be updated with a modified learning rate α /. –v Ni*(d) (q) = v Ni*(d) (q-1) + α(p k (q) – v Ni*(d) (q-1))

JingTao Yao 12 Growing Hierarchical Self-Organizing Maps for Web Mining Bidirectional Update Propagation Let w i * be the winning neuron in SOM W j,k for input k. To propagate upwards: –Calculate Par(w i * ) = W j-1,m, where Lev(W j-1,m ) < Lev(W j,k ). –Update all neurons w a contained in W j-1,m that are similar to w i *. –v a * (q) = v a * (q-1) + β(p k (q) – v a * (q-1))

JingTao Yao 13 Growing Hierarchical Self-Organizing Maps for Web Mining Bidirectional Update Propagation To propagate downwards: –Calculate Chd(w i * ) = A * j+1, where j+1 is the next level in the hierarchy succeeding level j. –Update the corresponding weight vectors for all neurons w b in SOM W j+1,t, where W j+1,t is on the lower level A * j+1. –v b * (q) = v b * (q-1) + γ(p k (q) – v b * (q-1)) The learning rates β and γ are derived from a value of α. Generally, updates to a parent neuron are not as strong as updates to children neurons.

JingTao Yao 14 Growing Hierarchical Self-Organizing Maps for Web Mining Web-based News Coverage Example The top-most level of the hierarchy contains news articles pertaining to high-level concepts and are arranged according to their features. The entire collection of Web documents on the online news site are presented through feature maps that abstract their similarities. Individual maps W 2,1, …, W 2,10 are Web documents pertaining to Global, Local, Political, Business, Weather, Entertainment, Technology, Sports, Opinion, and Health news respectively.

JingTao Yao 15 Growing Hierarchical Self-Organizing Maps for Web Mining Web-based News Coverage Example Feature map W 2,10 with neurons linking to three children maps: W 3,10,1, W 3,10,2, W 3,10,3. Articles in W 2,10 relate to Health News. W 3,10,1 relates to Health Research Funding. W 3,10,2 relates to Health Outbreak Crises. W 3,10,3 relates to Doctor shortages. New Health-related articles are coming in rapidly relating to a recent international outbreak. Neurons are added to W 2,10 in the Health Outbreak Crises cluster, that point to the SOM W 3,10,2.

JingTao Yao 16 Growing Hierarchical Self-Organizing Maps for Web Mining Conclusion The Web mining model of growing hierarchical self- organizing maps minimizes the effect of the dynamic data and high-dimensionality problems. Bidirection Update Propagation allows for changes in pattern abstractions to be reflect on multiple levels in the hierarchy. The Web-based News Coverage example demonstrates the effectiveness of growing hierarchical self-organizing maps when used in conjunction with bidirectional update propagation.

JingTao Yao Growing Hierarchical Self-Organizing Maps for Web Mining Joseph P. Herbert and JingTao Yao Department of Computer Science, University or Regina CANADA S4S 0A2