Template design only ©copyright 2008 Ohio UniversityMedia Production 740.597-2521 Spring Quarter  A hierarchical neural network structure for text learning.

Slides:



Advertisements
Similar presentations
Sparse Coding in Sparse Winner networks Janusz A. Starzyk 1, Yinyin Liu 1, David Vogel 2 1 School of Electrical Engineering & Computer Science Ohio University,
Advertisements

Improved Neural Network Based Language Modelling and Adaptation J. Park, X. Liu, M.J.F. Gales and P.C. Woodland 2010 INTERSPEECH Bang-Xuan Huang Department.
Analog Circuits for Self-organizing Neural Networks Based on Mutual Information Janusz Starzyk and Jing Liang School of Electrical Engineering and Computer.
B.Macukow 1 Lecture 3 Neural Networks. B.Macukow 2 Principles to which the nervous system works.
Un Supervised Learning & Self Organizing Maps. Un Supervised Competitive Learning In Hebbian networks, all neurons can fire at the same time Competitive.
[1].Edward G. Jones, Microcolumns in the Cerebral Cortex, Proc. of National Academy of Science of United States of America, vol. 97(10), 2000, pp
Brain-like design of sensory-motor programs for robots G. Palm, Uni-Ulm.
Hybrid Pipeline Structure for Self-Organizing Learning Array Yinyin Liu 1, Ding Mingwei 2, Janusz A. Starzyk 1, 1 School of Electrical Engineering & Computer.
Self-Organizing Hierarchical Neural Network
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Neural Networks: Concepts (Reading:
EE141 1 Design of Self-Organizing Learning Array for Intelligent Machines Janusz Starzyk School of Electrical Engineering and Computer Science Heidi Meeting.
1 OPTIMIZED INTERCONNECTIONS IN PROBABILISTIC SELF-ORGANIZING LEARNING Janusz Starzyk, Mingwei Ding, Haibo He School of EECS Ohio University, Athens, OH.
Un Supervised Learning & Self Organizing Maps Learning From Examples
COGNITIVE NEUROSCIENCE
Fourth International Symposium on Neural Networks (ISNN) June 3-7, 2007, Nanjing, China A Hierarchical Self-organizing Associative Memory for Machine Learning.
Future Hardware Realization of Self-Organizing Learning Array and Its Software Simulation Adviser: Dr. Janusz Starzyk Student: Tsun-Ho Liu Ohio University.
EE141 1 Broca’s area Pars opercularis Motor cortexSomatosensory cortex Sensory associative cortex Primary Auditory cortex Wernicke’s area Visual associative.
EE141 1 Design of Self-Organizing Learning Array for Intelligent Machines Janusz Starzyk School of Electrical Engineering and Computer Science Heidi Meeting.
Mental Development and Representation Building through Motivated Learning Janusz A. Starzyk, Ohio University, USA, Pawel Raif, Silesian University of Technology,
Chapter Seven The Network Approach: Mind as a Web.
Associative Learning in Hierarchical Self Organizing Learning Arrays Janusz A. Starzyk, Zhen Zhu, and Yue Li School of Electrical Engineering and Computer.
IT 691 Final Presentation Pace University Created by: Robert M Gust Mark Lee Samir Hessami Mark Lee Samir Hessami.
MACHINE LEARNING 12. Multilayer Perceptrons. Neural Networks Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Neural Networks Lab 5. What Is Neural Networks? Neural networks are composed of simple elements( Neurons) operating in parallel. Neural networks are composed.
Fourth International Symposium on Neural Networks (ISNN) June 3-7, 2007, Nanjing, China Online Dynamic Value System for Machine Learning Haibo He, Stevens.
Biologically Inspired Robotics Group,EPFL Associative memory using coupled non-linear oscillators Semester project Final Presentation Vlad TRIFA.
A Hybrid Self-Organizing Neural Gas Network James Graham and Janusz Starzyk School of EECS, Ohio University Stocker Center, Athens, OH USA IEEE World.
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
A Genetic Algorithms Approach to Feature Subset Selection Problem by Hasan Doğu TAŞKIRAN CS 550 – Machine Learning Workshop Department of Computer Engineering.
Lecture 12 Self-organizing maps of Kohonen RBF-networks
Chapter 14: Artificial Intelligence Invitation to Computer Science, C++ Version, Third Edition.
© N. Kasabov Foundations of Neural Networks, Fuzzy Systems, and Knowledge Engineering, MIT Press, 1996 INFO331 Machine learning. Neural networks. Supervised.
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
What is a neural network? Collection of interconnected neurons that compute and generate impulses. Components of a neural network include neurons, synapses,
Advances in Modeling Neocortex and its impact on machine intelligence Jeff Hawkins Numenta Inc. VS265 Neural Computation December 2, 2010 Documentation.
Neural Network with Memory and Cognitive Functions Janusz A. Starzyk, and Yue Li School of Electrical Engineering and Computer Science Ohio University,
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
NEURAL NETWORKS FOR DATA MINING
Hierarchical Temporal Memory as a Means for Image Recognition by Wesley Bruning CHEM/CSE 597D Final Project Presentation December 10, 2008.
KNOWLEDGE BASED TECHNIQUES INTRODUCTION many geographical problems are ill-structured an ill-structured problem "lacks a solution algorithm.
Self Organizing Feature Map CS570 인공지능 이대성 Computer Science KAIST.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 8: Neural Networks.
CS 478 – Tools for Machine Learning and Data Mining Perceptron.
Lecture 5 Neural Control
Why Can't A Computer Be More Like A Brain?. Outline Introduction Turning Test HTM ◦ A. Theory ◦ B. Applications & Limits Conclusion.
IE 585 History of Neural Networks & Introduction to Simple Learning Rules.
Efficient Pairwise Key Establishment Scheme Based on Random Pre-Distribution Keys in Wireless Sensor Networks Source: Lecture Notes in Computer Science,
“Principles of Soft Computing, 2 nd Edition” by S.N. Sivanandam & SN Deepa Copyright  2011 Wiley India Pvt. Ltd. All rights reserved. CHAPTER 2 ARTIFICIAL.
Presented By Dr. Paul Cottrell Company: Reykjavik.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
March 31, 2016Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms I 1 … let us move on to… Artificial Neural Networks.
Bayesian Brain - Chapter 11 Neural Models of Bayesian Belief Propagation Rajesh P.N. Rao Summary by B.-H. Kim Biointelligence Lab School of.
Symbolic Reasoning in Spiking Neurons: A Model of the Cortex/Basal Ganglia/Thalamus Loop Terrence C. Stewart Xuan Choo Chris Eliasmith Centre for Theoretical.
Machine Learning 12. Local Models.
Hybrid BDD and All-SAT Method for Model Checking
Multilayer Perceptrons
Fall 2004 Perceptron CS478 - Machine Learning.
Learning in Neural Networks
Artificial Intelligence (CS 370D)
Copyright © 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 2 Database System Concepts and Architecture.
Dynamic Routing Using Inter Capsule Routing Protocol Between Capsules
Machine Learning Today: Reading: Maria Florina Balcan
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
Lecture 22 Clustering (3).
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
The Network Approach: Mind as a Web
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Template design only ©copyright 2008 Ohio UniversityMedia Production Spring Quarter  A hierarchical neural network structure for text learning is obtained through self-organization  Similar representation for text based semantic network was used [2]  An input layer takes in characters, then learns and activates words stored in memory  Direct activation of words requires large computational cost for large dictionaries  Extension to phrases, sentences or paragraphs would render such a network impractical due to associated computational cost  Computer memory required would also be tremendously large  This leads to a sparse hierarchical structure.  The higher layers represent more complex concepts.  Basic nodes in this network are capable of differentiating input sequences.  Sequence learning is prerequisite to building spatio-temporal memories.  This is performed using laminar minicolumn [3] LTM cells (Fig.1)  In such networks, the interconnection scheme is naturally obtained through sequence learning and structural self-organization.  No prior assumption about locality of connections or structure sparsity is made.  Machine learns only inputs useful to its objectives, a process that is regulated by reinforcement signals and self organization. Hierarchical Neural Network for Text Based Learning Janusz A Starzyk, Basawaraj Ohio University, Athens, OH Introduction References Hierarchical Network  Traditional approach is to describe semantic network structure and/or probabilities of transition in associated Markov models  Biological networks learn  Different Neural Network structures, but common goal  Simple and efficient to solve the given problem  Sparsity is essential  Size of the network and time to train important for large data sets  Hierarchical structure of identical processing units was proposed [1]  Layered organization and sparse structure is biologically inspired  Neurons on different layers interact through trained links  Mountcastle, V. B., et. al, Response Properties of Neurons of Cat’s Somatic Sensory Cortex to Peripheral Stimuli, J. Neurophysiology, vol. 20, 1957, pp  Rogers, T. T., McClelland, J. L., Semantic Cognition text: A parallel Distributed Processing Approach, 2004, MIT Press.  Grossberg, S., How does the cerebral cortex work? Learning, attention and grouping by the laminar circuits of visual cortex. Spatial Vision, 12, ,1999.  Starzyk J.A., Liu Y., Hierarchical spatio-temporal memory for machine learning based on laminar minicolumn structure, 11 th ICCNS, Boston, Network Simplification  Proposed approach uses intermediate neurons to lower the computational cost  Intermediate neurons decrease number of activations associated with higher level neurons  This concept can be extended to associations of words  Small number of rules for concurrent processing are used  We can arrive at local optimum of network structure / performance  The network topology is self-organizing through addition and removal of neurons and redirecting of neuron connections  Neurons are described by their sets of input and output neurons  Local optimization criteria are checked by searching the set SL A before the structure is updated when creating or merging the neurons.  where IL X is the input list of neuron X and  OL A is the output list of neuron A Fig. 3 If create a new node C. B A B Fig. 4 Neuron “A” with single output is merged with Neuron “B”, and “A” is removed. B A A Fig. 5 Neuron “B” with single input is merged with Neuron “A”, and “B” is removed. Batch Mode:  All words used for training are available at initiation.  Network simplification & optimization is done by processing - all the words in the training set.  Total number of neurons is 23% higher than the reference (6000) Dynamic Mode:  Words used for training are increased incrementally, - one word at a time  Simplification & optimization is done by processing - one word at a time.  Total number of neurons is 68% higher than the reference (6000) Implementation Results and Conclusion  Tests were run with dictionary up to 6000 words  The percent reduction in number of interconnections increases (by up to 65 – 70%) as the number of words increase.  The time required to process network activation for all the words used decreases as the number of words increases (reduction by a factor of 55, in batch mode; and 35, in dynamic mode; for 6000 words).  Dynamic implementation takes longer compared to the batch implementation, mainly due to the additional overhead required for bookkeeping.  The savings (connections and activations) obtained in case of dynamic implementation are less compared to the batch implementation  Combination of both methods is advisable for continuous learning and self-organization. A B X4X4 X1X1 X2X2 X3X3 AB X4X4 C X1X1 X2X2 X3X3 Rules for Self-Organization Fig. 2 If create a new node C. X1X1 B X3X3 X4X4 X2X2 A B A C X4X4 X3X3 X2X2 X1X1  Few simple rules used for self organization  Rules Contd.