Chapter 9 Neural Network.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Slides from: Doug Gray, David Poole
1 Data Mining: and Knowledge Acquizition — Chapter 5 — BIS /2014 Summer.
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Kostas Kontogiannis E&CE
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Perceptron.
Machine Learning Neural Networks
Lecture 14 – Neural Networks
Decision Support Systems
Neural Networks Basic concepts ArchitectureOperation.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Neural Networks. R & G Chapter Feed-Forward Neural Networks otherwise known as The Multi-layer Perceptron or The Back-Propagation Neural Network.
Neural Networks Chapter Feed-Forward Neural Networks.
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
Before we start ADALINE
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Data Mining with Neural Networks (HK: Chapter 7.5)
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Gini Index (IBM IntelligentMiner)
CS 484 – Artificial Intelligence
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial Neural Network
Chapter 5 Data mining : A Closer Look.
Introduction to Directed Data Mining: Neural Networks
BEE4333 Intelligent Control
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Introduction to Neural Networks Debrup Chakraborty Pattern Recognition and Machine Learning 2006.
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
NEURAL NETWORKS FOR DATA MINING
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Chapter 8 The k-Means Algorithm and Genetic Algorithm.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Artificial Neural Networks An Introduction. What is a Neural Network? A human Brain A porpoise brain The brain in a living creature A computer program.
Multi-Layer Perceptron
Akram Bitar and Larry Manevitz Department of Computer Science
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Chapter 8: Adaptive Networks
Hazırlayan NEURAL NETWORKS Backpropagation Network PROF. DR. YUSUF OYSAL.
Neural Networks 2nd Edition Simon Haykin
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Previous Lecture Perceptron W  t+1  W  t  t  d(t) - sign (w(t)  x)] x Adaline W  t+1  W  t  t  d(t) - f(w(t)  x)] f’ x Gradient.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks Lecture 11: Learning in recurrent networks Geoffrey Hinton.
“Principles of Soft Computing, 2 nd Edition” by S.N. Sivanandam & SN Deepa Copyright  2011 Wiley India Pvt. Ltd. All rights reserved. CHAPTER 2 ARTIFICIAL.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
DATA MINING TECHNIQUES (DECISION TREES ) Presented by: Shweta Ghate MIT College OF Engineering.
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
A Document-Level Sentiment Analysis Approach Using Artificial Neural Network and Sentiment Lexicons Yan Zhu.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Neural networks.
Data Mining, Neural Network and Genetic Programming
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Neural Networks Advantages Criticism
Presentation transcript:

Chapter 9 Neural Network

Chapter Objective Understand how feed-forward networks are used to solve estimation problems. Know how input and output data conversions are performed for neural networks. Understand how feed-forward neural networks learn through backpropagation. Know how genetic learning is applied to train feed-forward neural networks. Know how self-organizing neural networks perform unsupervised clustering. List the strengths and weaknesses of neural networks. Chapter 9

Feed-Forward Neural Network Chapter 9

Feed-Forward Neural Network Chapter 9

Neural Network Training: A Conceptual View Chapter 9

Neural Network Training: A Conceptual View Chapter 9

Neural Network Training: A Conceptual View Chapter 9

Neural Network Explanation Sensitivity analysis is a technique that has been successfully applied to gain insight into the effect individual attributes have on neural network output. The general process consists of the following steps: 1. Divide the data into a training set and a test dataset. 2. Train the network with the training data. Chapter 9

Neural Network Explanation 3. Use the test set data to create a new instance I. Each attribute value for I is the average of all attribute values within the test data. 4. For each attribute: a. Vary the attribute value within instance I and present the modification of I to the network for classification. b. Determine the effect the variations have on the output of the neural network. c. The relative importance of each attribute is measured by the effect of attribute variations on network output. Chapter 9

General Considerations The following is a partial list of choices that affect the performance of a neural network model: What input attributes will be used to build the network? How will the network output be represented? How many hidden layers should the network contain? How many nodes should there be in each hidden layer? What condition will terminate network training? Chapter 9

Neural Network Training: A Detailed View Chapter 9

Neural Networks Advantages Criticism prediction accuracy is generally high robust, works when training examples contain errors output may be discrete, real-valued, or a vector of several discrete or real-valued attributes fast evaluation of the learned target function Criticism long training time difficult to understand the learned function (weights) not easy to incorporate domain knowledge Chapter 9

- f A Neuron mk å weighted sum Input vector x output y Activation function weight vector w å w0 w1 wn x0 x1 xn The n-dimensional input vector x is mapped into variable y by means of the scalar product and a nonlinear function mapping Chapter 9

Network Training The ultimate objective of training Steps obtain a set of weights that makes almost all the tuples in the training data classified correctly Steps Initialize weights with random values Feed the input tuples into the network..... one by one For each unit Compute the net input to the unit as a linear combination of all the inputs to the unit Compute the output value using the activation function Compute the error Update the weights and the bias Chapter 9

Multi-Layer Perceptron Output vector Output nodes Hidden nodes wij Input nodes Input vector: xi Chapter 9

Network Pruning and Rule Extraction Fully connected network will be hard to articulate N input nodes, h hidden nodes and m output nodes lead to h(m+N) weights Pruning: Remove some of the links without affecting classification accuracy of the network Chapter 9

Network Pruning and Rule Extraction Extracting rules from a trained network Discretize activation values; replace individual activation value by the cluster average maintaining the network accuracy Enumerate the output from the discretized activation values to find rules between activation value and output Find the relationship between the input and activation value Combine the above two to have rules relating the output to input Chapter 9

Chapter Summary A neural network is parallel computing system of several interconnected processor nodes. The input to individual network nodes is restricted to numeric values falling in the closed interval range [0,1]. Because of this, categorical data must be transformed prior to network training. Chapter 9

Chapter Summary Developing a neural network involves first training the network to carry out the desired computations and then applying the trained network to solve new problems. During the learning phase, training data is used to modify the connection weights between pairs of nodes so as to obtain a best result for the output node (s). The feed-forward neural network architecture is commonly used for supervised learning. Feed-forward neural networks contain a set of layered nodes and weighted connections between nodes in adjacent layers. Chapter 9

Chapter Summary Feed-forward neural networks are often trained using a backpropagation learning scheme. Backpropagation learning works by making modifications in weight values starting at the output layer then moving backward through the hidden layers of the network. Genetic learning can also be applied to train feed-forward networks. Chapter 9

Chapter Summary The self-organizing Kohonen neural network architecture is a popular model for unsupervised clustering. A self-organizing neural network learns by having several output nodes complete for the training instances. For each instance, the output node whose weight vectors most closely match the attribute values of the input instance is the winning node. Chapter 9

Chapter Summary As a result, the winning node has its associated input weights modified to more closely match the current training instance. When unsupervised learning is complete, output nodes winning the most instances are saved. After this, test data is applied and the clusters formed by the test set data are analyzed to help determine the meaning of what has been found. Chapter 9

Chapter Summary A central issue surrounding neural network is their inability to explain what has been learned. Despite this, neural network have been successfully applied to solve problems in both the business and scientific world. Although we have discussed the most popular neural network models, several other architectures and learning rules have been developed. Jain, Mao, and Mohiuddin (1996) provide a good starting point for learning more about neural networks. Chapter 9

Key Terms Average member technique. An unsupervised clustering neural network explanation technique where the most typical member of each cluster is computed by finding the average value for each class attribute. Backpropagation learning. A training method used with many feed-forward networks that works by making modifications in weight values starting at the output layer then moving backward through the hidden layer. Delta rule. A neural network learning rule designed to minimize the sum of squared errors between computed and target network output. Chapter 9

Key Terms Epoch. One complete pass of the training data through a neural network. Feed-forward neural network. A neural network architecture where all weights at one layer are directed toward nodes at the next network layer. Weights do not cycle back as inputs to previous layers. Fully connected. A neural network structure where all nodes at one layer of the network are connected to all nodes in the next layer. Kohonen network. A two-layer neural network used for unsupervised clustering. Chapter 9

Key Terms Neural network. A parallel computing system consisting of several interconnected processors. Neurode. A neural network processor node. Several neurodes are connected to form a complete neural network structure. Sensitivity analysis. A neural network explanation technique that allows us to determine a rank ordering for the relative importance of individual attributes. Sigmoid function. One of several commonly used neural network evaluation functions. The sigmoid function is continuous and outputs a value between 0 or 1. Chapter 9

Key Terms Linearly separable. Two classes, A and B, are said to be linearly separable if a straight line can be drawn to separate the instances of class B. Perceptron neural network. A simple feed-forward neural network architecture consisting of an input layer and a single output layer. Chapter 9

Reference Data Mining: Concepts and Techniques (Chapter 7 Slide for textbook), Jiawei Han and Micheline Kamber, Intelligent Database Systems Research Lab, School of Computing Science, Simon Fraser University, Canada Chapter 9