Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
CS 678 –Boltzmann Machines1 Boltzmann Machine Relaxation net with visible and hidden units Learning algorithm Avoids local minima (and speeds up learning)
Kostas Kontogiannis E&CE
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Classification Neural Networks 1
CS 678 –Relaxation and Hopfield Networks1 Relaxation and Hopfield Networks Totally connected recurrent relaxation networks Bidirectional weights (symmetric)
Machine Learning Neural Networks
Neural Networks Basic concepts ArchitectureOperation.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Rutgers CS440, Fall 2003 Neural networks Reading: Ch. 20, Sec. 5, AIMA 2 nd Ed.
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
Artificial Neural Networks
CS 4700: Foundations of Artificial Intelligence
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Neural Networks. Plan Perceptron  Linear discriminant Associative memories  Hopfield networks  Chaotic networks Multilayer perceptron  Backpropagation.
Artificial Neural Networks
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Computer Science and Engineering
Artificial Neural Networks
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Multi-Layer Perceptrons Michael J. Watts
Machine Learning Chapter 4. Artificial Neural Networks
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
1 Chapter 6: Artificial Neural Networks Part 2 of 3 (Sections 6.4 – 6.6) Asst. Prof. Dr. Sukanya Pongsuparb Dr. Srisupa Palakvangsa Na Ayudhya Dr. Benjarath.
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
NEURAL NETWORKS FOR DATA MINING
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Classification / Regression Neural Networks 2
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
ADVANCED PERCEPTRON LEARNING David Kauchak CS 451 – Fall 2013.
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks - lecture 51 Multi-layer neural networks  Motivation  Choosing the architecture  Functioning. FORWARD algorithm  Neural networks as.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Chapter 18 Connectionist Models
Chapter 6 Neural Network.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Machine Learning Supervised Learning Classification and Regression
Fall 2004 Backpropagation CS478 - Machine Learning.
Learning with Perceptrons and Neural Networks
Learning in Neural Networks
Real Neurons Cell structures Cell body Dendrites Axon
CSSE463: Image Recognition Day 17
CSC 578 Neural Networks and Deep Learning
Classification Neural Networks 1
network of simple neuron-like computing elements
Artificial Neural Networks
Backpropagation.
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 18
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 17
David Kauchak CS158 – Spring 2019
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Neural Networks Ellen Walker Hiram College

Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing elements –Large number of weighted connections between these elements –Highly parallel, distributed control –Emphasis on automatic learning of internal representations (weights)

Basic Connectionist Unit (Perceptron)

Classes of Connectionist Architectures Constraint networks –Positive and negative connections denote constraints between the values of nodes –Weights set by programmer Layered networks –Weights represent contribution from one intermediate value to the next –Weights are learned using feedback

Hopfield Network A constraint network Every node is connected to every other node –If the weight is 0, the connection doesn’t matter To use the network, set the values of the nodes and let the nodes adjust their values according to the weights. The “result” is the set of all values in the stabilized network.

Hopfield Network as CAM Nodes represent features of objects Compatible features support each other (weights > 0) Stable states (local minima) are “valid” interpretations Noise features (incompatible) will be suppressed (network will fall into nearest stable state)

Hopfield Net Example

Relaxation Algorithm to find stable state for Hopfield network (serial or parallel) –Pick a node –Compute [incoming weights]*[neighbors] –If above sum > 0, node =1, else node = -1 When values aren’t changing, network is stable Result can depend on order of nodes chosen

Line Labeling and Relaxation Given an object, each vertex contrains the labels of its connected lines

Hopfield Network for Labeling Each gray box contains 4 mutually exclusive nodes (with negative links between them) Lines denote positive links between compatible labels

Boltzmann Machine Alternative training method for a Hopfield network, based on simulated annealing Goal: to find the most stable state (rather than the nearest) Boltzmann rule is probabilistic, based on the “temperature” of the system

Deterministic vs. Boltzman Deterministic update rule Probabilistic update rule –As temperature decreases, probabilistic rule approaches deterministic one

Networks and Function Fitting We earlier talked about function fitting –Finding a function that approximates a set of data so that Function fits the data well Function generalized to fit additional data

What Can a Neuron Compute? n inputs (i 0 =1, i 1 …in) n+1 weights (w 0 …w n ) 1 output: –1 if g(i) > 0 –0 if g(i) < 0 –g(i) = G denotes a linear surface, and the output is 1 if the point is above this surface

Classification by a Neuron

Training a Neuron 1.Initialize weights randomly 2.Collect all misclassified examples 3.If there are none, we’re done. 4.Else compute gradient & update weights –Add all points that should have fired, subtract all points that should not have fired –Add a constant (0<C<1) * gradient back to the weights. 5.Repeat steps 2-5 until done (Guaranteed to converge -- loop will end)

Training Example

Perceptron Problem We have a model and a training algorithm, but we can only compute linearly separable functions! Most interesting functions are not linearly separable. Solution: use more than one line (multiple perceptrons)

Multilayered Network Layered, fully-connected (between layers), feed-forward input hidden output

Backpropagation Training Compute a result: –input->hidden->output Compute error for each hidden node, based on desired result Propagate errors back: –Output->hidden, hidden->input –Weights are adjusted using gradient

Backpropagation Training (cont’d) Repeat above for every example in the training set (one epoch) Repeat above until stopping criterion is reached –Good enough average performance on training set –Little enough change in network Hundreds of epochs…

Generalization If the network is trained correctly, results will generalize to unseen data If overtrained, network will “memorize” training data, random outputs otherwise Tricks to avoid memorization –Limit number of hidden nodes –Insert noise into training data

Unsupervised Network Learning Kohonen network for classification

Training Kohonen Network Create inhibitory links among nodes of output layer (“winner take all”) For each item in training data: –Determine an input vector –Run network - find max output node –Reinforce (increase) weights to maximum node –Normalize weights so they sum to 1

Representations in Networks Distributed representation –Concept = pattern –Examples: Hopfield, backpropagation Localist representation –Concept = single node –Example: Kohonen Distributed can be more robust, also more efficient