Connectionist Modelling Summer School Lecture Three.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Perceptron Lecture 4.
Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Classification Neural Networks 1
Perceptron.
Simple Neural Nets For Pattern Classification
The back-propagation training algorithm
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
Before we start ADALINE
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
September 28, 2010Neural Networks Lecture 7: Perceptron Modifications 1 Adaline Schematic Adjust weights i1i1i1i1 i2i2i2i2 inininin …  w 0 + w 1 i 1 +
Neural Networks Lecture 8: Two simple learning algorithms
Artificial neural networks:
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Where We’re At Three learning rules  Hebbian learning regression  LMS (delta rule) regression  Perceptron classification.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Artificial Neural Networks
Waqas Haider Khan Bangyal. Multi-Layer Perceptron (MLP)
Brief Overview of Connectionism to understand Learning Walter Schneider P2476 Cognitive Neuroscience of Human Learning & Instruction
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Modelling Language Evolution Lecture 1: Introduction to Learning Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
Feed-Forward Neural Networks 主講人 : 虞台文. Content Introduction Single-Layer Perceptron Networks Learning Rules for Single-Layer Perceptron Networks – Perceptron.
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
Linear Classification with Perceptrons
ADALINE (ADAptive LInear NEuron) Network and
Perceptrons Gary Cottrell. Cognitive Science Summer School 2 Perceptrons: A bit of history Frank Rosenblatt studied a simple version of a neural net called.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Chapter 2 Single Layer Feedforward Networks
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Neural Networks 2nd Edition Simon Haykin
Previous Lecture Perceptron W  t+1  W  t  t  d(t) - sign (w(t)  x)] x Adaline W  t+1  W  t  t  d(t) - f(w(t)  x)] f’ x Gradient.
Chapter 6 Neural Network.
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
1 Technological Educational Institute Of Crete Department Of Applied Informatics and Multimedia Intelligent Systems Laboratory.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Connectionist Modelling Summer School Lecture Two.
Announcements 1. Textbook will be on reserve at library 2. Topic schedule change; modified reading assignment: This week: Linear discrimination, evaluating.
Lecture 2 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 2/1 Dr.-Ing. Erwin Sitompul President University
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Neural networks.
Fall 2004 Backpropagation CS478 - Machine Learning.
Chapter 2 Single Layer Feedforward Networks
Artificial neural networks:
Real Neurons Cell structures Cell body Dendrites Axon
Ranga Rodrigo February 8, 2014
CSE 473 Introduction to Artificial Intelligence Neural Networks
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CSE P573 Applications of Artificial Intelligence Neural Networks
Simple learning in connectionist networks
Disadvantages of Discrete Neurons
Biological and Artificial Neuron
Classification Neural Networks 1
Biological and Artificial Neuron
of the Artificial Neural Networks.
Biological and Artificial Neuron
Neural Networks Chapter 5
Capabilities of Threshold Neurons
Back Propagation and Representation in PDP Networks
Neuro-Computing Lecture 2 Single-Layer Perceptrons
Simple learning in connectionist networks
Presentation transcript:

Connectionist Modelling Summer School Lecture Three

Using an Error Signal Orthogonality Constraint Number of patterns limited by dimensionality of network. Input patterns must be orthogonal to each other Similarity effects. Perceptron Convergence Rule Learning in a single weight network Assume a teacher signal t out Adaptation of Connection and Threshold (Rosenblatt 1958) Note that threshold always changes if incorrect output. Blame is apportioned to a connection in proportion to the activity of the input line. x y z Input Neurons Output Neurons w a in a out

Using an Error Signal Perceptron Convergence Rule “ The perceptron convergence rule guarantees to find a solution to a mapping problem, provided a solution exists.” (Minsky & Papert 1969 ) An Example of Perceptron Learning Boolean Or Training the network InputOutput a out w 20 w 21 InOutW 20 W 21 θ a out δ ΔθΔθ ΔwΔw

Gradient Descent Least Mean Square Error (LMS) Define the error measure as the square of the discrepancy between the actual output and the desired output. (Widrow-Hoff 1960) Plot an error curve for a single weight network Make weight adjustments by performing gradient descent – always move down the slope. Calculating the Error Signal Note that Perceptron Convergence and LMS use similar learning algorithms – the Delta Rule Error Landscapes Gradient descent algorithms adapt by moving downhill in a multi- dimensional landscape – the error surface. Ball bearing analogy. In a smooth landscape, the bottom will always be reached. However, bottom may not correspond to zero error. Weight Value Error

Past Tense Revisited Vocabulary Discontinuity –Up to 10 epochs – 8 irregulars + 2 regulars. Thereafter – 420 verbs – mostly regular. –Justification: Irregulars are more frequent than regulars Lack of Evidence –Vocabulary spurt at 2 years whereas overregularizations occur at 3 years. Furthermore, vocabulary spurt consists mostly of nouns. –Pinker and Prince (1988) show that regulars and irregulars are relatively balanced in early productive vocabularies

Longitudinal evidence Stages or phases in development? –Initial error-free performance. –Protracted period of overregularisation but at low rates (typically < 10%). –Gradual recovery from error. –Rate of overregularisation is much less the rate of regularisation of regular verbs. 1992

Longitudinal evidence Error Characteristics –High frequency irregulars are robust to overregularisation. –Some errors seem to be phonologically conditioned. –Irregularisations.

Single system account Multi-layered Perceptrons –Hidden unit representation –Error correction technique –Plunkett & Marchman 1991 –Type/Token distinction –Continuous training set

Single system account Incremental Vocabularies –Plunkett & Marchman (1993) –Initial small training set –Gradual expansion Overregularisation –Initial error-free performance. –Protracted period of overregularisation but at low rates (typically < 5%). –High frequency irregulars are robust to overregularisation.