1 Machine Learning: Connectionist 11 11.0Introduction 11.1Foundations of Connectionist Networks 11.2Perceptron Learning 11.3Backpropagation Learning 11.4.

Slides:



Advertisements
Similar presentations
Artificial Intelligence 12. Two Layer ANNs
Advertisements

1 Machine Learning: Connectionist Introduction 11.1Foundations of Connectionist Networks 11.2Perceptron Learning 11.3Backpropagation Learning 11.4.
Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Introduction to Training and Learning in Neural Networks n CS/PY 399 Lab Presentation # 4 n February 1, 2001 n Mount Union College.
PDP: Motivation, basic approach. Cognitive psychology or “How the Mind Works”
Biological and Artificial Neurons Michael J. Watts
Artificial Neural Networks - Introduction -
Artificial Intelligence (CS 461D)
Simple Neural Nets For Pattern Classification
PERCEPTRON. Chapter 3: The Basic Neuron  The structure of the brain can be viewed as a highly interconnected network of relatively simple processing.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Rutgers CS440, Fall 2003 Neural networks Reading: Ch. 20, Sec. 5, AIMA 2 nd Ed.
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
1 Chapter 20 Section Slide Set 1 Prologue to neural networks: the neurons in the human brain Additional sources used in preparing the slides: Various.
How does the mind process all the information it receives?
Data Mining with Neural Networks (HK: Chapter 7.5)
CS 4700: Foundations of Artificial Intelligence
Artificial Neural Networks - Introduction -
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Traffic Sign Recognition Using Artificial Neural Network Radi Bekker
BEE4333 Intelligent Control
Artificial neural networks:
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
Artificial Intelligence Neural Networks ( Chapter 9 )
Introduction to Artificial Neural Network Models Angshuman Saha Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg.
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Presented by Scott Lichtor An Introduction to Neural Networks.
1 Chapter 20 Section Slide Set 2 Perceptron examples Additional sources used in preparing the slides: Nils J. Nilsson’s book: Artificial Intelligence:
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
George F Luger ARTIFICIAL INTELLIGENCE 6th edition Structures and Strategies for Complex Problem Solving Machine Learning: Connectionist Luger: Artificial.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
CS344 : Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 29 Introducing Neural Nets.
Introduction to Artificial Intelligence (G51IAI) Dr Rong Qu Neural Networks.
ADVANCED PERCEPTRON LEARNING David Kauchak CS 451 – Fall 2013.
CS 478 – Tools for Machine Learning and Data Mining Perceptron.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Artificial Neural Networks Chapter 4 Perceptron Gradient Descent Multilayer Networks Backpropagation Algorithm 1.
Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
November 21, 2013Computer Vision Lecture 14: Object Recognition II 1 Statistical Pattern Recognition The formal description consists of relevant numerical.
COMP53311 Other Classification Models: Neural Network Prepared by Raymond Wong Some of the notes about Neural Network are borrowed from LW Chan’s notes.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CSC321: Neural Networks Lecture 1: What are neural networks? Geoffrey Hinton
Where are we? What’s left? HW 7 due on Wednesday Finish learning this week. Exam #4 next Monday Final Exam is a take-home handed out next Friday in class.
1 Azhari, Dr Computer Science UGM. Human brain is a densely interconnected network of approximately neurons, each connected to, on average, 10 4.
Computational Intelligence Semester 2 Neural Networks Lecture 2 out of 4.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Learning with Perceptrons and Neural Networks
Artificial Intelligence (CS 370D)
Artificial neural networks:
Real Neurons Cell structures Cell body Dendrites Axon
Machine Learning: Connectionist
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Artificial Intelligence Lecture No. 28
Artificial Intelligence 12. Two Layer ANNs
David Kauchak CS158 – Spring 2019
EE 193/Comp 150 Computing with Biological Parts
Presentation transcript:

1 Machine Learning: Connectionist Introduction 11.1Foundations of Connectionist Networks 11.2Perceptron Learning 11.3Backpropagation Learning 11.4 Competitive Learning 11.5Hebbian Coincidence Learning 11.6Attractor Networks or “Memories” 11.7Epilogue and References 11.8Exercises Additional sources used in preparing the slides: Robert Wilensky’s AI lecture notes, Various sites that explain how a neuron works

2 Chapter Objectives Learn about the neurons in the human brain Learn about single neuron systems Introduce neural networks

3 Inspiration: The human brain We seem to learn facts and get better at doing things without having to run a separate “learning procedure.” It is desirable to integrate learning more with doing.

4 Biology The brain doesn’t seem to have a CPU. Instead, it’s got lots of simple, parallel, asynchronous units, called neurons. Every neuron is a single cell that has a number of relatively short fibers, called dendrites, and one long fiber, called an axon.  The end of the axon branches out into more short fibers  Each fiber “connects” to the dendrites and cell bodies of other neurons  The “connection” is actually a short gap, called a synapse  Axons are transmitters, dendrites are receivers

5 Neuron

6 How neurons work The fibers of surrounding neurons emit chemicals (neurotransmitters) that move across the synapse and change the electrical potential of the cell body  Sometimes the action across the synapse increases the potential, and sometimes it decreases it.  If the potential reaches a certain threshold, an electrical pulse, or action potential, will travel down the axon, eventually reaching all the branches, causing them to release their neurotransmitters. And so on...

7 How neurons work (cont’d)

8 How neurons change There are changes to neurons that are presumed to reflect or enable learning:  The synaptic connections exhibit plasticity. In other words, the degree to which a neuron will react to a stimulus across a particular synapse is subject to long- term change over time (long-term potentiation).  Neurons also will create new connections to other neurons.  Other changes in structure also seem to occur, some less well understood than others.

9 Neurons as devices How many neurons are there in the human brain? - around (with, perhaps, or so synapses) Neurons are slow devices. - Tens of milliseconds to do something. - Feldman translates this into the “100 step program constraint: Most of the AI tasks we want to do take people less than a second. So any brain “program” can’t be longer than 100 neural “instructions.” No particular unit seems to be important. - Destroying any one brain cell has little effect on overall processing.

10 How do neurons do it? Basically, all the billions of neurons in the brain are active at once. - So, this is truly massive parallelism. But, probably not the kind of parallelism that we are used to in conventional Computer Science. - Sending messages (i.e., patterns that encode information) is probably too slow to work. - So information is probably encoded some other way, e.g., by the connections themselves.

11 AI / Cognitive Science Implication Explain cognition by richly connected networks transmitting simple signals. Sometimes called - connectionist computing (by Jerry Feldman) - Parallel Distributed Processing (PDP) (by Rumelhart, McClelland, and Hinton) - neural networks (NN) - artificial neural networks (ANN) (emphasizing that the relation to biology is generally rather tenuous)

12 From a neuron to a perceptron All connectionist models use a similar model of a neuron There is a collection of units each of which has  a number of weighted inputs from other units  inputs represent the degree to which the other unit is firing  weights represent how much the units wants to listen to other units  a threshold that the sum of the weighted inputs are compared against  the threshold has to be crossed for the unit to do something (“fire”)  a single output to another bunch of units  what the unit decided to do, given all the inputs and its threshold

13 A unit (perceptron) x i are inputs w i are weights w n is usually set for the threshold with x n =1 (bias) y is the weighted sum of inputs including the threshold (activation level) o is the output. The output is computed using a function that determines how far the perceptron’s activation level is below or above 0 x1x1 x2x2 x3x3 xnxn w1w1 w2w2 w3w3 wnwn y=  w i x i O=f(y)

14 Notes The perceptrons are continuously active - Actually, real neurons fire all the time; what changes is the rate of firing, from a few to a few hundred impulses a second The weights of the perceptrons are not fixed - Indeed, learning in a NN system is basically a matter of changing weights

15 Interesting questions for NNs How do we wire up a network of perceptrons? - i.e., what “architecture” do we use? How does the network represent knowledge? - i.e., what do the nodes mean? How do we set the weights? - i.e., how does learning take place?

16 The simplest architecture: a single perceptron A perceptron computes o = sign (X. W), where X.W = w 1 * x 1 + w 2 * x 2 + … + w n * 1, and sign(x) = 1 if x > 0 and -1 otherwise A perceptron can act as a logic gate interpreting 1 as true and -1 (or 0) as false x1x1 x2x2 x3x3 xnxn w2w2 w3w3 wnwn y=  w i x i o w1w1

17 Logical function and x + y - 2 x x  y y

18 Logical function or x + y - 1 x x  y y 1 +1

19 Training perceptrons We can train perceptrons to compute the function of our choice The procedure  Start with a perceptron with any values for the weights (usually 0)  Feed the input, let the perceptron compute the answer  If the answer is right, do nothing  If the answer is wrong, then modify the weights by adding or subtracting the input vector (perhaps scaled down)  Iterate over all the input vectors, repeating as necessary, until the perceptron learns what we want

20 Training perceptrons: the intuition If the unit should have gone on, but didn’t, increase the influence of the inputs that are on: - adding the input (or fraction thereof) to the weights will do so; If it should have been off, but was on, decrease influence of the units that were on: - subtracting the input from the weights does this

21 Example: teaching the logical or function Want to learn this: Initially the weights are all 0, i.e., the weight vector is (0 0 0) The next step is to cycle through the inputs and change the weights as necessary

22 The training cycle InputWeightsResultAction 1. ( )(0 0 0)f(0) = -1correct, do nothing 2. (1 -1 1)(0 0 0)f(0) = -1should have been 1, so add inputs to weights (1 -1 1)(0 0 0) + (1 -1 1) = (1 -1 1) 3. (1 1 -1) (1 -1 1)f(-1) = -1 should have been 1, so add inputs to weights (2 0 0)(1 -1 1) + (1 1 -1) = (2 0 0) 4. (1 1 1)(2 0 0) f(1) = 1 correct, but keep going! 1. ( )(2 0 0)f(2) = 1 should be have been -1, so subtract inputs from weights (1 1 1)(2 0 0) - ( ) = (1 1 1) These do the trick!

23 The final set of weights The learned set of weights does the right thing for all the data: ( ). ( 1 1 1) = -1  f(-1) = -1 (1 -1 1). (1 1 1) = 1  f(1) = 1 (1 1 -1). (1 1 1) = 1  f(1) = 1 (1 1 1). (1 1 1) = 3  f(3) = 1

24 The general procedure Start with a perceptron with any values for the weights (usually 0) Feed the input, let the perceptron compute the answer If the answer is right, do nothing If the answer is wrong, then modify the weights by adding or subtracting the input vector  w i = c (d - f) x i Iterate over all the input vectors, repeating as necessary, until the perceptron learns what we want (i.e., the weight vector converges)

25 More on  w i = c (d - f) x i c is the learning constant d is the desired output f is the actual output (d - f ) is either 0 (correct), or (1 - (-1))= 2, or (-1 - 1) = -2. The net effect is: When the actual output is -1 and should be 1, increment the weights on the ith line by 2cx i. When the actual output is 1 and should be -1, decrement the weights on the ith line by 2cx i.

26 A data set for perceptron classification

27 A two-dimensional plot of the data points

28 The good news The weight vector converges to ( ) after 500 iterations. The equation of the line found is -1.3 * x * x = 0 I had different weight vectors in iterations

29 The bad news: the exclusive-or problem No straight line in two-dimensions can separate the (0, 1) and (1, 0) data points from (0, 0) and (1, 1). A single perceptron can only learn linearly separable data sets.

30 The solution: multi-layered NNs

31 The adjustment for w ki depends on the total contribution of node i to the error at the output

32 Comments on neural networks Parallelism in AI is not new. - spreading activation, etc. Neural models for AI is not new. - Indeed, is as old as AI, some subdisciplines such as computer vision, have continuously thought this way. Much neural network works makes biologically implausible assumptions about how neurons work - backpropagation is biologically implausible. - “neurally inspired computing” rather than “brain science.”

33 Comments on neural networks (cont’d) None of the neural network models distinguish humans from dogs from dolphins from flatworms. - Whatever distinguishes “higher” cognitive capacities (language, reasoning) may not be apparent at this level of analysis. Relation between NN and “symbolic AI”? - Some claim NN models don’t have symbols and representations. - Others think of NNs as simply being an “implementation-level” theory. - NNs started out as a branch of statistical pattern classification, and is headed back that way.

34 Nevertheless NNs give us important insights into how to think about cognition NNs have been used in solving lots of problems  learning how to pronounce words from spelling (NETtalk, Sejnowski and Rosenberg, 1987)  Controlling kilns (Ciftci, 2001)