ANNs (Artificial Neural Networks). THE PERCEPTRON.

Slides:



Advertisements
Similar presentations
Perceptron Lecture 4.
Advertisements

Slides from: Doug Gray, David Poole
1 Image Classification MSc Image Processing Assignment March 2003.
Artificial Neural Networks (1)
Perceptron Learning Rule
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.

Decision Support Systems
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation.
The Perceptron CS/CMPE 333 – Neural Networks. CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS2 The Perceptron – Basics Simplest and one.
20.5 Nerual Networks Thanks: Professors Frank Hoffmann and Jiawei Han, and Russell and Norvig.
Chapter 6: Multilayer Neural Networks
Artificial Neural Networks
Artificial Neural Networks
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial neural networks:
1 Introduction to Artificial Neural Networks Andrew L. Nelson Visiting Research Faculty University of South Florida.
Artificial Neural Networks
1 Mehran University of Engineering and Technology, Jamshoro Department of Electronic, Telecommunication and Bio-Medical Engineering Neural Networks Mukhtiar.
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Networks
Multi Layer NN and Bit-True Modeling of These Networks SILab presentation Ali Ahmadi September 2007.
Multi-Layer Perceptrons Michael J. Watts
Explorations in Neural Networks Tianhui Cai Period 3.
Machine Learning Chapter 4. Artificial Neural Networks
1 Chapter 6: Artificial Neural Networks Part 2 of 3 (Sections 6.4 – 6.6) Asst. Prof. Dr. Sukanya Pongsuparb Dr. Srisupa Palakvangsa Na Ayudhya Dr. Benjarath.
Appendix B: An Example of Back-propagation algorithm
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
NEURAL NETWORKS FOR DATA MINING
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Classification / Regression Neural Networks 2
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
An informal description of artificial neural networks John MacCormick.
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
Multi-Layer Perceptron
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
CS621 : Artificial Intelligence
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 32: sigmoid neuron; Feedforward.
COSC 4426 AJ Boulay Julia Johnson Artificial Neural Networks: Introduction to Soft Computing (Textbook)
NEURAL NETWORKS LECTURE 1 dr Zoran Ševarac FON, 2015.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Modelleerimine ja Juhtimine Tehisnärvivõrgudega Identification and Control with artificial neural networks.
Perceptrons Michael J. Watts
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Chapter 6 Neural Network.
Neural Networks Si Wu Dept. of Informatics PEV III 5c7 Spring 2008.
Neural Networks References: “Artificial Intelligence for Games” "Artificial Intelligence: A new Synthesis"
Artificial Neural Networks By: Steve Kidos. Outline Artificial Neural Networks: An Introduction Frank Rosenblatt’s Perceptron Multi-layer Perceptron Dot.
1 Neural Networks MUMT 611 Philippe Zaborowski April 2005.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Neural networks.
Modelleerimine ja Juhtimine Tehisnärvivõrgudega
CSE 473 Introduction to Artificial Intelligence Neural Networks
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
CSE P573 Applications of Artificial Intelligence Neural Networks
CSC 578 Neural Networks and Deep Learning
XOR problem Input 2 Input 1
CSE 573 Introduction to Artificial Intelligence Neural Networks
Neural Networks Chapter 5
Artificial Neural Networks
Presentation transcript:

ANNs (Artificial Neural Networks)

THE PERCEPTRON

Perceptron X is a (column) vector of inputs. W is a (column) vector of weights.  or w 0 is the bias or threshold weight.

Perceptron Usage: 1.training/learning – local vs. global minima – supervised vs. unsupervised 2.feedforward (or testing or usage or application) – Indicate class i if output g(X)>0; not class i otherwise.

Perceptron

The bias or threshold can be represented as simply another weight (w 0 ) with a constant input of 1 (x 0 =1).

Perceptron This is the dot product of two vectors, W and X.

Activation functions 1.linear 2.threshold 3.sigmoid 4.tanh

Relationship between sigmoid and tanh

Perceptron training/learning 1.Initialize weights (including thresholds, w 0 ) to random numbers in [-0.5,…,+0.5] (uniformly distributes). 2.Select training input vector, X, and desired output, d. 3.Calculate actual output, y. 4.Learn. (Only perform this step when output is incorrect.) where  is in [0..1] and is the gain or learning rate.

“The proof of the perceptron learning theorem (Rosenblatt 1962) demonstrated that a perceptron could learn anything it could represent.” [3] So what can it represent? – Any problem that is linearly separable. – Are all problems linear separable?

Consider a binary function of n binary inputs. – “A neuron with n binary inputs can have 2 n different input patterns, consisting of ones and zeros. Because each input pattern can produce two different binary outputs, 1 and 0, there are 2 2 n different functions of n variables.” [3] – How many of these are separable? Not many! For n=6, = 1.8x10 19 but only 5,028,134 are linearly separable. – AND and OR are linearly separable but XOR is not!

How about adding additional layers? “Multilayer networks provide no increase in computational power over a single-layer network unless there is a nonlinear function between layers.” [3] – That’s because matrix multiplication is associative. (XW 1 )W 2 = X(W 1 W 2 )

“It is natural to ask if every decision can be implemented by such a three-layer network … The answer, due ultimately to Kolmogorov …, is “yes” – any continuous function from input to output can be implemented in a three-layer net, given sufficient number of hidden units, proper nonlinearities, and weights.” [1]

MULTILAYER NETWORKS

Note threshold nodes from “Usefulness of artificial neural networks to predict follow-up dietary protein intake in hemodialysis patients”

Backpropagation (learning in a multilayer network)

Sigmoid/logistic function (and its derivative)

References 1.R.O. Duda, P.E. Hart, D.G. Stork, “Pattern Classification,” John Wiley and Sons, S.K. Rogers, M. Kabrisky, “An introduction to biological and artificial neural networks for pattern recognition,” SPIE Optical Engineering Press, P.D. Wasserman, “Neural computing,” Van Nostrand Reinhold, 1989.