Biological and Artificial Neuron

Slides:



Advertisements
Similar presentations
Slides from: Doug Gray, David Poole
Advertisements

NEURAL NETWORKS Backpropagation Algorithm
Introduction to Neural Networks Computing
NEURAL NETWORKS Perceptron
Longin Jan Latecki Temple University
Machine Learning Neural Networks
Performance Optimization
Supervised learning 1.Early learning algorithms 2.First order gradient methods 3.Second order gradient methods.
1 Part I Artificial Neural Networks Sofia Nikitaki.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Spring 2002 Shreekanth Mandayam Robi Polikar ECE Department.
Goals of Adaptive Signal Processing Design algorithms that learn from training data Algorithms must have good properties: attain good solutions, simple.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Fall 2004 Shreekanth Mandayam ECE Department Rowan University.
Artificial Neural Networks
CS 4700: Foundations of Artificial Intelligence
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2010 Shreekanth Mandayam ECE Department Rowan University.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Artificial Neural Networks
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
DIGITAL IMAGE PROCESSING Dr J. Shanbehzadeh M. Hosseinajad ( J.Shanbehzadeh M. Hosseinajad)
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
Lecture 3 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 3/1 Dr.-Ing. Erwin Sitompul President University
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
Artificial Intelligence Chapter 3 Neural Networks Artificial Intelligence Chapter 3 Neural Networks Biointelligence Lab School of Computer Sci. & Eng.
Features of Biological Neural Networks 1)Robustness and Fault Tolerance. 2)Flexibility. 3)Ability to deal with variety of Data situations. 4)Collective.
Non-Bayes classifiers. Linear discriminants, neural networks.
ADALINE (ADAptive LInear NEuron) Network and
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Chapter 18 Connectionist Models
NEURAL NETWORKS LECTURE 1 dr Zoran Ševarac FON, 2015.
Chapter 6 Neural Network.
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
1 Technological Educational Institute Of Crete Department Of Applied Informatics and Multimedia Intelligent Systems Laboratory.
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
Lecture 2 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 2/1 Dr.-Ing. Erwin Sitompul President University
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Machine Learning Supervised Learning Classification and Regression
Multiple-Layer Networks and Backpropagation Algorithms
Chapter 2 Single Layer Feedforward Networks
One-layer neural networks Approximation problems
第 3 章 神经网络.
Real Neurons Cell structures Cell body Dendrites Axon
Ranga Rodrigo February 8, 2014
DEPARTMENT: COMPUTER SC. & ENGG. SEMESTER : VII
A Simple Artificial Neuron
Derivation of a Learning Rule for Perceptrons
Classification with Perceptrons Reading:
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Machine Learning Today: Reading: Maria Florina Balcan
CSC 578 Neural Networks and Deep Learning
Biological and Artificial Neuron
Artificial Neural Network & Backpropagation Algorithm
BACKPROPAGATION Multlayer Network.
Artificial Intelligence Chapter 3 Neural Networks
Ch2: Adaline and Madaline
Biological and Artificial Neuron
Artificial Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Machine Learning: Lecture 4
Machine Learning: UNIT-2 CHAPTER-1
Artificial Intelligence Chapter 3 Neural Networks
Chapter - 3 Single Layer Percetron
Artificial Intelligence Chapter 3 Neural Networks
Artificial Neural Networks ECE /ECE Fall 2006
Artificial Intelligence Chapter 3 Neural Networks
Presentation transcript:

Biological and Artificial Neuron Neural Networks Learning Processes Biological and Artificial Neuron Weights, need to be determined Biological neuron Bias, need to be determined Artificial neuron

Application of Neural Networks Learning Processes Application of Neural Networks Function approximation and prediction Pattern recognition Signal processing Modeling and control Machine learning

Building a Neural Network Neural Networks Learning Processes Building a Neural Network Select Structure: design the way that the neurons are interconnected. Select weights: decide the strengths with which the neurons are interconnected. Weights are selected to get a “good match” of network output to the output of a training set. Training set is a set of inputs and desired outputs. The weight selection is conducted by the use of a learning algorithm.

Artificial neural network Neural Networks Learning Processes Learning Process Stage 1: Network Training Artificial neural network Training Data Learning Process Knowledge Input and output sets, adequate coverage In the form of a set of optimized synaptic weights and biases Stage 2: Network Validation Artificial neural network Output Prediction Unseen Data Implementation Phase From the same range as the training data

Neural Networks Learning Processes Learning Process Learning is a process by which the free parameters of a neural network are adapted through a process of stimulation by the environment in which the network is embedded. In most cases, due to complex optimization plane, the optimized weights and biases are obtained as a result of a number of learning iterations. ANN [w,b] x y Initialize: Iteration (0) [w,b]0 x y(0) Iteration (1) [w,b]1 x y(1) … Iteration (n) [w,b]n x y(n) ≈ d d : desired output

Learning Rules Error Correction Learning Neural Networks Learning Processes Learning Rules Error Correction Learning Delta Rule or Widrow-Hoff Rule Memory Based Learning Nearest Neighbor Rule Hebbian Learning Synchronous activation increases the synaptic strength Asynchronous activation decreases the synaptic strength Competitive Learning Boltzmann Learning

Error-Correction Learning Neural Networks Learning Processes Error-Correction Learning Activation function wk1(n) x1 Desired output dk (n) wk2(n) x2 + Output yk (n) S f(.) S Inputs - Synaptic weights wkm(n) Error signal xm bk(n) ek (n) Bias 1 Learning Rule

Delta Rule (Widrow-Hoff Rule) Neural Networks Learning Processes Delta Rule (Widrow-Hoff Rule) Minimization of a cost function (or performance index)

Delta Rule (Widrow-Hoff Rule) Neural Networks Learning Processes Delta Rule (Widrow-Hoff Rule) wkj(0) = 0 n = 0 “Least Mean Square” Rule yk(n) = S [wkj(n) xj(n)] wkj(n+1) = wkj(n) + h [dk(n) – yk(n)] xj(n) h : learning rate, [0…1] n = n+1

Learning Paradigm Supervised Unsupervised S ANN ANN Environment (Data) Neural Networks Learning Processes Learning Paradigm Supervised Unsupervised Environment (Data) Delay ANN Delayed Reinforcement Learning Cost Function S ANN Error Desired Actual + - Environment (Data) Teacher (Expert)

Single Layer Perceptrons Neural Networks Single Layer Perceptrons Single Layer Perceptrons Single-layer perceptron network is a network with all the inputs connected directly to the output(s). Output unit is independent of the others. Analysis can be limited to single output perceptron.

Derivation of a Learning Rule for Perceptrons Neural Networks Single Layer Perceptrons Derivation of a Learning Rule for Perceptrons Key idea: Learning is performed by adjusting the weights in order to minimize the sum of squared errors on a training. Weights are updated repeatedly (in each epoch/iteration). Sum of squared errors is a classical error measure (e.g. commonly used in linear regression). E(w) Learning can be viewed as an optimization search problem in weight space. w1 w2

Derivation of a Learning Rule for Perceptrons Neural Networks Single Layer Perceptrons Derivation of a Learning Rule for Perceptrons The learning rule performs a search within the solution's vector space towards a global minimum. The error surface itself is a hyper-paraboloid but is seldom as smooth as is depicted below. In most problems, the solution space is quite irregular with numerous pits and hills which may cause the network to settle down in a local minimum (not the best overall solution). Epochs are repeated until stopping criterion is reached (error magnitude, number of iterations, change of weights, etc).

Derivation of a Learning Rule for Perceptrons Neural Networks Single Layer Perceptrons Derivation of a Learning Rule for Perceptrons x1 x2 xm wk1 wk2 wkm .  Adaline (Adaptive Linear Element) Widrow [1962] Goal:

Least Mean Squares (LMS) Neural Networks Single Layer Perceptrons Least Mean Squares (LMS) The following cost function (error function) should be minimized:

Least Mean Squares (LMS) Neural Networks Single Layer Perceptrons Least Mean Squares (LMS) Letting f(wk) = f (wk1, wk2,…, wkm) be a function over Rm, then Defining

Gradient Operator f w f w f w df : positive df : zero Neural Networks Single Layer Perceptrons Gradient Operator f w f w f w df : positive df : zero df : negative go uphill plain go downhill To minimize f , we choose df is thus guaranteed to be always negative

Adaline Learning Rule With then As already obtained before, Defining Neural Networks Single Layer Perceptrons Adaline Learning Rule With then As already obtained before, Weight Modification Rule Defining we can write

Adaline Learning Modes Neural Networks Single Layer Perceptrons Adaline Learning Modes Batch Learning Mode Incremental Learning Mode

Adaline Learning Rule -Learning Rule LMS Algorithm Neural Networks Single Layer Perceptrons Adaline Learning Rule -Learning Rule LMS Algorithm Widrow-Hoff Learning Rule

Generalization and Early Stopping Neural Networks Single Layer Perceptrons Generalization and Early Stopping By proper training, a neural network may produce reasonable output for inputs not seen during training  Generalization Generalization is particularly useful for the analysis of a “noisy” data (e.g. time–series) “Overtraining” will not improve the ability of a neural network to produce good output. On the contrary, it will try to take noise as the real data and lost its generality.

Generalization and Early Stopping Neural Networks Single Layer Perceptrons Generalization and Early Stopping Overfitting vs Generalization

Neural Networks Single Layer Perceptrons Homework 2 Given a function y = 4x2, you are required to find the value of x that will result y = 2 by using the Least Mean Squares method. Use initial estimate x0 = 1 and learning rate η = 0.01. Write down the results of the first 10 epochs/iterations. Give conclusion about your result. Note: Calculation can be done manually or using Matlab.

Neural Networks Single Layer Perceptrons Homework 2A Given a function y = 2x3 + cos2x, you are required to find the value of x that will result y = 5 by using the Least Mean Squares method. Use initial estimate x0 = 0.2*Student ID and learning rate η = 0.01. Write down the results of the first 10 epochs/iterations. Give conclusion about your result. Note: Calculation can be done manually or using Matlab/Excel.