Artificial neurons Nisheeth 10th January 2019.

Slides:



Advertisements
Similar presentations
Artificial Intelligence 12. Two Layer ANNs
Advertisements

Slides from: Doug Gray, David Poole
Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Kostas Kontogiannis E&CE
Biological and Artificial Neurons Michael J. Watts
Artificial Neural Networks - Introduction -
1 Part I Artificial Neural Networks Sofia Nikitaki.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
CS 484 – Artificial Intelligence
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial Neural Networks
Artificial neural networks:
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Artificial Neural Networks
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University EE459 Neural Networks The Structure.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Kim HS Introduction considering that the amount of MRI data to analyze in present-day clinical trials is often on the order of hundreds or.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Back Propagation and Representation in PDP Networks
Neural networks.
Back Propagation and Representation in PDP Networks
Regression.
Fall 2004 Backpropagation CS478 - Machine Learning.
Regression.
CS 388: Natural Language Processing: Neural Networks
Neural Networks.
Artificial Neural Networks
Artificial neural networks
Learning with Perceptrons and Neural Networks
Advanced information retreival
Artificial neural networks:
Real Neurons Cell structures Cell body Dendrites Axon
CSE 473 Introduction to Artificial Intelligence Neural Networks
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
CSE P573 Applications of Artificial Intelligence Neural Networks
Simple learning in connectionist networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
CS621: Artificial Intelligence
Hebb and Perceptron.
CSC 578 Neural Networks and Deep Learning
Biological Neuron Cell body Dendrites Axon Impulses Synapses
Perceptron as one Type of Linear Discriminants
CSE 573 Introduction to Artificial Intelligence Neural Networks
Neural Network - 2 Mayank Vatsa
G5AIAI Introduction to AI
Artificial Intelligence Lecture No. 28
Backpropagation.
Multilayer Perceptron: Learning : {(xi, f(xi)) | i = 1 ~ N} → W
Artificial Intelligence 12. Two Layer ANNs
Back Propagation and Representation in PDP Networks
Simple learning in connectionist networks
Introduction to Neural Network
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
David Kauchak CS158 – Spring 2019

PYTHON Deep Learning Prof. Muhammad Saeed.
Presentation transcript:

Artificial neurons Nisheeth 10th January 2019

Last class We started talking about associativity

Modeling classical conditioning Most popular approach for years was the Rescorla-Wagner model Could reproduce a number of empirical observations in classical conditioning experiments Some versions replace Vtot with Vx; what is the difference? http://users.ipfw.edu/abbott/314/Rescorla2.htm

https://en.wikipedia.org/wiki/Rescorla%E2%80%93Wagner_model

Bayesian models of classical conditioning

Problems for Bayesian conditioning models How to incorporate the role of time? How to incorporate the role of attention and salience? How to model the process by which animals learn the higher-order structure within latent causes?

Explaining associativity

Hebbian learning If two information processing units are activated simultaneously, their activity will become correlated, such that activity in one will make activity in the other more likely We will see How Hebbian learning works in artificial neurons How Hebbian learning works in real neurons

Hebbian learning i j wij

Hebb’s rule in action Common variant Problem No way to weaken weights Weights can climb up without bound Solution: normalization?

Normalizing Hebb’s rule Oja’s rule can be derived as an approximate normalized version of Hebb’s rule Other formulations are also possible, e.g. BCM, generalized Hebb Hebbian learning partially explains how neurons can encode new information And explains how this learning is fundamentally associative

Real neurons

What neurons do – the ML version A neuron collects signals from dendrites (not quite) Sends out spikes of electrical activity through an axon, which splits into thousands of branches. At the end of each branch, a synapse converts activity into either exciting or inhibiting activity of a dendrite at another neuron (more complicated than this) Neuron fires when exciting activity surpasses inhibitory activity above some threshold (neurotransmitters govern the threshold and activity rate) Learning changes the strength of synapses (but what sort of learning?)

Artificial neurons Logical abstraction of natural neurons to perform these functions

ANN Forward Propagation

ANN Forward Propagation Bias Nodes Add one node to each layer that has constant output Forward propagation Calculate from input layer to output layer For each neuron: Calculate weighted average of input Calculate activation function

Neuron Model Firing Rules: Threshold rules: Perceptron rule Calculate weighted average of input Fire if larger than threshold Perceptron rule Calculate weighted average of input input Output activation level is

Neuron Model Firing Rules: Sigmoid functions: Hyperbolic tangent function Logistic activation function

ANN Forward Propagation Apply input vector X to layer of neurons. Calculate where xi is the activation of previous layer neuron i wji is the weight of going from node i to node j p is the number of neurons in the previous layer Calculate output activation

ANN Forward Propagation Example: Three layer network Calculates xor of inputs -4.8 1 2 3 5.9 5.1 4 4.6 5.2 -5.2 -3.2 -2.6 -2.7 Bias

ANN Forward Propagation Input (0,0) -4.8 1 2 3 5.9 5.1 4 4.6 5.2 -5.2 -3.2 -2.6 -2.7 Bias

ANN Forward Propagation Input (0,0) Node 2 activation is (-4.8  0+4.6  0 - 2.6)= 0.0691 -4.8 1 2 3 5.9 5.1 4 4.6 5.2 -5.2 -3.2 -2.6 -2.7 Bias

ANN Forward Propagation Input (0,0) Node 3 activation is (5.1  0 - 5.2  0 - 3.2)= 0.0392 -4.8 1 2 3 5.9 5.1 4 4.6 5.2 -5.2 -3.2 -2.6 -2.7 Bias

ANN Forward Propagation Input (0,0) Node 4 activation is (5.9  0.069 + 5.2  0.069 – 2.7)= 0.110227 -4.8 1 2 3 5.9 5.1 4 4.6 5.2 -5.2 -3.2 -2.6 -2.7 Bias

ANN Forward Propagation Input (0,1) Node 2 activation is (4.6 -2.6)= 0.153269 -4.8 1 2 3 5.9 5.1 4 4.6 5.2 -5.2 -3.2 -2.6 -2.7 Bias

ANN Forward Propagation Input (0,1) Node 3 activation is (-5.2 -3.2)= 0.000224817 -4.8 1 2 3 5.9 5.1 4 4.6 5.2 -5.2 -3.2 -2.6 -2.7 Bias

ANN Forward Propagation Input (0,1) Node 4 activation is (5.9  0.153269 + 5.2  0.000224817 -2.7 )= 0.923992 -4.8 1 2 3 5.9 5.1 4 4.6 5.2 -5.2 -3.2 -2.6 -2.7 Bias

ANN Forward Propagation Network can learn a non-linearly separated set of outputs. Need to map output (real value) into binary values.

ANN Training Weights are determined by training Back-propagation: On given input, compare actual output to desired output. Adjust weights to output nodes. Work backwards through the various layers Start out with initial random weights Best to keep weights close to zero (<<10)

ANN Training Weights are determined by training Need a training set Should be representative of the problem During each training epoch: Submit training set element as input Calculate the error for the output neurons Calculate average error during epoch Adjust weights

ANN Training Error is the mean square of differences in output layer y – observed output t – target output

ANN Training Error of training epoch is the average of all errors.

ANN Training Update weights and thresholds using Weights Bias  is a possibly time-dependent learning rate that should prevent overcorrection https://machinelearningmastery.com/implement-perceptron-algorithm-scratch-python/