CS621 : Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 21 Computing power of Perceptrons and Perceptron Training.

Slides:



Advertisements
Similar presentations
CS344: Principles of Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 11, 12: Perceptron Training 30 th and 31 st Jan, 2012.
Advertisements

CS344: Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 15, 16: Perceptrons and their computing power 6 th and.
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
B.Macukow 1 Lecture 3 Neural Networks. B.Macukow 2 Principles to which the nervous system works.
1 Human Nervous System. 2 What are we going to learn ? Function of nervous system Parts of nervous system Types of neurons Conduction of impulse through.
Neural Networks (II) Simple Learning Rule
A Review: Architecture
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
The Brain Chapter 2.
Artificial Neural Network Yalong Li Some slides are from _24_2011_ann.pdf.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 25: Perceptrons; # of regions;
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
CS344: Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31 and 32– Brain and Perceptron.
Biology and Behavior Chapter 3. The Nervous System Central Nervous System – consists of the brain and spinal cord. Central Nervous System – consists of.
1 Chapter 20 Section Slide Set 2 Perceptron examples Additional sources used in preparing the slides: Nils J. Nilsson’s book: Artificial Intelligence:
The Muscular System Muscles contribute to the outward appearance of animals and are essential for movement, posture, breathing, circulation, digestion,
CS621: Artificial Intelligence Lecture 11: Perceptrons capacity Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Elements of Neuronal Biophysics The human brain Seat of consciousness and cognition Perhaps the most complex information processing machine in nature.
CS621 : Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 19: Fuzzy Logic and Neural Net Based IR.
Instructor: Prof. Pushpak Bhattacharyya 13/08/2004 CS-621/CS-449 Lecture Notes CS621/CS449 Artificial Intelligence Lecture Notes Set 6: 22/09/2004, 24/09/2004,
CS344 : Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 29 Introducing Neural Nets.
Prof. Pushpak Bhattacharyya, IIT Bombay 1 CS 621 Artificial Intelligence Lecture /10/05 Prof. Pushpak Bhattacharyya Linear Separability,
CS623: Introduction to Computing with Neural Nets Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
CS621: Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 41,42– Artificial Neural Network, Perceptron, Capacity 2 nd, 4 th Nov,
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 30: Perceptron training convergence;
The Nervous System. Key Concepts Muscle Motor Neuro n Interneuron Skin receptors Sensory Neuron Brain Know the function and divisions of the nervous system.
Linear Discrimination Reading: Chapter 2 of textbook.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 29: Perceptron training and.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Instructor: Prof. Pushpak Bhattacharyya 13/08/2004 CS-621/CS-449 Lecture Notes CS621/CS449 Artificial Intelligence Lecture Notes Set 4: 24/08/2004, 25/08/2004,
Parts of the Brain BY Jovian Cardona.
The Nervous System Chapter 11.
CS621 : Artificial Intelligence
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 32: sigmoid neuron; Feedforward.
Chapter 1 The Physical Foundation of Behavior. © Copyright 2009 Delmar, Cengage Learning. All Rights Reserved.2 The Influence of DNA DNA: complex set.
CS621 : Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 21: Perceptron training and convergence.
CS 621 Artificial Intelligence Lecture /11/05 Guest Lecture by Prof
November 21, 2013Computer Vision Lecture 14: Object Recognition II 1 Statistical Pattern Recognition The formal description consists of relevant numerical.
The Brain. Studying the Brain Three different methods to view Brain –MRI –fMRI –PET.
The Human Nervous System. The Nervous System has TWO Major Divisions. The Central Nervous System The Peripheral Nervous System.
CS621 : Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 20: Neural Net Basics: Perceptron.
CS621: Artificial Intelligence Lecture 10: Perceptrons introduction Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
CS 621 Artificial Intelligence Lecture /11/05 Guest Lecture by Prof. Rohit Manchanda Biological Neurons - II.
CS623: Introduction to Computing with Neural Nets (lecture-5)
CS344: Introduction to Artificial Intelligence (associated lab: CS386)
Pushpak Bhattacharyya Computer Science and Engineering Department
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
8.3 The Central Nervous System
CS621: Artificial Intelligence
CS623: Introduction to Computing with Neural Nets (lecture-2)
Pushpak Bhattacharyya Computer Science and Engineering Department
CS623: Introduction to Computing with Neural Nets (lecture-4)
Artificial Intelligence Lecture No. 28
CS 621 Artificial Intelligence Lecture /10/05 Prof
Functions of the Brain Year 10 – Psychology.
CS621: Artificial Intelligence
CS621: Artificial Intelligence Lecture 12: Counting no
CS623: Introduction to Computing with Neural Nets (lecture-5)
CS623: Introduction to Computing with Neural Nets (lecture-3)
CS344 : Introduction to Artificial Intelligence
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
Prof. Pushpak Bhattacharyya, IIT Bombay
CS621: Artificial Intelligence Lecture 14: perceptron training
CS621: Artificial Intelligence Lecture 18: Feedforward network contd
CS621 : Artificial Intelligence
CS621: Artificial Intelligence Lecture 17: Feedforward network (lecture 16 was on Adaptive Hypermedia: Debraj, Kekin and Raunak) Pushpak Bhattacharyya.
Presentation transcript:

CS621 : Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 21 Computing power of Perceptrons and Perceptron Training

The human brain Seat of consciousness and cognition Perhaps the most complex information processing machine in nature

Brain Map Forebrain (Cerebral Cortex): Language, maths, sensation, movement, cognition, emotion Cerebellum: Motor Control Midbrain: Information Routing; involuntary controls Hindbrain: Control of breathing, heartbeat, blood circulation Spinal cord: Reflexes, information highways between body & brain

Maslow’s Hierarchy of Needs

Brain’s algorithms? Evolutionarily, brain has developed algorithms most suitable for survival Algorithms unknown: the search is on Brain astonishing in the amount of information it processes –Typical computers: 10 9 operations/sec –Housefly brain: operations/sec

Brain facts & figures Basic building block of nervous system: nerve cell (neuron) ~ neurons in brain ~ connections between them Connections made at “synapses” The speed: events on millisecond scale in neurons, nanosecond scale in silicon chips

Computing Power of Perceptron

The Perceptron Model A perceptron is a computing element with input lines having associated weights and the cell having a threshold value. The perceptron model is motivated by the biological neuron. Output = y wnwn W n-1 w1w1 X n-1 x1x1 Threshold = θ

θ 1 y Step function / Threshold function y = 1 for Σw i x i >=θ =0 otherwise ΣwixiΣwixi

Concept of Hyper-planes ∑ w i x i = θ defines a linear surface in the (W,θ) space, where W= is an n-dimensional vector. A point in this (w,θ) space defines a perceptron. y x1x1... θ w1w1 w2w2 w3w3 wnwn x2x2 x3x3 xnxn

Functions computed by the simplest perceptron (single input) f4f3f2f1x θ≥0 w≤θ θ≥0 w> θ θ<0 w≤ θ θ<0 W< θ 0-function Identity Function Complement Function True-Function

Counting the number of functions for the simplest perceptron For the simplest perceptron, the equation is w.x=θ. Substituting x=0 and x=1, we get θ=0 and w=θ. These two lines intersect to form four regions, which correspond to the four functions. θ=0 w=θ R1 R2 R3 R4 θ w

Fundamental Observation The number of TFs computable by a perceptron is equal to the number of regions produced by 2 n hyper-planes,obtained by plugging in the values in the equation ∑ i=1 n w i x i = θ

The geometrical observation Problem: m linear surfaces called hyper- planes (each hyper-plane is of (d-1)-dim) in d-dim, then what is the max. no. of regions produced by their intersection? i.e. R m,d = ?

Case of 2-input perceptrons Output = y w2w2 w1w1 x1x1 Threshold = θ x2x2

Basic equation w 1 x 1 +w 2 x 2 =θ There are 4 values of the input: (0,0), (0,1), (1,0), (1,1) The relevant space is the (w 1,w 2,θ) coordinate system w1w1 w2w2 θ

4 planes All go through the origin –With (0,0), θ= 0 -- (1) –With (1,0), w 1 = θ --(2) –With (0,1), w 2 = θ --(3) –With (1,1), w 1 + w 2 = θ --(4) How many regions do they produce? That is equal to the number of functions computable by a 2-input perceptron

How to think about this counting problem? Whenever a new plane comes in, the existing planes intersect the new plane is a set of lines all going through the origin These lines produce some regions on the new plane (notice that we are not worrying about the regions in the space, but are thinking about the regions on the new plane) These regions produced on the new plane is the additional number of regions produced in the (w,θ) space

Counting the maximum no. of regions produced by 4 planes passing through origin Plane numberAdditional Regions produced 1 st 2 2 nd (1 st plane cuts this plane to form a line through the origin) 2 3 rd (1 st and 2 nd planes cut this plane to form two lines through the origin) 4 4 th (1 st, 2 nd and 3 rd planes cut this plane to form thre lines through the origin) 6

It is clear why 2-input perceptron computes 14 functions = regions are produced by the planes. Hence only 14 functions computed Terminology: Functions computable by a perceptron are called threshold functions

General case A perceptron with n weights and the threshold defines an (n+1)-dimensional space. From the basic equation Σ 1 n w i x i =θ, 2 n planes passing through origin are produced How many regions do they produce in the space?

Recurrence relation produced by m hyperplanes in d-dimension –C(m,d)= C(m-1,d)+C(m-1,d-1) Boundary conditions: –C(m,1)=2 (degenerate case of m points ‘passing through’ origin) –C(1,d)=2 Existing regions Additional regions on the d th plance

Threshold functions miniscule compared to Boolean Function Solution of the recurrence relation leads to the max. no. of regions as 2^n 2 vs. 2^2 n boolean functions