COMP53311 Other Classification Models: Neural Network Prepared by Raymond Wong Some of the notes about Neural Network are borrowed from LW Chan’s notes.

Slides:



Advertisements
Similar presentations
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Advertisements

Artificial Neural Networks
Artificial Neural Networks - Introduction -
Artificial Neural Networks - Introduction -
Machine Learning Neural Networks
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2008 Shreekanth Mandayam ECE Department Rowan University.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2008 Shreekanth Mandayam ECE Department Rowan University.
Artificial Intelligence Statistical learning methods Chapter 20, AIMA (only ANNs & SVMs)
20.5 Nerual Networks Thanks: Professors Frank Hoffmann and Jiawei Han, and Russell and Norvig.
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Fall 2004 Shreekanth Mandayam ECE Department Rowan University.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2010 Shreekanth Mandayam ECE Department Rowan University.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Spring 2002 Shreekanth Mandayam Robi Polikar ECE Department.
Data Mining with Neural Networks (HK: Chapter 7.5)
Artificial Neural Networks
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Radial Basis Function (RBF) Networks
Neural Networks.
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
Multi-Layer Perceptrons Michael J. Watts
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
1 Pattern Classification X. 2 Content General Method K Nearest Neighbors Decision Trees Nerual Networks.
Artificial Neural Networks An Introduction. What is a Neural Network? A human Brain A porpoise brain The brain in a living creature A computer program.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
ADVANCED PERCEPTRON LEARNING David Kauchak CS 451 – Fall 2013.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 24 Nov 2, 2005 Nanjing University of Science & Technology.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 8: Neural Networks.
Artificial Neural Network Building Using WEKA Software
Soft Computing Lecture 19 Part 2 Hybrid Intelligent Systems.
Non-Bayes classifiers. Linear discriminants, neural networks.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
CS 478 – Tools for Machine Learning and Data Mining Perceptron.
Artificial Neural Networks Students: Albu Alexandru Deaconescu Ionu.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
CS621 : Artificial Intelligence
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
EEE502 Pattern Recognition
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
“Principles of Soft Computing, 2 nd Edition” by S.N. Sivanandam & SN Deepa Copyright  2011 Wiley India Pvt. Ltd. All rights reserved. CHAPTER 2 ARTIFICIAL.
Intro. ANN & Fuzzy Systems Lecture 3 Basic Definitions of ANN.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Fall 2004 Perceptron CS478 - Machine Learning.
Other Classification Models: Neural Network
Classification 3 (Nearest Neighbor Classifier)
Artificial Intelligence (CS 370D)
Artificial neural networks:
Other Classification Models: Neural Network
Neural Networks Dr. Peter Phillips.
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Data Mining with Neural Networks (HK: Chapter 7.5)
XOR problem Input 2 Input 1
Other Classification Models: Support Vector Machine (SVM)
The Network Approach: Mind as a Web
Prof. Pushpak Bhattacharyya, IIT Bombay
Presentation transcript:

COMP53311 Other Classification Models: Neural Network Prepared by Raymond Wong Some of the notes about Neural Network are borrowed from LW Chan’s notes Presented by Raymond Wong

COMP53312 What we learnt for Classification Decision Tree Bayesian Classifier Nearest Neighbor Classifier

COMP53313 Other Classification Models Neural Network

COMP53314 Neural Network A computing system made up of simple and highly interconnected processing elements Other terminologies: Connectionist Models Parallel distributed processing models (PDP) Artificial Neural Networks Computational Neural Networks Neurocomputers

COMP53315 Neural Network This approach is inspired by the way that brains process information, which is entirely differ from the way that conventional computers do Information processing occurs at many identical and simple processing elements called neurons (or also called units, cells or nodes) Interneuron connection strengths known as synaptic weights are used to store the knowledge

COMP53316 Advantages of Neural Network Parallel Processing – each neuron operates individually Fault tolerance – if a small number of neurons break down, the whole system is still able to operate with slight degradation in performance.

COMP53317 Neuron Network Neuron Network for OR Neuron Network input output x1x1 x2x2 y x1x1 x2x2 y OR Function

COMP53318 Neuron Network Neuron Network for OR input output x1x1 x2x2 y x1x1 x2x2 y OR Function Neuron

COMP53319 Neuron Network Neuron Network for OR input output x1x1 x2x2 y x1x1 x2x2 y OR Function Front Back net

COMP Neuron Network Neuron Network for OR input output x1x1 x2x2 y x1x1 x2x2 y OR Function Front Back net w1w1 w2w2 Weight net = w 1 x 1 + w 2 x 2 + b

COMP Neuron Network Neuron Network for OR input output x1x1 x2x2 y x1x1 x2x2 y OR Function Front Back net w1w1 w2w2 Activation function -Linear function: y = net or y = a. net -Non-linear function

COMP Activation Function Non-linear functions Threshold function, Step Function, Hard Limiter Piecewise-Linear Function Sigmoid Function Radial Basis Function

COMP Threshold function, Step Function, Hard Limiter y = if net  0 if net  < net y 1

COMP Piecewise-Linear Function y = if net  a if –a < net  < a net y 1 if net  -a ½(1/a x net + 1) -aa

COMP Sigmoid Function y = 0 net y e -net

COMP Radial Basis Function y = 0 net y 1 -net 2 e

COMP Neuron Network Neuron Network for OR input output x1x1 x2x2 y x1x1 x2x2 y OR Function Front Back net w1w1 w2w2 net = w 1 x 1 + w 2 x 2 + b Threshold function y = if net  0 if net  <0 1 0

COMP Learning Let  be the learning rate (a real number) Learning is done by w i  w i +  (d – y)x i where d is the desired output y is the output of our neural network b  b +  (d – y)

COMP Neuron Network Neuron Network for OR input output x1x1 x2x2 y x1x1 x2x2 d OR Function Front Back net w1w1 w2w2 net = w 1 x 1 + w 2 x 2 + b Threshold function y = if net  0 if net  <0 1 0 net = w 1 x 1 + w 2 x 2 + b y = if net  0 if net  <0 1 0

COMP Neuron Network x1x1 x2x2 d net = w 1 x 1 + w 2 x 2 + b y = if net  0 if net  <0 1 0 net = w 1 x 1 + w 2 x 2 + b bw1w1 w2w2 111 =1 y = 1 w 1 = w 1 +  (d – y)x 1 = 1+0.8*(0-1)*0 = 1 w 2 = w 2 +  (d – y)x 2 = 1+0.8*(0-1)*0 = 1 b = b +  (d – y) = 1+0.8*(0-1) = Incorrect!  = 0.8

COMP Neuron Network x1x1 x2x2 d net = w 1 x 1 + w 2 x 2 + b y = if net  0 if net  <0 1 0 net = w 1 x 1 + w 2 x 2 + b bw1w1 w2w =1.2 y = 1 w 1 = w 1 +  (d – y)x 1 = 1+0.8*(1-1)*0 = 1 w 2 = w 2 +  (d – y)x 2 = 1+0.8*(1-1)*1 = 1 b = b +  (d – y) = *(1-1) = Correct!  = 0.8

COMP Neuron Network x1x1 x2x2 d net = w 1 x 1 + w 2 x 2 + b y = if net  0 if net  <0 1 0 net = w 1 x 1 + w 2 x 2 + b bw1w1 w2w =1.2 y = 1 w 1 = w 1 +  (d – y)x 1 = 1+0.8*(1-1)*1 = 1 w 2 = w 2 +  (d – y)x 2 = 1+0.8*(1-1)*0 = 1 b = b +  (d – y) = *(1-1) = Correct!  = 0.8

COMP Neuron Network x1x1 x2x2 d net = w 1 x 1 + w 2 x 2 + b y = if net  0 if net  <0 1 0 net = w 1 x 1 + w 2 x 2 + b bw1w1 w2w =2.2 y = 1 w 1 = w 1 +  (d – y)x 1 = 1+0.8*(1-1)*1 = 1 w 2 = w 2 +  (d – y)x 2 = 1+0.8*(1-1)*1 = 1 b = b +  (d – y) = *(1-1) = Correct!  = 0.8

COMP Neuron Network x1x1 x2x2 d net = w 1 x 1 + w 2 x 2 + b y = if net  0 if net  <0 1 0 net = w 1 x 1 + w 2 x 2 + b bw1w1 w2w =0.2 y = 1 w 1 = w 1 +  (d – y)x 1 = 1+0.8*(0-1)*0 = 1 w 2 = w 2 +  (d – y)x 2 = 1+0.8*(0-1)*0 = 1 b = b +  (d – y) = *(0-1) = Incorrect! We repeat the above process until the neural networks output the correct values of y (i.e., y = d for each possible input)  = 0.8

COMP Neuron Network Neuron Network for OR input output x1x1 x2x2 y x1x1 x2x2 y OR Function Front Back net w1w1 w2w2 net = w 1 x 1 + w 2 x 2 + b Threshold function y = if net  0 if net  <0 1 0

COMP Limitation It can only solve linearly separable problems

COMP Multi-layer Perceptron (MLP) Neuron Network input output x1x1 x2x2 y input output x1x1 x2x2 y Neuron

COMP Multi-layer Perceptron (MLP) Neuron Network input output x1x1 x2x2 y input output x1x1 x2x2 y

COMP Multi-layer Perceptron (MLP) input output x1x1 y1y1 x2x2 x3x3 x4x4 x5x5 y2y2 y3y3 y4y4 Input layerHidden layer Output layer

COMP Advantages of MLP Can solve Linear separable problems Non-linear separable problems A universal approximator MLP has proven to be a universal approximator, i.e., it can model all types function y = f(x)