LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Multi-Layer Perceptron (MLP)
NEURAL NETWORKS Biological analogy
Slides from: Doug Gray, David Poole
Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Artificial Neural Networks
Artificial Neural Networks
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Lecture 14 – Neural Networks
Artificial Intelligence (CS 461D)
1 Part I Artificial Neural Networks Sofia Nikitaki.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Data Mining with Neural Networks (HK: Chapter 7.5)
Artificial Neural Networks
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
CS 484 – Artificial Intelligence
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial neural networks:
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Networks
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Explorations in Neural Networks Tianhui Cai Period 3.
2101INT – Principles of Intelligent Systems Lecture 10.
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University EE459 Neural Networks The Structure.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
Features of Biological Neural Networks 1)Robustness and Fault Tolerance. 2)Flexibility. 3)Ability to deal with variety of Data situations. 4)Collective.
Non-Bayes classifiers. Linear discriminants, neural networks.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Artificial Neural Networks Chapter 4 Perceptron Gradient Descent Multilayer Networks Backpropagation Algorithm 1.
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
EEE502 Pattern Recognition
Each neuron has a threshold value Each neuron has weighted inputs from other neurons The input signals form a weighted sum If the activation level exceeds.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Kim HS Introduction considering that the amount of MRI data to analyze in present-day clinical trials is often on the order of hundreds or.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Neural networks.
Artificial Neural Networks
Learning with Perceptrons and Neural Networks
Learning in Neural Networks
Artificial Intelligence (CS 370D)
Artificial neural networks:
Real Neurons Cell structures Cell body Dendrites Axon
CSE 473 Introduction to Artificial Intelligence Neural Networks
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CSE P573 Applications of Artificial Intelligence Neural Networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
Data Mining with Neural Networks (HK: Chapter 7.5)
Chapter 3. Artificial Neural Networks - Introduction -
CSE 573 Introduction to Artificial Intelligence Neural Networks
Neural Network - 2 Mayank Vatsa
Lecture Notes for Chapter 4 Artificial Neural Networks
ARTIFICIAL NEURAL networks.
David Kauchak CS158 – Spring 2019
Presentation transcript:

LINEAR CLASSIFICATION

Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected to the others through synapses  Properties of the brain  It can learn, reorganize itself from experience  It adapts to the environment  It is robust and fault tolerant

Biological neuron ( simplified model )  A neuron has  A branching input ( dendrites )  A branching output ( the axon )  The information circulates from the dendrites to the axon via the cell body  The cell body sums up the inputs in some way and fires – generates a signal through the axon – if the result is greater than some threshold

An Artificial Neuron Activation Function Usually not pictured (we’ll see why), but you can imagine a threshold parameter here.

Same Idea using the Notation in the Book

The Output of a Neuron  As described so far… This simplest form of a neuron is also called a perceptron.

The Output of a Neuron  Other possibilities, such as the sigmoid function for continuous output.

Linear Regression using a Perceptron  Linear regression :

Linear Regression As an Optimization Problem  Finding the optimal weights could be solved through :  Gradient descent  Simulated annealing  Genetic algorithms  … and now Neural Networks

Linear Regression using a Perceptron

The Bias Term  So far we have defined the output of a perceptron as controlled by a threshold x 1 w 1 + x 2 w 2 + x 3 w 3 … + x n w n >= t  But just like the weights, this threshold is a parameter that needs to be adjusted  Solution : make it another weight x 1 w 1 + x 2 w 2 + x 3 w 3 … + x n w n + (1)(- t ) >= 0 The bias term.

A Neuron with a Bias Term

Another Example  Assign weights to perform the logical OR operation.

Artificial Neural Network ( ANN )  A mathematical model to solve engineering problems  Group of highly connected neurons to realize compositions of non linear functions  Tasks  Classification  Discrimination  Estimation

Feed Forward Neural Networks  The information is propagated from the inputs to the outputs  There are no cycles between outputs and inputs  the state of the system is not preserved from one iteration to another x1 x2xn….. 1st hidden layer 2nd hidden layer Output layer

ANN Structure  Finite number of inputs  Zero or more hidden layers  One or more outputs  All nodes at the hidden and output layers contain a bias term.

Examples  Handwriting character recognition  Control of a virtual agent

ALVINN Neural Network controlled AGV (1994)

Learning  The procedure that consists in estimating the weight parameters so that the whole network can perform a specific task  The Learning process ( supervised )  Present the network a number of inputs and their corresponding outputs  See how closely the actual outputs match the desired ones  Modify the parameters to better approximate the desired outputs

Perceptron Learning Rule

Linear Separability  Perceptrons can classify any input that is linearly separable.  For more complex problems we need a more complex model.

Different Non - Linearly Separable Problems Structure Types of Decision Regions Exclusive-OR Problem Classes with Meshed regions Most General Region Shapes Single-Layer Two-Layer Three-Layer Half Plane Bounded By Hyperplane Convex Open Or Closed Regions Arbitrary (Complexity Limited by No. of Nodes) A AB B A AB B A AB B B A B A B A

Calculating the Weights Perceptron learning rule is pretty much gradient descent.

Learning the Weights in a Neural Network  Perceptron learning rule ( gradient descent ) worked before, but it required us to know the correct output of the node.  How do we know the correct output of a given hidden node ??

Backpropagation Algorithm  Gradient descent over entire network weight vector  Easily generalized to arbitrary directed graphs  Will find a local, not necessarily global error minimum  in practice often works well ( can be invoked multiple times with different initial weights )

Backpropagation Algorithm

Intuition  General idea : hidden nodes are “ responsible ” for some of the error at the output nodes it connects to  The change in the hidden weights is proportional to the strength ( magnitude ) of the connection between the hidden node and the output node  This is the same as the perceptron learning rule, but for a sigmoid decision function instead of a step decision function ( full derivation on p. 726)

Intuition  General idea : hidden nodes are “ responsible ” for some of the error at the output nodes it connects to  The change in the hidden weights is proportional to the strength ( magnitude ) of the connection between the hidden node and the output node

Intuition  When expanded, the update to the output nodes is almost the same as the perceptron rule  Slight difference is that the algorithm uses a sigmoid function instead of a step function  ( full derivation on p. 726)

Questions