Dimensions of Neural Networks Ali Akbar Darabi Ghassem Mirroshandel Hootan Nokhost.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Slides from: Doug Gray, David Poole
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Introduction to Neural Networks Computing
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Artificial Neural Networks
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Modular Neural Networks CPSC 533 Franco Lee Ian Ko.
Machine Learning Neural Networks
Non-linear classification problem using NN Fainan May 2006 Pattern Classification and Machine Learning Course Three layers Feedforward Neural Network (FFNN)
Simple Neural Nets For Pattern Classification
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
November 2, 2010Neural Networks Lecture 14: Radial Basis Functions 1 Cascade Correlation Weights to each new hidden node are trained to maximize the covariance.
Chapter 6: Multilayer Neural Networks
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Aula 4 Radial Basis Function Networks
Radial Basis Function (RBF) Networks
Radial-Basis Function Networks
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Neural networks.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Artificial Neural Networks
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Cascade Correlation Architecture and Learning Algorithm for Neural Networks.
Artificial Neural Networks
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
1 RECENT DEVELOPMENTS IN MULTILAYER PERCEPTRON NEURAL NETWORKS Walter H. Delashmit Lockheed Martin Missiles and Fire Control Dallas, TX 75265
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 1. Stat 231. A.L. Yuille. Fall Perceptron Rule and Convergence Proof Capacity.
Neural Networks 2nd Edition Simon Haykin
Perceptrons Michael J. Watts
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning Supervised Learning Classification and Regression
Neural networks.
Neural networks and support vector machines
Fall 2004 Backpropagation CS478 - Machine Learning.
Deep Learning Amin Sobhani.
CSE 473 Introduction to Artificial Intelligence Neural Networks
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Counter propagation network (CPN) (§ 5.3)
Neuro-Computing Lecture 4 Radial Basis Function Network
Neural Network - 2 Mayank Vatsa
Neural Networks Geoff Hulten.
Capabilities of Threshold Neurons
Lecture Notes for Chapter 4 Artificial Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Multilayer Perceptron: Learning : {(xi, f(xi)) | i = 1 ~ N} → W
The Network Approach: Mind as a Web
Artificial Intelligence Chapter 3 Neural Networks
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Dimensions of Neural Networks Ali Akbar Darabi Ghassem Mirroshandel Hootan Nokhost

Outline Motivation Neural Networks Power Kolmogorov Theory Cascade Correlation

Motivation Consider you are an engineer and you know ANN You encounter a problem that can not be solved with common analytical approaches You decide to use ANN

But… Some questions  Is this problem solvable using ANN?  How many neurons?  How many layers?  …

Two Approaches Fundamental Analyses  Kolmogrov Theory Adaptive Networks  Cascade Correlation

Outline Motivation Neural Networks Power Kolmogorov Theory Cascade Correlation

Single layer Networks Limitations of the perceptron and linear classifiers

A Solution

Network Construction (x,y)→(x^2, y^2,x*y) 1 2

Network Construction (con…)

Network Construction (Con…)

Learning Mechanism Using Error Function Gradient Descent

Outline Motivation Neural Networks Power Kolmogorov Theory Cascade Correlation

Kolmogorov theorem (concept) An example:  Any continuous function of n dimensions can be completely characterized by a dimensional continuous functions

An Idea Suppose we want to construct f (x, y) A simple idea: find a mapping  (x, y) → r Then define a function g such that:  g(r) = f(x, y) x y r g

An Example Suppose we have a discrete function:  We choose a mapping  We define the 1-dimentional function  So 

Kolmogrov theorem In the illustrated example we had: 

Applying to the neural networks

Universal Approximation Neural Networks with a hidden layer can approximate any continuous function with arbitrary precision  Use independent function from main function  approximate the network with traditional networks

A kolmogorov Network We have to define:  Mapping  Function g

Spline Function Linear combination of several 3-dimensional functions Used to approximate functions with given points

Mapping x y

Example X1=2.5 X2=4.5 x1 x

Function g Now for each unique input value of a we should define a output value g corresponding to f We choose the value of f in the center of the square X1=2.5 X2=4.5

Function g (Con…)

Reduce Error Shifting defined patterns N different patterns will be generated Use avg y

Replace the function With sufficiently large number of knots:

Outline Motivation Neural Networks Power Kolmogorov Theory Cascade Correlation

Dynamic size, depth, topology Single layer learning in spite of multilayer structure Fast learning

Architecture

Algorithm step 1

Adding Hidden Layer

Correlation Residual error for output unit for pattern p  Average residual error for output unit  Computed activation for input vector x(p)  Z(p) Average activation, over all patterns, of candidate unit 

Correlation Use Variance as a similarity criteria Update weights similar to gradient descent

Algorithm step 2

Adding Hiding Neuron

Algorithm Step 3

Final Result

An Example 100 Run 1700 epochs on avg Beats standard backprob with factor 10 with the same complexity

Results Cascade Correlation is either better  Only forward pass  Many of epochs are run while the network is very small  Cashing mechanism

Network Steps

Network Steps (con..)

Another Example N-input parity problem Standard backprob takes 2000 epoches on N=8 with 16 hidden neurons

Discussion There is no need to guess the size, depth and the connectivity pattern Learns fast Can build deep networks (high order feature detector) Herd effect Results can be cashed

Conclusion A Network with a hidden layer can define complex boundaries and can approximate any function The number of neurons in the hidden layer determines the amount of approximation Dynamic Networks