4.3 Feedforward Net. Applications

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Multi-Layer Perceptron (MLP)
Neural networks Introduction Fitting neural networks
Introduction to Neural Networks Computing
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Machine Learning Lecture 4 Multilayer Perceptrons G53MLE | Machine Learning | Dr Guoping Qiu1.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Classification Neural Networks 1
Machine Learning Neural Networks
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
Back-Propagation Algorithm
Chapter 6: Multilayer Neural Networks
Image Compression Using Neural Networks Vishal Agrawal (Y6541) Nandan Dubey (Y6279)
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
SOMTIME: AN ARTIFICIAL NEURAL NETWORK FOR TOPOLOGICAL AND TEMPORAL CORRELATION FOR SPATIOTEMPORAL PATTERN LEARNING.
7-Speech Recognition Speech Recognition Concepts
Multi Layer NN and Bit-True Modeling of These Networks SILab presentation Ali Ahmadi September 2007.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
Neural Networks Chapter 6 Joost N. Kok Universiteit Leiden.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Classification / Regression Neural Networks 2
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
Akram Bitar and Larry Manevitz Department of Computer Science
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
CS621 : Artificial Intelligence
Chapter 2 Single Layer Feedforward Networks
Neural Networks - lecture 51 Multi-layer neural networks  Motivation  Choosing the architecture  Functioning. FORWARD algorithm  Neural networks as.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks Lecture 4 out of 4. Practical Considerations Input Architecture Output.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Introduction to Neural Networks
Learning: Neural Networks Artificial Intelligence CMSC February 3, 2005.
Today’s Lecture Neural networks Training
Neural networks.
Neural networks and support vector machines
Handwritten Digit Recognition Using Stacked Autoencoders
Deep Feedforward Networks
an introduction to: Deep Learning
Natural Language and Text Processing Laboratory
Randomness in Neural Networks
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Computer Science and Engineering, Seoul National University
Chapter 2 Single Layer Feedforward Networks
Radial Basis Function G.Anuradha.
Intelligent Information System Lab
LECTURE 28: NEURAL NETWORKS
Convolutional Networks
Classification / Regression Neural Networks 2
Classification Neural Networks 1
of the Artificial Neural Networks.
Basics of Deep Learning No Math Required
LECTURE 28: NEURAL NETWORKS
McCulloch–Pitts Neuronal Model :
Ch4: Backpropagation (BP)
Neural networks (1) Traditional multi-layer perceptrons
Neural networks (3) Regularization Autoencoder
Autoencoders Supervised learning uses explicit labels/correct output in order to train a network. E.g., classification of images. Unsupervised learning.
Prediction Networks Prediction A simple example (section 3.7.3)
An introduction to: Deep Learning aka or related to Deep Neural Networks Deep Structural Learning Deep Belief Networks etc,
Introduction to Neural Networks
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Ch4: Backpropagation (BP)
Artificial Intelligence Chapter 3 Neural Networks
Akram Bitar and Larry Manevitz Department of Computer Science
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

4.3 Feedforward Net. Applications MLP: Signum - Classification Sigmoid - Mapping/Classification Some weights may be fixed (at zero for partially connected) or shared. Neural computing can offer relatively simple solutions to complex pattern classification problems. ‘Black Box' solution (1) M-class Classification(with sigmoidal nonlinearity) Train with {x, ek} if x  ωk ek = unit vector along kth dimension = (0 … 0 1 0 … 0) Test : If yk is MAX then x  ωk For large training set No local minimum trap Enough Plasticity Neural Network is a nonparametric probability density estimator with Perfect training. - [Ref. Rumelhart PDP Vol.1 Ch.8] k x N y MLP classifier yk  P (ωkx)  A posteriori class probability

(2) Parity (XOR / n-bit Parity) (3) Encoder-Decoder (Autoassociator / Autoencoder) Compression Expansion 1 O O 1 Ex. 4 - 2 - 4 5 - 3 - 5 N - M - N ( M < N) O O O O O O O O (4) ’87 Cottrell : Image Compression with limited channel capacity [too small to allow transmission of color and intensity ] (HDTV) − redundancy elimination via self-supervised BP. Train on random overlapping patches, test on a complete set of non-overlapping patches of the same image or even of very different images. * PCA Network with linear activation function can compress better. patch 8 Input Image 64 16 Output Compression Expansion (Encoding) (Decoding)

(5) T-C problem (6) Engineering T h i s i s t h e input (7-letter window slides over text) 80 Hidden units (6) Engineering a. Speech Synthesis - NETtalk by Sejnowski - trained with 1024 words – intelligible after 10 epochs - 95% accuracy after 50 epochs - like a child learning to talk - 78% generalization after complete training (still intelligible) - NN is easy to construct, can be used even when a problem is not fully understood. Adding noise to connections or removing units only degrades performance gracefully. cf. Commercial DECTALK ( 10 years of analysis by many linguists, rule-based ) b. Signal Processing : Ref. NN for SP, Kosko, 91.

c. Signal Prediction a time T into the future – Time Series Analysis ) ( t x 1 - z T + Neural Network current future t t +T Past input Benchmark Problem : <Mackey-Glass Differential Delay Eq.> - Dynamic system that is more chaotic if  is bigger. - NN models the dynamics of this system. It predicts better than traditional methods.

Question: IV must rapidly adapt to external changing Env. How ? d. Carnegie Mellon University Autonomous Land Vehicle ** To be shown on Video Question: IV must rapidly adapt to external changing Env. How ? CMU Navlab and the NN Based Autonomous Driving

Some Real World Data that Can be Used for NN Design

Students’ Questions from 2005 I think that learning in Speech Recognition also needs a desired output. While in English some sound follows some other sound, in Korean there will be far more combinations of consonants, vowels, and undersymbols. Do we need to compare all of them ? With that amount of data, learning time will explode. If data gets reduced in face recognition, then also the size of the face DB for comparison diminish ? In HDTV, why use 8x8 patches ? How about using 16x16 or more ? Do they use NN in practice for HDTV ? In the HDTV compression, even under no redundancy, some important features such as in face recognition might be extracted. Do you foresee any problem in this ? What happens to the indirect control when multiple inverses exist, i.e., when the next desired state differs even for the same current state ?