Introduction to Neural Networks

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Slides from: Doug Gray, David Poole
Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Computer Science Department FMIPA IPB 2003 Neural Computing Yeni Herdiyeni Computer Science Dept. FMIPA IPB.
Navneet Goyal, BITS-Pilani Perceptrons. Labeled data is called Linearly Separable Data (LSD) if there is a linear decision boundary separating the classes.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Brian Merrick CS498 Seminar.  Introduction to Neural Networks  Types of Neural Networks  Neural Networks with Pattern Recognition  Applications.
Artificial Intelligence (CS 461D)
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
20.5 Nerual Networks Thanks: Professors Frank Hoffmann and Jiawei Han, and Russell and Norvig.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Rutgers CS440, Fall 2003 Neural networks Reading: Ch. 20, Sec. 5, AIMA 2 nd Ed.
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
AN INTERACTIVE TOOL FOR THE STOCK MARKET RESEARCH USING RECURSIVE NEURAL NETWORKS Master Thesis Michal Trna
Artificial Neural Networks
Lecture 4 Neural Networks ICS 273A UC Irvine Instructor: Max Welling Read chapter 4.
Data Mining with Neural Networks (HK: Chapter 7.5)
Artificial Neural Networks
LOGO Classification III Lecturer: Dr. Bo Yuan
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
ICS 273A UC Irvine Instructor: Max Welling Neural Networks.
CS 484 – Artificial Intelligence
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
1 Introduction to Artificial Neural Networks Andrew L. Nelson Visiting Research Faculty University of South Florida.
Some more Artificial Intelligence
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Neural Networks. Plan Perceptron  Linear discriminant Associative memories  Hopfield networks  Chaotic networks Multilayer perceptron  Backpropagation.
Artificial Neural Networks
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Artificial Neural Networks An Overview and Analysis.
Chapter 9 Neural Network.
ANNs (Artificial Neural Networks). THE PERCEPTRON.
Waqas Haider Khan Bangyal. Multi-Layer Perceptron (MLP)
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
NEURAL NETWORKS FOR DATA MINING
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Artificial Neural Networks An Introduction. What is a Neural Network? A human Brain A porpoise brain The brain in a living creature A computer program.
Neural Networks and Backpropagation Sebastian Thrun , Fall 2000.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Lecture 5 Neural Control
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
Perceptrons Michael J. Watts
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Lecture 12. Outline of Rule-Based Classification 1. Overview of ANN 2. Basic Feedforward ANN 3. Linear Perceptron Algorithm 4. Nonlinear and Multilayer.
Computational Intelligence Semester 2 Neural Networks Lecture 2 out of 4.
Artificial Neural Networks By: Steve Kidos. Outline Artificial Neural Networks: An Introduction Frank Rosenblatt’s Perceptron Multi-layer Perceptron Dot.
Neural networks.
Artificial Neural Networks
Learning in Neural Networks
Artificial Intelligence (CS 370D)
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Perceptron as one Type of Linear Discriminants
Lecture Notes for Chapter 4 Artificial Neural Networks
David Kauchak CS158 – Spring 2019
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Introduction to Neural Networks Holger Wittel

Outline Motivation Neural Networks & Topologies Perceptron Multi-layered networks Training Time dependence Neural Networks for Filters Example usages Summary

Motivation Computers are fast, still certain tasks are difficult for them Your brain took ~100 cycles (0.1s) to classify the picture as ‘cat‘, what can your computer do in 100 cycles? brain computer No. of processing units 1011 109 Type of processing units neurons transistors Type of calculation massively parallel usually serial Switching time 10-3 sec 10-9 sec Possible switching operations 1014 per sec 1018 per sec Table taken from: D.Kriesel: Neuronal Netze, http://www.dkriesel.com/_media/science/neuronalenetze-en-zeta2-2col-dkrieselcom.pdf

A real Neuron Dendrite Axon terminal Node of Ranvier Soma Axon # neurons roundworm 302 ant 104 fly 105 mouse 5·106 cat 3·108 human 1011 elephant 2·1011 Schwann cell Myelin sheath Nucleus SEM image of a neuron, cut open to show sacks of neurotransmitters Images: http://en.wikipedia.org/wiki/Neuron http://www.cellimagelibrary.org/images/214

Neuron Model ..... Out = resp(Σ inputi · wi) weight w1 Input2 weight w2 response function Out = resp(Σ inputi · wi) ..... weight wi Inputi Most important response functions: Linear: resp(x)= x Sigmoid: resp(x)= 1 1+ 𝑒 −𝑥

The Perceptron / Input Neurons Output sigmoid → use for classification The magic happens in the connections & weights! Input neurons Output neurons weight w1 weight w2 Output =response(Σ inputi · wi)

The Bias Neuron -1 Input neurons Output neurons No inputs, constantly outputs bias Often omitted in drawings Implement as input neuron Input neurons Output neurons weight w1 weight w2 weight w3 -1

... ... Two-Layered Networks Always linear, no processing Response depends on desired output Input layer Output layer w11 w12 w21 w22 ... ...

History: OR / XOR 1958: Perceptron 1969: Minsky & Papert: Perceptron can only classify linearly separable data This killed nearly all research on NNs for the next 15 years No XOR as popular example: XOR i1 i2 or xor 1 OR 1 1 1 1

Multi-layered Networks Adding a middle (‘hidden‘) layer solves the problem! At least one nonlinear layer A 3-layered perceptron can represent virtually anything! -1 -1

Feed forward network (fully connected) Network Topologies Feed forward network (fully connected) with shortcuts -1 -1

Recurrent network with lateral connections Network Topologies Recurrent network with lateral connections -1 -1

Recurrent network with direct feedbacks Network Topologies Recurrent network with direct feedbacks -1 -1

Recurrent network with indirect feedback Network Topologies Recurrent network with indirect feedback -1 -1

Fully connected network (omitted some connections) Network Topologies Fully connected network (omitted some connections)

Three-Layer Perceptron We consider a 3-layer feed forward Perceptron By far most used NN in practice -1 -1

Training How to determine the right connection weights? Prepare training data set: inputs & desired outputs Optimize weights : Backpropagation Standard text book method Gradient based Useful for adaptive networks May get stuck in local minimum local minimum Genetic algorithm Biology inspired Finds global minimum May be slow global minimum

Excurs: Genetic Algorithms Start with random population Evaluate each individual & select best Make next generation: Migration – copy best individuals Mutation – copy best and randomly change some attributes Crossover – copy best and mix their attributes Next Generation

Adding time-dependence Until now the NNs cannot represent relationships in time, they need memory Input (t) output (t) Input (t-1) This is an IIR filter! This is an FIR filter! Input (t-2) ...

Uses of NNs abcdef ghijk ? Pattern recognition: Predicting the future Face recognition Character recognition Sound recognition Predicting the future Stock market Weather Signal processing abcdef ghijk ? DAX 2000 - today

Example: NETTALK Written language → spoken language Nontrivial problem: pronounce ‘a‘ in have, brave, read Simple Network: 3-layer perceptron After 10.000 words trained Reading a new text Training Ref: Sejnowsk, Rosenberg: ‘Parallel Networks that Learn to Pronounce English Text’, Complex Systems 1 (1987) 145-168

Seismic feed forward system NN of your choice Training with GA: Inputs: Seismometer Desired output: Interferometer channel, for example: michelson rotation 300 Generations, training time ~ mins 1 hidden neuron

Adaptive Filters Arbitrary sine wave Power in GEO, unlocked, flashing

Pitfalls Creation: Training: Which type of network? Number of neurons? Number of layers? Training: Which algorithm for training? When to stop training? Does the training converge to the global optimum? Training can take long Make sure the NN doesn‘t memorize the training data and loses ability to generalize

Ideas for NNs Automated classification Locklosses Detector state Filters Nonlinear filters Adaptive filters Predict the future Identify and characterize nonlinear couplings Nonlinear noise projections

Summary NNs can be used in many ways Easy to implement 3-Layer perceptron can model (virtually) any function or coupling, even nonlinear couplings NNs can be used as filters Training can be an issue & time intensive Computationally expensive for high frequency data For linear coupling no advantage over ‘traditional‘ (Wiener) Filters

Thank you for your attention! References: http://dkriesel.com/science/neural_networks Handbook of Neural Network Signal Processing, CRC Press

NARX Nonlinear AutoRegressive network with eXogenous inputs -1 -1 29

Nonlinear Noise Projections? Unexplained noise