CAP6938 Neuroevolution and Artificial Embryogeny Neural Network Weight Optimization Dr. Kenneth Stanley January 18, 2006.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
also known as the “Perceptron”
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Fall 2004 Shreekanth Mandayam ECE Department Rowan University.
Supervised learning 1.Early learning algorithms 2.First order gradient methods 3.Second order gradient methods.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
20.5 Nerual Networks Thanks: Professors Frank Hoffmann and Jiawei Han, and Russell and Norvig.
Neural Networks Marco Loog.
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
Before we start ADALINE
Data Mining with Neural Networks (HK: Chapter 7.5)
An Introduction To The Backpropagation Algorithm Who gets the credit?
LOGO Classification III Lecturer: Dr. Bo Yuan
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2006 Shreekanth Mandayam ECE Department Rowan University.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Neural Networks. Plan Perceptron  Linear discriminant Associative memories  Hopfield networks  Chaotic networks Multilayer perceptron  Backpropagation.
Artificial Neural Networks
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Computer Science and Engineering
Cascade Correlation Architecture and Learning Algorithm for Neural Networks.
Artificial Neural Networks
Neural Networks Chapter 6 Joost N. Kok Universiteit Leiden.
Explorations in Neural Networks Tianhui Cai Period 3.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
Appendix B: An Example of Back-propagation algorithm
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
Lecture 3 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 3/1 Dr.-Ing. Erwin Sitompul President University
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
CAP6938 Neuroevolution and Developmental Encoding Basic Concepts Dr. Kenneth Stanley August 23, 2006.
Non-Bayes classifiers. Linear discriminants, neural networks.
Neural Networks and Backpropagation Sebastian Thrun , Fall 2000.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
CS621 : Artificial Intelligence
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
EEE502 Pattern Recognition
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Announcements 1. Textbook will be on reserve at library 2. Topic schedule change; modified reading assignment: This week: Linear discrimination, evaluating.
An Introduction To The Backpropagation Algorithm.
Evolutionary Computation Evolving Neural Network Topologies.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Neural Networks.
The Gradient Descent Algorithm
Artificial neural networks
Real Neurons Cell structures Cell body Dendrites Axon
CSE 473 Introduction to Artificial Intelligence Neural Networks
Derivation of a Learning Rule for Perceptrons
Dr. Kenneth Stanley September 6, 2006
CSE P573 Applications of Artificial Intelligence Neural Networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
Prof. Carolina Ruiz Department of Computer Science
Data Mining with Neural Networks (HK: Chapter 7.5)
CSE 573 Introduction to Artificial Intelligence Neural Networks
network of simple neuron-like computing elements
Neural Network - 2 Mayank Vatsa
Lecture Notes for Chapter 4 Artificial Neural Networks
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
David Kauchak CS158 – Spring 2019
Prof. Carolina Ruiz Department of Computer Science
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

CAP6938 Neuroevolution and Artificial Embryogeny Neural Network Weight Optimization Dr. Kenneth Stanley January 18, 2006

Review Remember, the values of the weights and the topology determine the functionality Given a topology, how are weights optimized? Weights are just parameters on a structure ? ? ? ?? ? ? ??

Two Cases Output targets are known Output targets are not known X1X1 X2X2 H1H1 H2H2 out 1 out 2 w 11 w 21 w 12

Decision Boundaries OR function: InputOutput OR is linearly separable Linearly separable problems do not require hidden nodes (nonlinearities) Bias

Decision Boundaries XOR is not linearly separable Requires at least one hidden node XOR function: InputOutput Bias

Hebbian Learning Change weights based on correlation of connected neurons Learning rules are local Simple Hebb Rule: Works best when relevance of inputs to outputs is independent Simple Hebb Rule grows weights unbounded Can be made incremental:

More Complex Local Learning Rules Hebbian Learning with a maximum magnitude: –Excitatory: –Inhibitory: Second terms are decay terms: forgetting –Happens when presynaptic node does not affect postsynaptic node Other rules are possible Videos: watch the connections change

Perceptron Learning Will converge on correct weights Single layer learning rule: Rule is applied until boundary is learned Bias

Backpropagation Designed for at least one hidden layer First, activation propagates to outputs Then, errors are computed and assigned Finally, weights are updated Sigmoid is a common activation function X1X1 X2X2 z1z1 z2z2 y1y1 y2y2 v 11 v 21 v 12 v 22 w 11 w 21 w 12 w 22 t1t1 t2t2 x’s are inputs z’s are hidden units y’s are outputs t’s are targets v’s are layer 1 weights w’s are layer 2 weights

Backpropagation Algorithm 1)Initialize weights 2)While stopping condition is false, for each training pair 1)Compute outputs by forward activation 2)Backpropagate error: 1)For each output unit, error 2) Weight correction 3)Send error back to hidden units 4)Calculate error contribution for each hidden unit: 5)Weight correction 3)Adjust weights by adding weight corrections (target minus output times slope) (Learning rate times error times hidden output)

Example Applications Anything with a set of examples and known targets XOR Character recognition NETtalk: reading English aloud Failure predicition Disadvantages: trapped in local optima

Output Targets Often Not Available (Stone, Sutton, and Kuhlmann 2005)

One Approach: Value Function Reinforcement Learning Divide the world into states and actions Assign values to states Gradually learn the most promising states and actions Start Goal

Learning to Navigate Start Goal Start Goal Start Goal Start Goal T=1 T=56 T=350 T=703

How to Update State/Action Values Q learning rule: Exploration increases Q-values’ accuracy The best actions to take in different states become known Works only in Markovian domains

Backprop In RL The state/action table can be estimated by a neural network The target learned by the network is the Q-value: NN Action State_description Value

Next Week: Evolutionary Computation For 1/23: Mitchell ch.1 (pp. 1-31) and ch.2 (pp ) Note Section 2.3 is "Evolving Neural Networks" For 1/25: Mitchell pp , paper: No Free Lunch Theorems for Optimization (1996)No Free Lunch Theorems for Optimization by David H. Wolpert, William G. Macready EC does not require targets EC can be a kind of RL EC is policy search EC is more than RL