For Wednesday Read chapter 19, sections 1-3 No homework.

Slides:



Advertisements
Similar presentations
Slides from: Doug Gray, David Poole
Advertisements

1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Tuomas Sandholm Carnegie Mellon University Computer Science Department
Kostas Kontogiannis E&CE
Reading for Next Week Textbook, Section 9, pp A User’s Guide to Support Vector Machines (linked from course website)
Artificial Neural Networks
Classification Neural Networks 1
Perceptron.
Machine Learning Neural Networks
Artificial Neural Networks
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Neural Networks Marco Loog.
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Back-Propagation Algorithm
Chapter 6: Multilayer Neural Networks
Artificial Neural Networks
Artificial Neural Networks
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
CS 4700: Foundations of Artificial Intelligence
INTRODUCTION TO ARTIFICIAL INTELLIGENCE
CS 484 – Artificial Intelligence
For Monday Read chapter 22, sections 1-3 No homework.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
CISC 4631 Data Mining Lecture 11: Neural Networks.
Neural Networks Chapter 18.7
Chapter 10 Boosting May 6, Outline Adaboost Ensemble point-view of Boosting Boosting Trees Supervised Learning Methods.
Artificial Neural Networks
1 CS 343: Artificial Intelligence Neural Networks Raymond J. Mooney University of Texas at Austin.
Artificial Neural Networks
For Wednesday Finish chapter 22 No homework. Program 4 Any questions?
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
1 Machine Learning Neural Networks Neural Networks NN belong to the category of trained classification models The learned classification model is an.
CS464 Introduction to Machine Learning1 Artificial N eural N etworks Artificial neural networks (ANNs) provide a general, practical method for learning.
Machine Learning Chapter 4. Artificial Neural Networks
For Friday Read chapter 18, section 7 Program 3 due.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
For Thursday No reading or homework Exam 2 Take-home due Tuesday Read chapter 22 for next Tuesday.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
For Monday Read Chapter 22, sections 4-6 No written homework.
For Wednesday No reading No homework. Exam 2 Friday. Will cover material through chapter 18. Take home is due Friday.
For Wednesday Read chapter 22, sections 4-6 Homework: –Chapter 18, exercise 7.
For Wednesday No reading Homework: –Chapter 18, exercise 25 (a & b)
Neural Networks and Backpropagation Sebastian Thrun , Fall 2000.
1 CS 391L: Machine Learning Neural Networks Raymond J. Mooney University of Texas at Austin.
For Friday No reading Take home exam due Exam 2. For Monday Read chapter 22, sections 1-3 FOIL exercise due.
Ensemble Methods in Machine Learning
Artificial Neural Networks (Cont.) Chapter 4 Perceptron Gradient Descent Multilayer Networks Backpropagation Algorithm 1.
1 Machine Learning Neural Networks Tomorrow no lesson! (will recover on Wednesday 11, WEKA !!)
Artificial Neural Networks Chapter 4 Perceptron Gradient Descent Multilayer Networks Backpropagation Algorithm 1.
For Friday Read ch. 20, sections 1-3 Program 4 due.
Artificial Neural Network
Chapter 6 Neural Network.
129 Feed-Forward Artificial Neural Networks AMIA 2003, Machine Learning Tutorial Constantin F. Aliferis & Ioannis Tsamardinos Discovery Systems Laboratory.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning Supervised Learning Classification and Regression
CS 388: Natural Language Processing: Neural Networks
Artificial Neural Networks
Artificial Neural Networks
Real Neurons Cell structures Cell body Dendrites Axon
Artificial Neural Networks
Artificial Neural Networks
Machine Learning Today: Reading: Maria Florina Balcan
Classification Neural Networks 1
Lecture Notes for Chapter 4 Artificial Neural Networks
Seminar on Machine Learning Rada Mihalcea
Presentation transcript:

For Wednesday Read chapter 19, sections 1-3 No homework

Program 4 Any questions?

Exam 2 Covers material through decision trees The rest of learning will be on the final

Backpropogation Learning Rule Each weight w ji is changed by  w ji =  j o i  j = o j (1 - o j ) (t j - o j )if j is an output unit  j = o j (1 - o j )   k w kj otherwise where  is a constant called the learning rate, t j is the correct output for unit j, d j is an error measure for unit j. First determine the error for the output units, then backpropagate this error layer by layer through the network, changing weights appropriately at each layer.

Backpropogation Learning Algorithm Create a three layer network with N hidden units and fully connect input units to hidden units and hidden units to output units with small random weights. Until all examples produce the correct output within e or the mean­squared error ceases to decrease (or other termination criteria): Begin epoch For each example in training set do: Compute the network output for this example. Compute the error between this output and the correct output. Backpropagate this error and adjust weights to decrease this error. End epoch Since continuous outputs only approach 0 or 1 in the limit, must allow for some e­approximation to learn binary functions.

Comments on Training There is no guarantee of convergence, may oscillate or reach a local minima. However, in practice many large networks can be adequately trained on large amounts of data for realistic problems. Many epochs (thousands) may be needed for adequate training, large data sets may require hours or days of CPU time. Termination criteria can be: –Fixed number of epochs –Threshold on training set error

Representational Power Multi­layer sigmoidal networks are very expressive. Boolean functions: Any Boolean function can be represented by a two layer network by simulating a two­layer AND­OR network. But number of required hidden units can grow exponentially in the number of inputs. Continuous functions: Any bounded continuous function can be approximated with arbitrarily small error by a two­layer network. Sigmoid functions provide a set of basis functions from which arbitrary functions can be composed, just as any function can be represented by a sum of sine waves in Fourier analysis. Arbitrary functions: Any function can be approximated to arbitarary accuracy by a three­layer network.

Sample Learned XOR Network Hidden unit A represents ¬(X  Y) Hidden unit B represents ¬(X  Y) Output O represents: A  ¬B ¬(X  Y)  (X  Y) X  Y A B XY

Hidden Unit Representations Trained hidden units can be seen as newly constructed features that re­represent the examples so that they are linearly separable. On many real problems, hidden units can end up representing interesting recognizable features such as vowel­detectors, edge­ detectors, etc. However, particularly with many hidden units, they become more “distributed” and are hard to interpret.

Input/Output Coding Appropriate coding of inputs and outputs can make learning problem easier and improve generalization. Best to encode each binary feature as a separate input unit and for multi­valued features include one binary unit per value rather than trying to encode input information in fewer units using binary coding or continuous values.

I/O Coding cont. Continuous inputs can be handled by a single input by scaling them between 0 and 1. For disjoint categorization problems, best to have one output unit per category rather than encoding n categories into log n bits. Continuous output values then represent certainty in various categories. Assign test cases to the category with the highest output. Continuous outputs (regression) can also be handled by scaling between 0 and 1.

Neural Net Conclusions Learned concepts can be represented by networks of linear threshold units and trained using gradient descent. Analogy to the brain and numerous successful applications have generated significant interest. Generally much slower to train than other learning methods, but exploring a rich hypothesis space that seems to work well in many domains. Potential to model biological and cognitive phenomenon and increase our understanding of real neural systems. –Backprop itself is not very biologically plausible

Beyond a Single Learner Ensembles of learners work better than individual learning algorithms Several possible ensemble approaches: –Ensembles created by using different learning methods and voting –Bagging –Boosting

Bagging Random selections of examples to learn the various members of the ensemble. Seems to work fairly well, but no real guarantees.

Boosting Most used ensemble method Based on the concept of a weighted training set. Works especially well with weak learners. Start with all weights at 1. Learn a hypothesis from the weights. Increase the weights of all misclassified examples and decrease the weights of all correctly classified examples. Learn a new hypothesis. Repeat