Learning Department of Computer Science & Engineering Indian Institute of Technology Kharagpur.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Learning from Observations Chapter 18 Section 1 – 3.
Slides from: Doug Gray, David Poole
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Back-propagation Chih-yun Lin 5/16/2015. Agenda Perceptron vs. back-propagation network Network structure Learning rule Why a hidden layer? An example:
CS 4700: Foundations of Artificial Intelligence
Kostas Kontogiannis E&CE
Decision making in episodic environments
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Learning From Observations
Neural Networks Marco Loog.
An Illustrative Example
LEARNING DECISION TREES
Learning decision trees derived from Hwee Tou Ng, slides for Russell & Norvig, AI a Modern Approachslides Tom Carter, “An introduction to information theory.
Learning decision trees
Learning decision trees derived from Hwee Tou Ng, slides for Russell & Norvig, AI a Modern Approachslides Tom Carter, “An introduction to information theory.
ICS 273A Intro Machine Learning
Classification and Prediction by Yen-Hsien Lee Department of Information Management College of Management National Sun Yat-Sen University March 4, 2003.
Learning: Introduction and Overview
Data Mining with Neural Networks (HK: Chapter 7.5)
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Machine learning Image source:
Machine learning Image source:
Machine Learning CPS4801. Research Day Keynote Speaker o Tuesday 9:30-11:00 STEM Lecture Hall (2 nd floor) o Meet-and-Greet 11:30 STEM 512 Faculty Presentation.
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Inductive learning Simplest form: learn a function from examples
1 st Neural Network: AND function Threshold(Y) = 2 X1 Y X Y.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Multiple-Layer Networks and Backpropagation Algorithms
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Chapter 9 Neural Network.
Appendix B: An Example of Back-propagation algorithm
Classification / Regression Neural Networks 2
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
LEARNING DECISION TREES Yılmaz KILIÇASLAN. Definition - I Decision tree induction is one of the simplest, and yet most successful forms of learning algorithm.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 21 Oct 28, 2005 Nanjing University of Science & Technology.
Learning from observations
Learning from Observations Chapter 18 Through
CHAPTER 18 SECTION 1 – 3 Learning from Observations.
Learning from Observations Chapter 18 Section 1 – 3, 5-8 (presentation TBC)
Learning from Observations Chapter 18 Section 1 – 3.
L6. Learning Systems in Java. Necessity of Learning No Prior Knowledge about all of the situations. Being able to adapt to changes in the environment.
Decision Trees. What is a decision tree? Input = assignment of values for given attributes –Discrete (often Boolean) or continuous Output = predicated.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Hazırlayan NEURAL NETWORKS Backpropagation Network PROF. DR. YUSUF OYSAL.
Introduction to Neural Networks Freek Stulp. 2 Overview Biological Background Artificial Neuron Classes of Neural Networks 1. Perceptrons 2. Multi-Layered.
Chapter 18 Section 1 – 3 Learning from Observations.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Learning From Observations Inductive Learning Decision Trees Ensembles.
Decision Tree Learning CMPT 463. Reminders Homework 7 is due on Tuesday, May 10 Projects are due on Tuesday, May 10 o Moodle submission: readme.doc and.
Multiple-Layer Networks and Backpropagation Algorithms
Learning from Observations
Learning from Observations
Introduce to machine learning
Presented By S.Yamuna AP/CSE
CSE P573 Applications of Artificial Intelligence Neural Networks
Prof. Carolina Ruiz Department of Computer Science
Classification Neural Networks 1
Decision making in episodic environments
Artificial Neural Network & Backpropagation Algorithm
General Aspects of Learning
CSE 573 Introduction to Artificial Intelligence Neural Networks
Backpropagation.
Learning from Observations
Learning from Observations
Machine Learning: Decision Tree Learning
Supplemental slides for CSE 327 Prof. Jeff Heflin
General Aspects of Learning
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

Learning Department of Computer Science & Engineering Indian Institute of Technology Kharagpur

CSE, IIT Kharagpur2 Paradigms of learning Supervised Learning –A situation in which both inputs and outputs are given –Often the outputs are provided by a friendly teacher. Reinforcement Learning –The agent receives some evaluation of its action (such as a fine for stealing bananas), but is not told the correct action (such as how to buy bananas). Unsupervised Learning –The agent can learn relationships among its percepts, and the trend with time

CSE, IIT Kharagpur3 Decision Trees A decision tree takes as input an object or situation described by a set of properties, and outputs a yes/no “decision”. Decision: Whether to wait for a table at a restaurant. 1.Alternate: whether there is a suitable alternative restaurant 2.Lounge: whether the restaurant has a lounge for waiting customers 3.Fri/Sat: true on Fridays and Saturdays 4.Hungry: whether we are hungry 5.Patrons: how many people are in it (None, Some, Full) 6.Price: the restaurant’ price range ( ★, ★★, ★★★ ) 7.Raining: whether it is raining outside 8.Reservation: whether we made a reservation 9.Type: the kind of restaurant (Indian, Chinese, Thai, Fastfood) 10.WaitEstimate: 0-10 mins, 10-30, 30-60, >60.

CSE, IIT Kharagpur4 Sample Decision Tree Patrons? Yes WaitEstimate?NoYes Reservation?Fri/Sat? Lounge?NoYes NoYes Alternate?NoHungry? Alternate?Yes Raining?Yes NoYes NoneSomeFull > NoYes No

CSE, IIT Kharagpur5 Decision Tree Learning Algorithm Function Decision-Tree-Learning( examples, attributes, default ) if examples is empty then return default else if all examples have the same classification then return the classification else if attributes is empty then return Majority-Value( examples ) else best  Choose-Attribute( attributes, examples ) tree  a new decision tree with root test best for each value v i of best do examples i  { elements of examples with best = v i } subtree  Decision-Tree-Learning( examples i, attributes – best, Majority-Value( examples )) add a branch to tree with label v i and subtree subtree end return tree

CSE, IIT Kharagpur6 Neural Networks A neural network consists of a set of nodes (neurons/units) connected by links –Each link has a numeric weight associated with it Each unit has: –a set of input links from other units, –a set of output links to other units, –a current activation level, and –an activation function to compute the activation level in the next time step.

CSE, IIT Kharagpur7 Basic definitions The total weighted input is the sum of the input activations times their respective weights: In each step, we compute: in i  aiai g Input Function Activation Function Output a i = g( in i ) Input Links ajaj W j,i Output Links

CSE, IIT Kharagpur8 Learning in Single Layer Networks IjIj W j,i OiOi If the output for a output unit is O, and the correct output should be T, then the error is given by: Err = T – O The weight adjustment rule is: W j  W j +  x I j x Err where  is a constant called the learning rate This method is unable to learn all types of functions can learn only linearly separable functions Used in competitive learning

CSE, IIT Kharagpur9 Two-layer Feed-Forward Network O i Output units W j,i a j Hidden units W k,j I k Input units

CSE, IIT Kharagpur10 Back-Propagation Learning Err i = (T i – O i ) is the error at the output node The weight update rule for the link from unit j to output unit i is: W j,i  W j,i +  x a j x Err i x g'(in i ) where g' is the derivative of the activation function g. Let  i = Err i g'(in i ). Then we have: W j,i  W j,i +  x a j x  i

CSE, IIT Kharagpur11 Error Back-Propagation For updating the connections between the inputs and the hidden units, we need to define a quantity analogous to the error term for the output nodes –Hidden node j is responsible for some fraction of the error  i each of the output nodes to which it connects –So, the  i values are divided according to the strength of the connection between the hidden node and the output node, and propagated back to provide the  j values for the hidden layer The weight update rule for weights between the inputs and the hidden layer is: W k,j  W k,j +  x I k x  j

CSE, IIT Kharagpur12 The theory The total error is given by: Only one of the terms in the summation over i and j depends on a particular W j,i, so all other terms are treated as constants with respect to W j,i Similarly: