Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.

Slides:



Advertisements
Similar presentations
INSTANCE-BASED LEARNING ALGORITHMS Presented by Yan T. Yang.
Advertisements

1 Classification using instance-based learning. 3 March, 2000Advanced Knowledge Management2 Introduction (lazy vs. eager learning) Notion of similarity.
DECISION TREES. Decision trees  One possible representation for hypotheses.
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
Instance Based Learning
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Classification and Decision Boundaries
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Nearest Neighbor. Predicting Bankruptcy Nearest Neighbor Remember all your data When someone asks a question –Find the nearest old data point –Return.
K nearest neighbor and Rocchio algorithm
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
Instance-Based Learning
Instance Based Learning. Nearest Neighbor Remember all your data When someone asks a question –Find the nearest old data point –Return the answer associated.
© Vipin Kumar CSci 8980 Fall CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance Computing Research Center Department of Computer.
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
1 Nearest Neighbor Learning Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AI.
CES 514 – Data Mining Lec 9 April 14 Mid-term k nearest neighbor.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Instance Based Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán) 1.
INSTANCE-BASE LEARNING
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor.
CS Instance Based Learning1 Instance Based Learning.
Module 04: Algorithms Topic 07: Instance-Based Learning
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Instance Selection. 1.Introduction 2.Training Set Selection vs. Prototype Selection 3.Prototype Selection Taxonomy 4.Description of Methods 5.Related.
K Nearest Neighborhood (KNNs)
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
1 Data Mining Lecture 5: KNN and Bayes Classifiers.
K Nearest Neighbors Saed Sayad 1www.ismartsoft.com.
K Nearest Neighbors Classifier & Decision Trees
1 Pattern Classification X. 2 Content General Method K Nearest Neighbors Decision Trees Nerual Networks.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
CSC 196k Semester Project: Instance Based Learning
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Statistical Inference (By Michael Jordon) l Bayesian perspective –conditional perspective—inferences.
1 Instance Based Learning Ata Kaban The University of Birmingham.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.7: Instance-Based Learning Rodney Nielsen.
CpSc 881: Machine Learning Instance Based Learning.
CpSc 810: Machine Learning Instance Based Learning.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 6.2: Classification Rules Rodney Nielsen Many.
COMP 2208 Dr. Long Tran-Thanh University of Southampton K-Nearest Neighbour.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Outline K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
CS Machine Learning Instance Based Learning (Adapted from various sources)
K-Nearest Neighbor Learning.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Presenter : Chien-Shing Chen Author: Gustavo.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
Data Mining Practical Machine Learning Tools and Techniques Chapter 6.5: Instance-based Learning Rodney Nielsen Many / most of these slides were adapted.
Data Mining – Algorithms: Instance-Based Learning
Instance Based Learning
Instance Based Learning (Adapted from various sources)
K Nearest Neighbor Classification
Instance Based Learning
COSC 4335: Other Classification Techniques
Chap 8. Instance Based Learning
Advanced Mathematics Hossein Malekinezhad.
Machine Learning: UNIT-4 CHAPTER-1
Nearest Neighbors CSC 576: Data Mining.
Data Mining Classification: Alternative Techniques
Nearest Neighbor Classifiers
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Presentation transcript:

Instance-Based Learning Evgueni Smirnov

Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based Learning Nearest Neighbor (NN) Algorithm Advantages and Disadvantages of the NN algorithm Approaches to overcome the Disadvantages of the NN algorithm Combining Eager and Instance-Based Learning

– Learning = storing all training instances – Classification = an instance gets a classification equal to the classification of the nearest instances to the instance. Instance-Based Learning

Different Learning Methods Eager Learning –Learning = acquiring an explicit structure of a classifier on the whole training set; –Classification = an instance gets a classification using the explicit structure of the classifier. Instance-Based Learning (Lazy Learning) –Learning = storing all training instances –Classification = an instance gets a classification equal to the classification of the nearest instances to the instance.

Different Learning Methods Eager Learning Any random movement =>It’s a mouse I saw a mouse!

Instance-Based Learning Its very similar to a Desktop!!

The Features of the Task of the NN Algorithm: the instance language I is a conjunctive language with a set A with n attributes a 1, a 2, … a n. The domain of each attribute a i, can be discrete or continuous. an instance x is represented as, where a i (x) is the value of the attribute a i for the instance x; the classes to be learned can be: –discrete. In this case we learn discrete function f(x) and the co- domain C of the function consists of the classes c to be learned. –continuous. In this case we learn continuous function f(x) and the co-domain C of the function consists of the classes c to be learned. Nearest-Neighbor Algorithm (NN)

Distance Functions The distance functions are composed from difference metrics d a w.r.t. attributes a defined for each two instances x i and x j. If the attribute a is numerical, then : If the attribute a is discrete, then :

Distance Functions The main distance function for determining nearest neighbors is the Euclidean distance:

The case of discrete set of classes. 1.Take the instance x to be classified 2.Find k nearest neighbors of x in the training data. 3.Determine the class c of the majority of the instances among the k nearest neighbors. 4.Return the class c as the classification of x. k-Nearest-Neighbor Algorithm

e1 1-nn:1-nn: q1 is positive 5-nn: q1 is classified as negative q1 Classification & Decision Boundaries

The case of continuous set of classes (Regression). 1.Take the instance x to be classified 2.Find k nearest neighbors of x in the training data. 3.Return the average of the classes of the k nearest neighbors as the classification of x. k-Nearest-Neighbor Algorithm

The case of discrete set of classes. 1.Take the instance x to be classified 2.Determine for each class c the sum 3.Return the class c with the greater S c. Distance Weighted Nearest-Neighbor Algorithm

Advantages of the NN Algorithm the NN algorithm can estimate complex target classes locally and differently for each new instance to be classified; the NN algorithm provides good generalisation accuracy on many domains; the NN algorithm learns very quickly; the NN algorithm is robust to noisy training data; the NN algorithm is intuitive and easy to understand which facilitates implementation and modification.

Disadvantages of the NN Algorithm the NN algorithm has large storage requirements because it has to store all the data; the NN algorithm is slow during instance classification because all the training instances have to be visited; the accuracy of the NN algorithm degrades with increase of noise in the training data; the accuracy of the NN algorithm degrades with increase of irrelevant attributes.

Condensed NN Algorithm The Condensed NN algorithm was introduced to reduce the storage requirements of the NN algorithm. The algorithm finds a subset S of the training data D s.t. each instance in D can be correctly classified by the NN algorithm applied on the subset S. The average reduction of the algorithm varies between 60% to 80%.

Condensed NN Algorithm This algorithm first randomly selects one instance for each class in D and puts it in S. Then each instance in D is classified using only the instances in S. If an instance is misclassified, it is added to S. This process is repeated until there are no instances in D that are misclassified. DS

Condensed NN Algorithm The CNN algorithm is especially sensitive to noise, because noisy instances will usually be misclassified by their neighbors, and thus will be retained. This causes two problems. storage reduction is hindered, because noisy instances are retained, and because they are there, often non-noisy instances nearby will also need to be retained. generalization accuracy is hurt because noisy instances are usually exceptions and thus do not represent the underlying function well.

The Edited Nearest Neighbor algorithm was proposed to stabilise the accuracy of the NN algorithm when there is increase of noise in the training data. The algorithm starts with the set S equal to the training data D, and then each instance in S is removed if it does not agree with the majority of its k nearest neighbors (with k=3, typically). The algorithm edits out noisy instances as well as close border cases, leaving smoother decision boundaries. It also retains all internal points; i.e., it does not reduce the space as much as most other reduction algorithms. Edited NN Algorithm

e1 The negative instance is removed! Edited NN Algorithm The average reduction of the algorithm varies between 20% to 40%.

The weighting-attribute technique was proposed in order to improve the accuracy of the NN algorithm in the presence of irrelevant attributes. The key idea is to find weights for all the attribute and to use them when the distance between instances is computed. Determining the weights of the attributes can be done by some search algorithm while determining the adequacy of the weights can be done with the process of cross validation. In a similar way we can choose the best k parameter for the NN algorithm! Weighting Attributes

Combining Decision Tress and the NN Algorithm Outlook sunnyovercastrainy HumidityWindy highnormal no falsetrue yes no

Combining Decision Tress and the NN Algorithm Outlook sunnyovercastrainy HumidityWindy highnormalfalsetrue yes Classify the instance using the NN algorithm applied on the training instances associated with the classification nodes (leaves)

Combining Decision Rules and the NN Algorithm Incrementally Compile InstancesInstances & Abstractions

Summary Points Instance-based learning is simple, efficient and accurate approach to concept learning and classification. Many of the problems of instance-based learning can be solved. Instance-based learning can be combined with eager approaches to concept learning.