1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.

Slides:



Advertisements
Similar presentations
Text Categorization.
Advertisements

Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
K-means method for Signal Compression: Vector Quantization
Text Categorization CSC 575 Intelligent Information Retrieval.
Instance Based Learning
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Classification and Decision Boundaries
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Nearest Neighbor. Predicting Bankruptcy Nearest Neighbor Remember all your data When someone asks a question –Find the nearest old data point –Return.
Learning for Text Categorization
K nearest neighbor and Rocchio algorithm
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
Instance Based Learning
Instance-Based Learning
Instance Based Learning IB1 and IBK Small section in chapter 20.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Instance Based Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán) 1.
INSTANCE-BASE LEARNING
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
CS Instance Based Learning1 Instance Based Learning.
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
Module 04: Algorithms Topic 07: Instance-Based Learning
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Nearest Neighbor Learning
PrasadL11Classify1 Vector Space Classification Adapted from Lectures by Raymond Mooney and Barbara Rosario.
K Nearest Neighborhood (KNNs)
DATA MINING LECTURE 10 Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines.
Nearest Neighbor (NN) Rule & k-Nearest Neighbor (k-NN) Rule Non-parametric : Can be used with arbitrary distributions, No need to assume that the form.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
1 Instance Based Learning Ata Kaban The University of Birmingham.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.7: Instance-Based Learning Rodney Nielsen.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 6: Nearest and k-nearest Neighbor Classification.
CpSc 881: Machine Learning Instance Based Learning.
CpSc 810: Machine Learning Instance Based Learning.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
KNN Classifier.  Handed an instance you wish to classify  Look around the nearby region to see what other classes are around  Whichever is most common—make.
Information Retrieval and Organisation Chapter 14 Vector Space Classification Dell Zhang Birkbeck, University of London.
DATA MINING LECTURE 10b Classification k-nearest neighbor classifier
CS Machine Learning Instance Based Learning (Adapted from various sources)
K-Nearest Neighbor Learning.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
CS 8751 ML & KDDInstance Based Learning1 k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning.
1 Instance Based Learning Soongsil University Intelligent Systems Lab.
Instance Based Learning
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Classification Nearest Neighbor
Instance Based Learning
Instance Based Learning (Adapted from various sources)
K Nearest Neighbor Classification
Classification Nearest Neighbor
Text Categorization Assigning documents to a fixed set of categories
Instance Based Learning
Nearest Neighbor Classifiers
Chap 8. Instance Based Learning
Machine Learning: UNIT-4 CHAPTER-1
Nearest Neighbors CSC 576: Data Mining.
Nearest Neighbor Classifiers
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
INFORMATION RETRIEVAL TECHNIQUES BY DR. ADNAN ABID
Presentation transcript:

1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin

2 Instance-Based Learning Unlike other learning algorithms, does not involve construction of an explicit abstract generalization but classifies new instances based on direct comparison and similarity to known training instances. Training can be very easy, just memorizing training instances. Testing can be very expensive, requiring detailed comparison to all past training instances. Also known as: –Case-based –Exemplar-based –Nearest Neighbor –Memory-based –Lazy Learning

3 Similarity/Distance Metrics Instance-based methods assume a function for determining the similarity or distance between any two instances. For continuous feature vectors, Euclidian distance is the generic choice: Where a p (x) is the value of the pth feature of instance x. For discrete features, assume distance between two values is 0 if they are the same and 1 if they are different (e.g. Hamming distance for bit vectors). To compensate for difference in units across features, scale all continuous values to the interval [0,1].

4 Other Distance Metrics Mahalanobis distance –Scale-invariant metric that normalizes for variance. Cosine Similarity –Cosine of the angle between the two vectors. –Used in text and other high-dimensional data. Pearson correlation –Standard statistical correlation coefficient. –Used for bioinformatics data. Edit distance –Used to measure distance between unbounded length strings. –Used in text and bioinformatics.

5 K-Nearest Neighbor Calculate the distance between a test point and every training instance. Pick the k closest training examples and assign the test instance to the most common category amongst these nearest neighbors. Voting multiple neighbors helps decrease susceptibility to noise. Usually use odd value for k to avoid ties.

6 5-Nearest Neighbor Example

7 Implicit Classification Function Although it is not necessary to explicitly calculate it, the learned classification rule is based on regions of the feature space closest to each training example. For 1-nearest neighbor with Euclidian distance, the Voronoi diagram gives the complex polyhedra segmenting the space into the regions closest to each point.

8 Efficient Indexing Linear search to find the nearest neighbors is not efficient for large training sets. Indexing structures can be built to speed testing. For Euclidian distance, a kd-tree can be built that reduces the expected time to find the nearest neighbor to O(log n) in the number of training examples. –Nodes branch on threshold tests on individual features and leaves terminate at nearest neighbors. Other indexing structures possible for other metrics or string data. –Inverted index for text retrieval.

9 Nearest Neighbor Variations Can be used to estimate the value of a real- valued function (regression) by taking the average function value of the k nearest neighbors to an input point. All training examples can be used to help classify a test instance by giving every training example a vote that is weighted by the inverse square of its distance from the test instance.

10 Feature Relevance and Weighting Standard distance metrics weight each feature equally when determining similarity. –Problematic if many features are irrelevant, since similarity along many irrelevant examples could mislead the classification. Features can be weighted by some measure that indicates their ability to discriminate the category of an example, such as information gain. Overall, instance-based methods favor global similarity over concept simplicity. + Training Data – + Test Instance ??

11 Rules and Instances in Human Learning Biases Psychological experiments show that people from different cultures exhibit distinct categorization biases. “Western” subjects favor simple rules (straight stem) and classify the target object in group 2. “Asian” subjects favor global similarity and classify the target object in group 1.

12 Other Issues Can reduce storage of training instances to a small set of representative examples. –Support vectors in an SVM are somewhat analogous. Can hybridize with rule-based methods or neural-net methods. –Radial basis functions in neural nets and Gaussian kernels in SVMs are similar. Can be used for more complex relational or graph data. –Similarity computation is complex since it involves some sort of graph isomorphism. Can be used in problems other than classification. –Case-based planning –Case-based reasoning in law and business.

13 Conclusions IBL methods classify test instances based on similarity to specific training instances rather than forming explicit generalizations. Typically trade decreased training time for increased testing time.