Reliability and Information Gain Ida Sprinkhuizen-Kuyper Evgueni Smirnov Georgi Nalbantov (UM/EUR)

Slides:



Advertisements
Similar presentations
Statistical Machine Learning- The Basic Approach and Current Research Challenges Shai Ben-David CS497 February, 2007.
Advertisements

Introduction to Support Vector Machines (SVM)
ECG Signal processing (2)
An Introduction of Support Vector Machine
An Introduction of Support Vector Machine
Support Vector Machines
SVM—Support Vector Machines
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
IMAN SAUDY UMUT OGUR NORBERT KISS GEORGE TEPES-NICA BARLEY SEEDS CLASSIFICATION.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Fuzzy Support Vector Machines (FSVMs) Weijia Wang, Huanren Zhang, Vijendra Purohit, Aditi Gupta.
Rutgers CS440, Fall 2003 Support vector machines Reading: Ch. 20, Sec. 6, AIMA 2 nd Ed.
Linear Learning Machines  Simplest case: the decision function is a hyperplane in input space.  The Perceptron Algorithm: Rosenblatt, 1956  An on-line.
Support Vector Machines Kernel Machines
Announcements  Project teams should be decided today! Otherwise, you will work alone.  If you have any question or uncertainty about the project, talk.
1 Computational Learning Theory and Kernel Methods Tianyi Jiang March 8, 2004.
Support Vector Machines
Oregon State University – Intelligent Systems Group 8/22/2003ICML Giorgio Valentini Dipartimento di Scienze dell Informazione Università degli Studi.
05/06/2005CSIS © M. Gibbons On Evaluating Open Biometric Identification Systems Spring 2005 Michael Gibbons School of Computer Science & Information Systems.
Optimization Theory Primal Optimization Problem subject to: Primal Optimal Value:
An SVM Based Voting Algorithm with Application to Parse Reranking Paper by Libin Shen and Aravind K. Joshi Presented by Amit Wolfenfeld.
Linear Discriminators Chapter 20 From Data to Knowledge.
Support Vector Machines Exercise solutions Ata Kaban The University of Birmingham.
An Introduction to Support Vector Machines Martin Law.
Classification III Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart Russell,
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Efficient Model Selection for Support Vector Machines
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
Transcription of Text by Incremental Support Vector machine Anurag Sahajpal and Terje Kristensen.
Support Vector Machines Mei-Chen Yeh 04/20/2010. The Classification Problem Label instances, usually represented by feature vectors, into one of the predefined.
Classification and Ranking Approaches to Discriminative Language Modeling for ASR Erinç Dikici, Murat Semerci, Murat Saraçlar, Ethem Alpaydın 報告者:郝柏翰 2013/01/28.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
An Introduction to Support Vector Machines (M. Law)
Artificial Intelligence Research Laboratory Bioinformatics and Computational Biology Program Computational Intelligence, Learning, and Discovery Program.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
1 CSC 4510, Spring © Paula Matuszek CSC 4510 Support Vector Machines (SVMs)
An Introduction to Support Vector Machine (SVM)
Support Vector Machines in Marketing Georgi Nalbantov MICC, Maastricht University.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
Support vector machine LING 572 Fei Xia Week 8: 2/23/2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A 1.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
CSSE463: Image Recognition Day 15 Announcements: Announcements: Lab 5 posted, due Weds, Jan 13. Lab 5 posted, due Weds, Jan 13. Sunset detector posted,
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
SVMs, Part 2 Summary of SVM algorithm Examples of “custom” kernels Standardizing data for SVMs Soft-margin SVMs.
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
SVMs in a Nutshell.
CSSE463: Image Recognition Day 14 Lab due Weds. Lab due Weds. These solutions assume that you don't threshold the shapes.ppt image: Shape1: elongation.
Ensembles of Classifiers Evgueni Smirnov. Outline 1 Methods for Independently Constructing Ensembles 1.1 Bagging 1.2 Randomness Injection 1.3 Feature-Selection.
Support Vector Machine (SVM) Presented by Robert Chen.
CSSE463: Image Recognition Day 15 Today: Today: Your feedback: Your feedback: Projects/labs reinforce theory; interesting examples, topics, presentation;
An Introduction of Support Vector Machine In part from of Jinwei Gu.
Introduction to Machine Learning Prof. Nir Ailon Lecture 5: Support Vector Machines (SVM)
Support Vector Machines Part 2. Recap of SVM algorithm Given training set S = {(x 1, y 1 ), (x 2, y 2 ),..., (x m, y m ) | (x i, y i )   n  {+1, -1}
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
An Introduction to Support Vector Machines
Support Vector Machines Introduction to Data Mining, 2nd Edition by
Support Vector Machines
CSSE463: Image Recognition Day 14
Support Vector Machines
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Introduction to Machine Learning
Presentation transcript:

Reliability and Information Gain Ida Sprinkhuizen-Kuyper Evgueni Smirnov Georgi Nalbantov (UM/EUR)

Outline Reliability vs. accuracy Version Spaces Support Vector Machines Version Space Support Vector Machines Beating existing algorithms Conclusions

Reliability vs. accuracy Accuracy is a global measure It gives no information about individual instances In practice we need reliability of the instance (patient, diagnosis, …) in order to take an acceptable decision

Version Spaces A version space is the set of all hypotheses consistent with the training set Strong point: Unanimous voting results in high reliability Problem: Noise can result in version space collapse

Support Vector Machines SVM’s and kernel methods make a trade-off between accuracy on the training set and complexity of hyperplane generated by the kernel used They try to find a hyperplane with margin as large as possible and error term as small as possible

VSM (2) Strong point: Natural way to handle noise Problem: How to measure reliability of an individual instance

VSSVM Combine the best of the two worlds! First attempt: Version Space Support Vector Machines Implement the concept of unanimous voting by using SVM’s: An instance is positive iff no SVM hyperplane exists that thinks it is negative

VSSVM (2) Find kernel (RBF or polynomial) and parameters (C: cost of error; G for RBF or E for Pol) such that the training set is separated by the corresponding SVM Classification: Add instance as positive (negative). If the new set is not separable, the instance is negative (positive). If the set is separable in both cases, the instance is not classified.

VSSVM (3) Results with Leave-one-Out: Data SetParamsVSSVM c VSSVM a SVM a IG Heart-St-LPol, E=2, C= HepatitisPol, E=1.4, C= LaborRBF, G=0.02, C= SonarRBF, G=0.65, C=

Beating Existing Algorithms VSSVM give complete reliability on the coverage c Consider an arbitrary algorithm with accuracy a (e.g. the best algorithm until now) How can we beat that?

Beating Existing Algorithms (2) Answer: information gain! Algorithm CovA uses the algorithm Cov with coverage c and 100% accuracy on c and the algorithm A with accuracy a Theorem: the information gain of CovA is positive with respect to both the algorithms A and Cov!

Reliability gives Information Gain Theorem: Let reliability information be given such that and let E a and E r be the entropy given by the accuracy a and the reliabilities r i, then the information gain IG = E a – E r is positive.

Conclusions Version spaces are powerful! Implementation by SVM handles noisy training sets Implementation by SVM and RBF kernels prevents Version Space collapse by noisy training instances Unanimous voting results in reliability Reliability results in information gain

Future Research Extension to more than two classes Extension to the nonseparable case Reliability of training instances …

Example