Pattern Recognition ->Machine Learning- >Data Analytics Supervised Learning Unsupervised Learning Semi-supervised Learning Reinforcement Learning.

Slides:



Advertisements
Similar presentations
ECG Signal processing (2)
Advertisements

Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
ONLINE ARABIC HANDWRITING RECOGNITION By George Kour Supervised by Dr. Raid Saabne.
Classification / Regression Support Vector Machines
Data Mining Classification: Alternative Techniques
An Introduction of Support Vector Machine
Support Vector Machines
Support vector machine
Machine learning continued Image source:
Transductive Reliability Estimation for Kernel Based Classifiers 1 Department of Computer Science, University of Ioannina, Greece 2 Faculty of Computer.
Announcements  Project proposal is due on 03/11  Three seminars this Friday (EB 3105) Dealing with Indefinite Representations in Pattern Recognition.
SVM Support Vectors Machines
Lecture 10: Support Vector Machines
Optimization Theory Primal Optimization Problem subject to: Primal Optimal Value:
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Support Vector Machines Exercise solutions Ata Kaban The University of Birmingham.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Step 3: Classification Learn a decision rule (classifier) assigning bag-of-features representations of images to different classes Decision boundary Zebra.
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
CSE 446 Perceptron Learning Winter 2012 Dan Weld Some slides from Carlos Guestrin, Luke Zettlemoyer.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Naive Bayes Classifier Christopher Gonzalez. Outline Bayes’ Theorem What is a Naive Bayes Classifier (NBC)? Why/when to use NBC? How does NBC work? Applications.
Making classifiers by supervised learning
Technological Educational Institute Of Crete Department Of Applied Informatics and Multimedia Intelligent Systems Laboratory Statistical Decision Making.
Kernels Usman Roshan CS 675 Machine Learning. Feature space representation Consider two classes shown below Data cannot be separated by a hyperplane.
An Introduction to Support Vector Machine (SVM)
12/7/20151 Math b Conditional Probability, Independency, Bayes Theorem.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
DATA MINING LECTURE 10b Classification k-nearest neighbor classifier
Supervised Machine Learning: Classification Techniques Chaleece Sandberg Chris Bradley Kyle Walsh.
Support Vector Machines Exercise solutions Ata Kaban The University of Birmingham.
Chapter 6. Classification and Prediction Classification by decision tree induction Bayesian classification Rule-based classification Classification by.
CSSE463: Image Recognition Day 14 Lab due Weds. Lab due Weds. These solutions assume that you don't threshold the shapes.ppt image: Shape1: elongation.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Knowledge-Based Nonlinear Support Vector Machine Classifiers Glenn Fung, Olvi Mangasarian & Jude Shavlik COLT 2003, Washington, DC. August 24-27, 2003.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
Neural networks and support vector machines
CSSE463: Image Recognition Day 14
CS 9633 Machine Learning Support Vector Machines
PREDICT 422: Practical Machine Learning
Machine Learning Fisher’s Criteria & Linear Discriminant Analysis
Support Vector Machines
Introduction to Machine Learning
Machine learning, pattern recognition and statistical data modelling
Geometrical intuition behind the dual problem
Nonparametric Methods: Support Vector Machines
Kernels Usman Roshan.
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
Linear Discriminators
Support Vector Machines Introduction to Data Mining, 2nd Edition by
Support Vector Machines
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
Statistical Learning Dong Liu Dept. EEIS, USTC.
CS 2750: Machine Learning Support Vector Machines
Minimax Probability Machine (MPM)
CSSE463: Image Recognition Day 14
CSSE463: Image Recognition Day 14
Machine Learning Week 3.
CSSE463: Image Recognition Day 14
Support Vector Machines
CSSE463: Image Recognition Day 14
Usman Roshan CS 675 Machine Learning
CSSE463: Image Recognition Day 15
CSSE463: Image Recognition Day 14
COSC 4368 Machine Learning Organization
Basics of ML Rohan Suri.
Machine Learning Support Vector Machine Supervised Learning
MAS 622J Course Project Classification of Affective States - GP Semi-Supervised Learning, SVM and kNN Hyungil Ahn
Presentation transcript:

Pattern Recognition ->Machine Learning- >Data Analytics Supervised Learning Unsupervised Learning Semi-supervised Learning Reinforcement Learning

Bayes Theorem Conditional probability: P(A|B), P(B|A) Marginal probability: P(A), P(B) Joint probability: P(A,B) P(AB) Bayes theorem: P(A), P(B), P(A|B) →P(B|A) P(A), P(B), P(B|A) →P(A|B)

Example P(C) = 0.01 (1%) P(pos|C) = 0.90 (90%) → 90% test is positive if you have C Sensitivity P(neg|~C) = 0.90 (90%) → 90% test is negative if you don’t have C Prior Specificity Question: if test is positive, the probability of having C ? P(C|pos) = ?

All people C

Naïve Bayes (Example: Text learning) Chris Sara LifeWorkFamily LifeWorkFamily P(Chris) = 0.5 P(Sara) = 0.5 Life Family Chris or Sara? Life Work Chris or Sara?

Naïve Bayes Prior: P(y 1 ), P(y 2 )…., P(y m ) Conditional prob: P(x 1,x 2,…,x n |y 1 ),…. P(x 1,x 2,…,x n |y m ) Solution: argmax(P(y j |x 1,x 2,…,x n ) ) j Naïve: assume independence of x 1, x 2,…, x n

Gaussian Naïve Bayes Conditional prob: P(x i |y) ~ N(µ y, σ y ) Implementation: Scikit learning-sklearn

Gaussian Naïve Bayes Self-driving car dataset

Gaussian Naïve Bayes

Support Vector Machine (SVM) Basic idea

Support Vector Machine (SVM) Basic idea

Support Vector Machine (SVM) Basic idea

Support Vector Machine (SVM) Basic idea Maximize distance to nearest point Maximize margin Support vector

Support Vector Machine (SVM) Basic idea Maximize robustness of classifier

Support Vector Machine (SVM) Basic idea

Support Vector Machine (SVM) Basic idea 1.Lower classifier error ! 2.Maximize margin

Support Vector Machine (SVM) Outlier outlier

Support Vector Machine (SVM) Outlier outlier

Support Vector Machine (SVM) Will SVMs work ? 1.Yes 2.No

Support Vector Machine (SVM) Trick Features x y SVM Label

Support Vector Machine (SVM) Is this linearly separable? x y SVM Label X 2 +y 2

Support Vector Machine (SVM) x y SVM Label Z=X 2 +y 2

Support Vector Machine (SVM) x y SVM Label Z=X 2 +y 2

Support Vector Machine (SVM) Add one more feature to make linearly separable 1.x 2 +y 2 2.|x| 3.|y|

Support Vector Machine (SVM) Add one or more non-linear features to make linearly separable Kernel Trick x, y x 1 x 2 x 3 x 4 x 5 Kernel Not separableSeparable Solution(linear boundary) Non-linear separable

Support Vector Machine (SVM) Nonlinear Kernels Sigmoid function

Implementation (sklearn SVM)

Naïve Bayes, Accuracy: Linear SVM, Accuracy: 0.92RBF SVM, Accuracy: 0.94