Welcome everyone. Been to good sessions, exciting ones coming up.

Slides:



Advertisements
Similar presentations
Outlines Background & motivation Algorithms overview
Advertisements

Computational Learning An intuitive approach. Human Learning Objects in world –Learning by exploration and who knows? Language –informal training, inputs.
CHAPTER 9: Decision Trees
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
Classification / Regression Support Vector Machines
Linear Classifiers/SVMs
SOFT LARGE MARGIN CLASSIFIERS David Kauchak CS 451 – Fall 2013.
Support Vector Machines and Kernels Adapted from slides by Tim Oates Cognition, Robotics, and Learning (CORAL) Lab University of Maryland Baltimore County.
Supervised Learning Recap
A Quick Overview By Munir Winkel. What do you know about: 1) decision trees 2) random forests? How could they be used?
What is Statistical Modeling
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Financial time series forecasting using support vector machines Author: Kyoung-jae Kim 2003 Elsevier B.V.
Margins, support vectors, and linear programming Thanks to Terran Lane and S. Dreiseitl.
Recommendations via Collaborative Filtering. Recommendations Relevant for movies, restaurants, hotels…. Recommendation Systems is a very hot topic in.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
An Introduction to Support Vector Machines Martin Law.
Machine Learning CUNY Graduate Center Lecture 1: Introduction.
Data mining and machine learning A brief introduction.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Learning Theory Reza Shadmehr logistic regression, iterative re-weighted least squares.
An Introduction to Support Vector Machines (M. Law)
Overview of the final test for CSC Overview PART A: 7 easy questions –You should answer 5 of them. If you answer more we will select 5 at random.
Principal Component Analysis Machine Learning. Last Time Expectation Maximization in Graphical Models – Baum Welch.
Lecture 6 Spring 2010 Dr. Jianjun Hu CSCE883 Machine Learning.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
CS558 Project Local SVM Classification based on triangulation (on the plane) Glenn Fung.
Support Vector Machines. Notation Assume a binary classification problem. –Instances are represented by vector x   n. –Training examples: x = (x 1,
6.S093 Visual Recognition through Machine Learning Competition Image by kirkh.deviantart.com Joseph Lim and Aditya Khosla Acknowledgment: Many slides from.
LECTURE 20: SUPPORT VECTOR MACHINES PT. 1 April 11, 2016 SDS 293 Machine Learning.
The goal of the project is to predict the survival of passengers based off a set of data. To do this we train a prediction system.
Titanic and Decision Trees Supplement. Titanic Predictions and Decision Trees Variable Selection Approaches – Hypothesis Driven – Data Driven – Kitchen.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
CSSE463: Image Recognition Day 14
PREDICT 422: Practical Machine Learning
General-Purpose Learning Machine
Large Margin classifiers
Support Vector Machines
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Predict whom survived the Titanic Disaster
Estimating Link Signatures with Machine Learning Algorithms
Support Vector Machines (SVM)
K-means and Hierarchical Clustering
Support Vector Machines
Neuro-Computing Lecture 4 Radial Basis Function Network
Mathematical Foundations of BME Reza Shadmehr
CSSE463: Image Recognition Day 14
Application of Logistic Regression Model to Titanic Data
CSSE463: Image Recognition Day 14
CSSE463: Image Recognition Day 15
Lecture 6: Introduction to Machine Learning
CSSE463: Image Recognition Day 15
Advanced Artificial Intelligence Classification
CSSE463: Image Recognition Day 14
CSSE463: Image Recognition Day 15
Support Vector Machines and Kernels
CSSE463: Image Recognition Day 15
Nearest Neighbors CSC 576: Data Mining.
Decision Maths Unit 7 Sorting Algorithms 3. Shell Sort.
CSSE463: Image Recognition Day 15
INTRODUCTION TO Machine Learning 2nd Edition
CS639: Data Management for Data Science
Decision Tree  Decision tree is a popular classifier.
Exercise 1: Entering data into SPSS
Decision Tree  Decision tree is a popular classifier.
Exploratory Analysis Report
Support Vector Machines 2
Presentation transcript:

Welcome everyone. Been to good sessions, exciting ones coming up. My first SQL Bits session. Introduction to me.

Machine Learning The Maths Behind My introduction to ML. Data analytics world getting more interested in ML. Noticed a trend. Black box A shame Need to understand to effectively use. More frustration. Missing fun. Things I want to put right today. To show how quick, 3 algorithms. 3 popular, used before, use again.

Quick overview – classic classification problem. Variables & class.

Classification Supervised – we know if it’s dog or cat. height weight

Decision Tree Classification Widely used, popular, easy to understand.

Would I have survived?

All passengers 500 : 809 Died

All passengers 500 : 809 Men 142 : 640 Women 308 : 112 Children 50 : 57 Survived Died Died

All passengers 500 : 809 Men 142 : 640 Women 308 : 112 Children 50 : 57 1st Class 55 : 114 2nd Class 22 : 138 3rd Class 65 : 388 1st Class 124 : 4 2nd Class 85 : 12 3rd Class 99 : 96 1st Class 21 : 5 2nd Class 12 : 8 3rd Class 17 : 44 Died Died Died Survived Survived Survived Survived Survived Died

14% All passengers 500 : 809 Men 142 : 640 Women 308 : 112 Children 50 : 57 1st Class 55 : 114 2nd Class 22 : 138 3rd Class 65 : 388 1st Class 124 : 4 2nd Class 85 : 12 3rd Class 99 : 96 1st Class 21 : 5 2nd Class 12 : 8 3rd Class 17 : 44 Died Died Died Survived Survived Survived Survived Survived Died 14%

Classification Demo

Which variable do we split on? The best variable pclass age name cabin fare boat body Which variable do we split on? The best variable parch ticket survived embarked home.dest sex sibsp

The best variable pclass age name cabin fare boat body parch ticket survived embarked home.dest sex sibsp

𝐺𝑖𝑛𝑖=1− 𝑖=1 𝐶 ( 𝑝 𝑖 ) 2 The Gini Index 𝐺𝑖𝑛𝑖=1− 𝑖=1 𝐶 ( 𝑝 𝑖 ) 2 Gini index originally used to calculate income inequality. Derived 1912 – present. Low Gini index is desired.

𝑔 𝑃 𝑠 = 1 − 500 1309 2 − 809 1309 2 ≈0.4721 𝑔 𝑃 𝑚 = 1 − 161 843 2 − 682 843 2 ≈0.3090 𝑔 𝑃 𝑓 = 1 − 339 466 2 − 127 466 2 ≈0.3965 𝐺=0.4721 − 843 1309 ×0.3090 − 466 1309 ×0.3965 ≈0.132

sex = 0.132 sex = male sex = female fare pclass sibsp parch age embarked age pclass fare parch embarked sibsp pclass fare sibsp embarked parch age

pclass 1 2 3 1.5 2.5 Continue recursively, until sub group reaches min size, or no improvement. Greedy. Cross validation performed to trim.

Used a lot in data analytics. Similar to classification.

Regression profit time

Regression Support Vector SVR similar to SVM, getting more popular. Works well with non-linear.

How much is my car worth?

𝑟= (𝑥− 𝑥 )(𝑦− 𝑦 ) (𝑥− 𝑥 ) 2 (𝑦− 𝑦 ) 2 𝑎= 𝑦 - b 𝑥 15,000- 𝑦=𝑎+𝑏𝑥 𝑟= (𝑥− 𝑥 )(𝑦− 𝑦 ) (𝑥− 𝑥 ) 2 (𝑦− 𝑦 ) 2 𝑎= 𝑦 - b 𝑥 𝑏=𝑟 𝜎 𝑦 𝜎 𝑥 price (€) Uses a Kernel function, this case Radial Basis function Kernel. 0 - | 1980 | 2015 year

10,000- 15,000- price (€) 3,000 - 0 - | 2003 | 1980 | 2011 | 2015 year

15,000- price (€) 0 - | 1980 | 2015 year

price (€) year 15,000- 𝑓 𝑥 + 𝜀 𝑓 𝑥 − 𝜀 0 - | 1980 | 2015 Outside margin points penalised. Principle of maximal margin. Important to prevent overfitting. 0 - | 1980 | 2015 year

Regression Demo

200,000- price (€) 0 - | 1980 | 2015 year

A third dimension …and so on forth third fifth sixth

Netflix clustering. Similar to HBO.

Clustering

K-Means Clustering K-Means – most popular, released decades ago, improvements.

What new music can I discover?

Euclidean distance= ( 𝑥 2 − 𝑥 1 ) 2 + ( 𝑦 2 − 𝑦 1 ) 2

Clustering Demo Better algorithms for problem: Gaussian mixture models, hierarchical clustering.

We’ve gone through 3 algorithms in 1 hour. List of good sources…

github.com/matt-willis https://s3.amazonaws.com/assets.datacamp.com/blog_assets/PySpark_SQL_Cheat_Sheet_Python.pdf www.listendata.com alex.smola.org/papers/2004/SmoSch04.pdf data-flair.training/blogs www.analyticsvidhya.com www.kdnuggets.com cowlet.org github.com/matt-willis hackernoon.com r4ds.had.co.nz blogs.adatis.co.uk towardsdatascience.com archive.ics.uci.edu/ml/datasets.html www.r-bloggers.com elitedatascience.com www.datacamp.com www.datasciencecentral.com trevorstephens.com kernelsvm.tripod.com https://runawayhorse001.github.io/LearningApacheSpark/clustering.html www.coursera.org www.learnbymarketing.com wiki.icub.org/images/8/82/OnlineSVR_Thesis.pdf

matt.willis@adatis.bg Email me. Speak to me. I hope you’ve learnt something. More importantly, changed your approach.