The Little Big Data Showdown ECE 8110 Pattern Recognition and Machine Learning Temple University Christian Ward.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

ECG Signal processing (2)
Relevant characteristics extraction from semantically unstructured data PhD title : Data mining in unstructured data Daniel I. MORARIU, MSc PhD Supervisor:
Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
Integrated Instance- and Class- based Generative Modeling for Text Classification Antti PuurulaUniversity of Waikato Sung-Hyon MyaengKAIST 5/12/2013 Australasian.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Support Vector Machines and Kernels Adapted from slides by Tim Oates Cognition, Robotics, and Learning (CORAL) Lab University of Maryland Baltimore County.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
K Means Clustering , Nearest Cluster and Gaussian Mixture
CS395: Visual Recognition Spatial Pyramid Matching Heath Vinicombe The University of Texas at Austin 21 st September 2012.
Large Scale Manifold Transduction Michael Karlen Jason Weston Ayse Erkan Ronan Collobert ICML 2008.
Discriminative and generative methods for bags of features
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Lecture Notes for CMPUT 466/551 Nilanjan Ray
MCS 2005 Round Table In the context of MCS, what do you believe to be true, even if you cannot yet prove it?
An Introduction to Kernel-Based Learning Algorithms K.-R. Muller, S. Mika, G. Ratsch, K. Tsuda and B. Scholkopf Presented by: Joanna Giforos CS8980: Topics.
Reduced Support Vector Machine
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
Lecture 14: Classification Thursday 18 February 2010 Reading: Ch – 7.19 Last lecture: Spectral Mixture Analysis.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
CS Instance Based Learning1 Instance Based Learning.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
July 11, 2001Daniel Whiteson Support Vector Machines: Get more Higgs out of your data Daniel Whiteson UC Berkeley.
Outline Separating Hyperplanes – Separable Case
This week: overview on pattern recognition (related to machine learning)
earthobs.nr.no Land cover classification of cloud- and snow-contaminated multi-temporal high-resolution satellite images Arnt-Børre Salberg and.
Hyperspectral Imaging Alex Chen 1, Meiching Fong 1, Zhong Hu 1, Andrea Bertozzi 1, Jean-Michel Morel 2 1 Department of Mathematics, UCLA 2 ENS Cachan,
Stochastic k-Neighborhood Selection for Supervised and Unsupervised Learning University of Toronto Machine Learning Seminar Feb 21, 2013 Kevin SwerskyIlya.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 24 – Classifiers 1.
1 Comparison of Principal Component Analysis and Random Projection in Text Mining Steve Vincent April 29, 2004 INFS 795 Dr. Domeniconi.
Experiments in Machine Learning COMP24111 lecture 5 Accuracy (%) A BC D Learning algorithm.
Csc Lecture 7 Recognizing speech. Geoffrey Hinton.
Transfer Learning with Applications to Text Classification Jing Peng Computer Science Department.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
A Toolkit for Remote Sensing Enviroinformatics Clustering Fazlul Shahriar, George Bonev Advisors: Michael Grossberg, Irina Gladkova, Srikanth Gottipati.
Classification Derek Hoiem CS 598, Spring 2009 Jan 27, 2009.
Interactive Learning of the Acoustic Properties of Objects by a Robot
Principal Component Analysis Machine Learning. Last Time Expectation Maximization in Graphical Models – Baum Welch.
Principal Manifolds and Probabilistic Subspaces for Visual Recognition Baback Moghaddam TPAMI, June John Galeotti Advanced Perception February 12,
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
DATA MINING WITH CLUSTERING AND CLASSIFICATION Spring 2007, SJSU Benjamin Lam.
An Approximate Nearest Neighbor Retrieval Scheme for Computationally Intensive Distance Measures Pratyush Bhatt MS by Research(CVIT)
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Introduction to Pattern Recognition (การรู้จํารูปแบบเบื้องต้น)
Nonparametric Density Estimation Riu Baring CIS 8526 Machine Learning Temple University Fall 2007 Christopher M. Bishop, Pattern Recognition and Machine.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Speech Lab, ECE, State University of New York at Binghamton  Classification accuracies of neural network (left) and MXL (right) classifiers with various.
CS378 Final Project The Netflix Data Set Class Project Ideas and Guidelines.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
Objectives: Terminology Components The Design Cycle Resources: DHS Slides – Chapter 1 Glossary Java Applet URL:.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
KERNELS AND PERCEPTRONS. The perceptron A B instance x i Compute: y i = sign(v k. x i ) ^ y i ^ If mistake: v k+1 = v k + y i x i x is a vector y is -1.
9.913 Pattern Recognition for Vision Class9 - Object Detection and Recognition Bernd Heisele.
ECE 471/571 – Lecture 22 Support Vector Machine 11/24/15.
A Parallel Mixture of SVMs for Very Large Scale Problems Ronan Collobert Samy Bengio Yoshua Bengio Prepared : S.Y.C. Neural Information Processing Systems,
Computational Intelligence: Methods and Applications Lecture 24 SVM in the non-linear case Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
SVMs in a Nutshell.
CSSE463: Image Recognition Day 14 Lab due Weds. Lab due Weds. These solutions assume that you don't threshold the shapes.ppt image: Shape1: elongation.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
CMPS 142/242 Review Section Fall 2011 Adapted from Lecture Slides.
Spectral Methods for Dimensionality
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
Juweek Adolphe Zhaoyu Li Ressi Miranda Dr. Shang
ECE539 final project Instructor: Yu Hen Hu Fall 2005
Minimax Probability Machine (MPM)
Presentation transcript:

The Little Big Data Showdown ECE 8110 Pattern Recognition and Machine Learning Temple University Christian Ward

What Data Is This? One set stationary, one set temporal Dimension’s greater than three Small test sets Let’s go with SVM, GM and KNN as our algorithms

Stationary Data with SVMs

SVM Continued, Correlation Matrix

SVM Continued, Correlation Probability

SVM, Accurate? Below: Baseline accuracy with best kernel Right Top: Training Scaled Right Bottom: Training and Testing Scaled Best effort 14% accuracy overall

KNN, How Many Neighbors Is Too Many? KNN unscaled inputs: k=7 with 60% accuracy This makes slight sense given the scaling from the SVM models and the covariance matrix. KNN scaled inputs: k=5 with 43% accuracy

Gaussian Mixtures, Are Only Hope? Distribution sizes from 1 to 379 Best error rate found around 54% near 60 clusters. Attempts to feed the algorithm with the means from KMEANS failed due to my inability to properly generate covariance matrix to meet the function’s needs.

Data Set 2, Too Much Data

SVM, Results Top Left: Baseline Results Top Right: Scaled Training Results Bottom: Scaled Training and Testing Data Results Accuracy for each specific classifier is quite high, but there are excessive errors and overlaps outside the correct classes. This results in an overall accuracy of only 60 percent.

KNN, Too Few Neighbors? This time the various metrics perform with the same results, which lends further credence to the difference between the two data sets.

Gaussian Mixture Models on Temporal The Gaussian approach fails quite hard given the temporal nature of the data and no trend develops as the number of clusters increases. This is not too surprising given that the 39x39 element covariance matrix shows quite a bit of