Mitchell Kossoris, Catelyn Scholl, Zhi Zheng

Slides:



Advertisements
Similar presentations
Classification.. continued. Prediction and Classification Last week we discussed the classification problem.. – Used the Naïve Bayes Method Today..we.
Advertisements

Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Notes Sample vs distribution “m” vs “µ” and “s” vs “σ” Bias/Variance Bias: Measures how much the learnt model is wrong disregarding noise Variance: Measures.
Indian Statistical Institute Kolkata
Lecture 22: Evaluation April 24, 2010.
A (very) brief introduction to multivoxel analysis “stuff” Jo Etzel, Social Brain Lab
What is Statistical Modeling
Evaluation.
Rich Caruana Alexandru Niculescu-Mizil Presented by Varun Sudhakar.
Assessing and Comparing Classification Algorithms Introduction Resampling and Cross Validation Measuring Error Interval Estimation and Hypothesis Testing.
Lesson learnt from the UCSD datamining contest Richard Sia 2008/10/10.
Sparse vs. Ensemble Approaches to Supervised Learning
Sparse vs. Ensemble Approaches to Supervised Learning
MACHINE LEARNING 6. Multivariate Methods 1. Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Motivating Example  Loan.
Classification and Prediction: Basic Concepts Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 07: Cost-Sensitive Measures.
Today Evaluation Measures Accuracy Significance Testing
A Multivariate Biomarker for Parkinson’s Disease M. Coakley, G. Crocetti, P. Dressner, W. Kellum, T. Lamin The Michael L. Gargano 12 th Annual Research.
Predicting Income from Census Data using Multiple Classifiers Presented By: Arghya Kusum Das Arnab Ganguly Manohar Karki Saikat Basu Subhajit Sidhanta.
Evaluation – next steps
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Evaluating Hypotheses Reading: Coursepack: Learning From Examples, Section 4 (pp )
1 1 Slide Evaluation. 2 2 n Interactive decision tree construction Load segmentchallenge.arff; look at dataset Load segmentchallenge.arff; look at dataset.
Evaluating What’s Been Learned. Cross-Validation Foundation is a simple idea – “ holdout ” – holds out a certain amount for testing and uses rest for.
Machine learning system design Prioritizing what to work on
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensemble Methods: Bagging and Boosting
Introduction Use machine learning and various classifying techniques to be able to create an algorithm that can decipher between spam and ham s. .
Evaluating Predictive Models Niels Peek Department of Medical Informatics Academic Medical Center University of Amsterdam.
An Exercise in Machine Learning
CSSE463: Image Recognition Day 11 Due: Due: Written assignment 1 tomorrow, 4:00 pm Written assignment 1 tomorrow, 4:00 pm Start thinking about term project.
A Brief Introduction and Issues on the Classification Problem Jin Mao Postdoc, School of Information, University of Arizona Sept 18, 2015.
Evaluating Classifiers. Reading for this topic: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)
Does one size really fit all? Evaluating classifiers in a Bag-of-Visual-Words classification Christian Hentschel, Harald Sack Hasso Plattner Institute.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Data Mining Introduction to Classification using Linear Classifiers
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Machine Learning – Classification David Fenyő
Evaluating Classifiers
An Empirical Comparison of Supervised Learning Algorithms
Summary of “Efficient Deep Learning for Stereo Matching”
Machine Learning: Methodology Chapter
COMP1942 Classification: More Concept Prepared by Raymond Wong
Evaluating Results of Learning
Statistical Techniques
Classification Evaluation And Model Selection
When to engage in interaction – and how
SOCIAL COMPUTING Homework 3 Presentation
Natural Language Processing of Knee MRI Reports
Generalization ..
Categorizing networks using Machine Learning
Intro to Machine Learning
Experiments in Machine Learning
Phd Candidate Computational Physiology Lab University of Houston
Computer Vision Chapter 4
CSCI N317 Computation for Scientific Applications Unit Weka
Model Evaluation and Selection
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Intro to Machine Learning
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Model generalization Brief summary of methods
Analysis for Predicting the Selling Price of Apartments Pratik Nikte
Evaluating Classifiers
Assignment 1: Classification by K Nearest Neighbors (KNN) technique
Machine Learning: Methodology Chapter
COSC 4368 Intro Supervised Learning Organization
Credit Card Fraudulent Transaction Detection
Information Organization: Evaluation of Classification Performance
An introduction to Machine Learning (ML)
Presentation transcript:

Mitchell Kossoris, Catelyn Scholl, Zhi Zheng Classifying Physiological and Response Data to Detect Distracted Driving Events Mitchell Kossoris, Catelyn Scholl, Zhi Zheng

Distracted Driving Dataset Controlled driving simulator 68 volunteers Same highway 4 driving stimuli types: No stimuli Cognitive stimuli Emotional stimuli Sensorimotor stimuli

Distracted Driving Dataset Data collected: Speed Acceleration Brake force Steering Lane position Palm EDA Heart Rate Breathing Rate Gaze position

Processing Normalization Used to account for differences in each participant’s range of each feature E.g. Person 1 has a resting heart rate of 70 bpm vs person 2 with 80 bpm

Processing - Interpolation Forward fills missing data Segments of less than 10 linear-filled More data points to be analyzed Less data removed

Processing - Data Removal Continuous missing segments of data removed Longer than 10 seconds Fewer data inaccuracies over large segments

Processing - Rolling Mean Mean of each consecutive 10 second interval Dampen large differences caused by devices Reduces outliers

Processing - Balancing Ensures classifiers are not biased toward one class Data is split evenly between classes Trained and tested on balanced data

Classifiers K-Nearest Neighbor Support Vector Machines Random Forest Naive Bayes Neural Network

Evaluation Accuracy Baseline accuracy of our data, using a basic percentage of correctly-classified data points

Evaluation Mean Squared Error Mean squared error regression loss between ground truth and estimated target values

Evaluation F1 Scores Weighted average of precision and recall where its best value is 1 and worst value is 0

Evaluation Confusion Matrix Count of true positives, false positives, true negatives, and false negatives.

Evaluation K-Fold Cross Validation The dataset was split into k consecutive folds Each fold was then used once as a validation while the k-1 remaining folds formed the training set

Results Random forest: 99.38% accurate K-Nearest Neighbor: 93.9% accurate Support Vector Machines: 80.4% accurate Neural Net: 79.2% accurate Naive Bayes: 75.7% accurate