Classify A to Z Problem Statement Technical Approach Results Dataset

Slides:



Advertisements
Similar presentations
Neural Networks Chapter Feed-Forward Neural Networks.
Advertisements

CONTENT-BASED BOOK RECOMMENDING USING LEARNING FOR TEXT CATEGORIZATION TRIVIKRAM BHAT UNIVERSITY OF TEXAS AT ARLINGTON DATA MINING CSE6362 BASED ON PAPER.
What is the Best Multi-Stage Architecture for Object Recognition Kevin Jarrett, Koray Kavukcuoglu, Marc’ Aurelio Ranzato and Yann LeCun Presented by Lingbo.
Classifiers, Part 3 Week 1, Video 5 Classification  There is something you want to predict (“the label”)  The thing you want to predict is categorical.
Machine Learning CS 165B Spring 2012
Slide Image Retrieval: A Preliminary Study Guo Min Liew and Min-Yen Kan National University of Singapore Web IR / NLP Group (WING)
Kuan-Chuan Peng Tsuhan Chen
Prediction of Molecular Bioactivity for Drug Design Experiences from the KDD Cup 2001 competition Sunita Sarawagi, IITB
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Text Categorization With Support Vector Machines: Learning With Many Relevant Features By Thornsten Joachims Presented By Meghneel Gore.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
COMP24111: Machine Learning Ensemble Models Gavin Brown
Competition II: Springleaf Sha Li (Team leader) Xiaoyan Chong, Minglu Ma, Yue Wang CAMCOS Fall 2015 San Jose State University.
BAYESIAN LEARNING. 2 Bayesian Classifiers Bayesian classifiers are statistical classifiers, and are based on Bayes theorem They can calculate the probability.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Convectional Neural Networks
A Simple Approach for Author Profiling in MapReduce
Big data classification using neural network
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
A Smart Tool to Predict Salary Trends of H1-B Holders
Tenacious Deep Learning
References and Related Work
Bagging and Random Forests
The Relationship between Deep Learning and Brain Function
Deep Learning Amin Sobhani.
Theory and Applications
Data Mining, Neural Network and Genetic Programming
Observations by Dance Move
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Estimating Link Signatures with Machine Learning Algorithms
COMP61011 : Machine Learning Ensemble Models
Basic machine learning background with Python scikit-learn
Natural Language Processing of Knee MRI Reports
NBA Draft Prediction BIT 5534 May 2nd 2018
Vincent Granville, Ph.D. Co-Founder, DSC
Machine Learning & Data Science
Categorizing networks using Machine Learning
Feature Film Features: Applying machine learning to movie genre identification  CSCI 5622 Group L: Grant Baker, John Dinkel, Derek Gorthy, Jeffrey Maierhofer,
Volume 10, Issue 6, Pages (June 2018)
A Comparative Study of Convolutional Neural Network Models with Rosenblatt’s Brain Model Abu Kamruzzaman, Atik Khatri , Milind Ikke, Damiano Mastrandrea,
iSRD Spam Review Detection with Imbalanced Data Distributions
Machine Learning 101 Intro to AI, ML, Deep Learning
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Lecture: Deep Convolutional Neural Networks
Machine Learning Interpretability
Tuning CNN: Tips & Tricks
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Analysis of Trained CNN (Receptive Field & Weights of Network)
Somi Jacob and Christian Bach
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Model generalization Brief summary of methods
Overview of deep learning
MTBI Personality Predictor using ML
Presentation By: Eryk Helenowski PURE Mentor: Vincent Bindschaedler
Heterogeneous convolutional neural networks for visual recognition
Predicting Loan Defaults
Reuben Feinman Research advised by Brenden Lake
Department of Computer Science Ben-Gurion University of the Negev
Semi-Supervised Learning
Introduction to Sentiment Analysis
Trusting Machine Learning Algorithms for Safeguards Applications
THE ASSISTIVE SYSTEM SHIFALI KUMAR BISHWO GURUNG JAMES CHOU
Credit Card Fraudulent Transaction Detection
Adrian E. Gonzalez , David Parra Department of Computer Science
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
An introduction to Machine Learning (ML)
Presentation transcript:

Classify A to Z Problem Statement Technical Approach Results Dataset ['item_name_stem'] ['item_name_stem'] Classify A to Z Jason Jung, Kate Larsen, Ethen Liu, Santiago Olivar | MSiA 490-30 Deep Learning | Spring 2017 | Northwestern University Applying Machine Learning to Product Categorization: http://cs229.stanford.edu/proj2011/LinShankar-Applying%20Machine%20Learning%20to%20Product%20Categorization.pdf Problem Statement Technical Approach Results INSERT LOSS CURVE Goal: Given the product’s text description and image from Amazon.com, classify its category Audience: Useful for companies that sell products online and wish to classify or organize them. Unlike large companies like Amazon, who can afford to use expensive software like IBM E-Commerce to organize product information, our software can be used by smaller companies Challenge: Each category includes hundreds of subcategories which can be very different from one another. Thus, the model has to learn the abstract concept of categories Similar works: Students in Stanford University classified the similar dataset except with just text descriptions by using algorithms like Naive Bayes and tree classifiers that produced 86% accuracy Training Text Image Stemmed words TF-IDF Trained machine learning models: SVM Naive Bayes Logistic Regression Random Forest Utilized Keras’ Inception network model that was pre-trained on the Imagenet dataset to perform transfer learning (more details in the section below) Ensemble Combined the image and text models’ outputs by using a weighted average of their prediction Explanation Used LIME technique (Local Interpretable Model agnostic Explanations) to explain the predictions of the classifier LIME shows the model focused on the cat (green) rather than the mouse or background (red) The accuracy measure on a hold out test set for the text model was 96.2% and 87.1% for the image model, while an ensemble (0.7/0.3 weight for text/image) resulted in a slight improvement of 96.8% Strength of Text Model: The R2D2 usb drive below shows how the text model correctly classifies the product focusing on the orange words below, while the image model performs poorly by focusing on the green parts of the image Dataset Raw Data & Features Extracted: Downloaded in json format from a publicly available dataset that labeled categories for different Amazon products Extracted text fields of title and descriptions and URL to download the images Challenges: Encountered few preprocessing issues due to clean json format Found some images were misclassified or overlapped categories Removed the Sports & Outdoors category due to troublesome images even for human accuracy Data Size: Balanced the classes to 35,000 per category leading to 140,000 observations in total Clothing, Shoes, & Jewelry Grocery & Gourmet Food Home & Kitchen Electronics LaCroix Sparkling Water, Orange, 12-Ounce Cans (Pack of 12) Network Architecture Comparison Inception Module Strength of Image Model: The headphones below show how the image model performs well by focusing on earpieces, while the text model performs poorly by focusing on the stemmed word “cherri.” Methodology Research Left: State-of-art Neural Network Architecture The x-axis shows the number of network parameters/operations and the y-axis shows network performance on the Imagenet challenge This shows that Inception-v3 network strikes a good balance between model performance and complexity Right: Inception Module The Inception network’s success was due to the Inception module At a high level, because input features are correlated, redundancy can be removed by combining them appropriately with the 1x1 convolutions before feeding them to more expensive 3x3 or 5x5 convolutions Conclusion References and Related Work Text alone is very powerful for predicting item category, however performance is best achieved when ensembling the text and image model with a 70% - 30% respective weight One limitation of our approach was that we structured our model to classify at a high level of category. Thus, a possible improvement to the project includes classifying at a more granular level such as classifying TV, cell phone as oppose to Electronics LIME is handy for visual explanation of black-box model’s decision An Analysis Of Deep Neural Network Models For Practical Applications: https://arxiv.org/pdf/1605.07678.pdf Dataset: http://jmcauley.ucsd.edu/data/amazon/ Going deeper with convolutions: https://arxiv.org/pdf/1409.4842.pdf LIME: https://github.com/marcotcr/lime