Classification of Breast Cancer Cells Using Artificial Neural Networks and Support Vector Machines Emmanuel Contreras Guzman.

Slides:



Advertisements
Similar presentations
ECG Signal processing (2)
Advertisements

Scott Wiese ECE 539 Professor Hu
Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Supervised Learning Recap
Machine Learning Week 1, Lecture 2. Recap Supervised Learning Data Set Learning Algorithm Hypothesis h h(x) ≈ f(x) Unknown Target f Hypothesis Set 5 0.
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Face Recognition & Biometric Systems Support Vector Machines (part 2)
Y.-J. Lee, O. L. Mangasarian & W.H. Wolberg
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Artificial Intelligence Statistical learning methods Chapter 20, AIMA (only ANNs & SVMs)
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Deep Belief Networks for Spam Filtering
CS 4700: Foundations of Artificial Intelligence
Greg GrudicIntro AI1 Introduction to Artificial Intelligence CSCI 3202: The Perceptron Algorithm Greg Grudic.
SVMs Finalized. Where we are Last time Support vector machines in grungy detail The SVM objective function and QP Today Last details on SVMs Putting it.
SVM Support Vectors Machines
1 Diagnosing Breast Cancer with Ensemble Strategies for a Medical Diagnostic Decision Support System David West East Carolina University Paul Mangiameli.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
1 Music Classification Using SVM Ming-jen Wang Chia-Jiu Wang.
SVMs, cont’d Intro to Bayesian learning. Quadratic programming Problems of the form Minimize: Subject to: are called “quadratic programming” problems.
Breast Cancer Diagnosis A discussion of methods Meena Vairavan.
Ranga Rodrigo April 5, 2014 Most of the sides are from the Matlab tutorial. 1.
Mathematical Programming in Support Vector Machines
Breast Cancer Diagnosis via Linear Hyper-plane Classifier Presented by Joseph Maalouf December 14, 2001 December 14, 2001.
197 Case Study: Predicting Breast Cancer Invasion with Artificial Neural Networks on the Basis of Mammographic Features MEDINFO 2004, T02: Machine Learning.
Neural Networks.
Prediction model building and feature selection with SVM in breast cancer diagnosis Cheng-Lung Huang, Hung-Chang Liao, Mu- Chen Chen Expert Systems with.
An Example of Course Project Face Identification.
WELCOME. Malay Mitra Lecturer in Computer Science & Application Jalpaiguri Polytechnic West Bengal.
Matlab Matlab Sigmoid Sigmoid Perceptron Perceptron Linear Linear Training Training Small, Round Blue-Cell Tumor Classification Example Small, Round Blue-Cell.
Breast Cancer Diagnosis via Neural Network Classification Jing Jiang May 10, 2000.
Non-Bayes classifiers. Linear discriminants, neural networks.
컴퓨터 과학부 김명재.  Introduction  Data Preprocessing  Model Selection  Experiments.
Neural Network Classification versus Linear Programming Classification in breast cancer diagnosis Denny Wibisono December 10, 2001.
DNA Microarray Data Analysis using Artificial Neural Network Models. by Venkatanand Venkatachalapathy (‘Venkat’) ECE/ CS/ ME 539 Course Project.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
Intelligent Numerical Computation1 Numerical Analysis MATLAB programming Numerical Methods Applications Contents.
Text Classification using Support Vector Machine Debapriyo Majumdar Information Retrieval – Spring 2015 Indian Statistical Institute Kolkata.
Supervised Machine Learning: Classification Techniques Chaleece Sandberg Chris Bradley Kyle Walsh.
Estimation of car gas consumption in city cycle with ANN Introduction  An ANN based approach to estimation of car fuel consumption  Multi Layer Perceptron.
An Effective Hybridized Classifier for Breast Cancer Diagnosis DISHANT MITTAL, DEV GAURAV & SANJIBAN SEKHAR ROY VIT University, India.
SVMs in a Nutshell.
Survival-Time Classification of Breast Cancer Patients and Chemotherapy Yuh-Jye Lee, Olvi Mangasarian & W. H. Wolberg UW Madison & UCSD La Jolla Computational.
語音訊號處理之初步實驗 NTU Speech Lab 指導教授: 李琳山 助教: 熊信寬
A Brief Introduction to Support Vector Machine (SVM) Most slides were from Prof. A. W. Moore, School of Computer Science, Carnegie Mellon University.
Hybrid Ant Colony Optimization-Support Vector Machine using Weighted Ranking for Feature Selection and Classification.
Neural networks and support vector machines
Artificial Neural Networks
ECE 5424: Introduction to Machine Learning
Mammogram Analysis – Tumor classification
Project 4: Facial Image Analysis with Support Vector Machines
Schizophrenia Classification Using
NEURAL NETWORK APPROACHES FOR AUTOMOBILE MPG PREDICTION
An Introduction to Support Vector Machines
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
Hyperparameters, bias-variance tradeoff, validation
By: Behrouz Rostami, Zeyun Yu Electrical Engineering Department
Predicting Breast Cancer Diagnosis From Fine-Needle Aspiration
Support Vector Machine _ 2 (SVM)
Somi Jacob and Christian Bach
Introduction to Radial Basis Function Networks
CSCE833 Machine Learning Lecture 9 Linear Discriminant Analysis
Fewer attributes are better if they are optimal
Linear Discrimination
CS+Social Good.
Adapted by Dr. Sarah from a talk by Dr. Catherine A. Gorini
Support Vector Machines 2
Presentation transcript:

Classification of Breast Cancer Cells Using Artificial Neural Networks and Support Vector Machines Emmanuel Contreras Guzman

The Motivation  Breast Cancer is the second most deadly type of cancer in women worldwide.  1.3 million women diagnosed worldwide  Nearly half-a-million women dying from this disease each year.  Very curable if diagnosed early.  Cervical Cancer is the third most deadly type of cancer in women worldwide.  Half-a-million women diagnosed.  250,000 women dying from this disease each year.  Also very curable if diagnosed early.

The Data Set Breast Cancer Wisconsin (Original) Data Set Samples collected from a minimally invasive fine-needle aspirate (FNA). 458 benign (65.5%) and 241 malignant (34.5%) 9 Features (scale from 1-10): Clump thickness Uniformity of cell size Uniformity of cell shape Marginal adhesion Single epithelial cell size Bare nuclei Bland chromatin Normal nucleoli Mitoses Cervical Cancer Data 5 Features: Amount of cytoplasm Nuclei count Nuclei shape Nuclei texture Nuclei area

Pre-Processing Unknown Samples  16 incomplete data samples - bare nuclei  Samples used for analysis: 683 Normalization  Normalize value to between 0 - 1

Artificial Neural Network Analysis  MATLAB driver program  Network Configurations:  1, 2 or 3 hidden layers  Each layer with 3, 5 or 7 perceptrons  70% training  15% for testing  15% for validation Transfer functions used:  Scaled Conjugate Gradient  Logistic  Tan Sigmoid Network retrained 50 times with random sample with replacement.

Support Vector Machine Analysis  MATLAB driver program  Parameters Tuned:  Kernels  Radial Basis Function  Linear Kernel  Polynomial Kernel - degrees 2 and 3  Box Constraint/C - support vector cost/penalty  1e-5 to 1e5 - increasing by factor of 10  Kernel Scale/Gamma - individual examples influence the hyperplane  1e-5 to 1e5 - increasing by factor of way cross validation

Artificial Neural Network Results Configurations with 97.4% accuracy  The Scaled Conjugate Gradient (SCG) without normalization and configuration: [5], [7], [7 7]  The Scaled Conjugate Gradient (SCG) with normalization and configuration: [5 5], [7]  The Logistic transfer function (logsig) without normalization and configuration: [5 5], [7]  The Logistic transfer function (logsig) with normalization and configuration: [5], [7]  The Tan-Sigmoid transfer function (tansig) without normalization:[5], [5 5], [7], [7 7]  The Tan-Sigmoid transfer function (tansig) with normalization:[7], [7 7], [7 7 7] Sensitivity: 98% Specificity: 96%

Support Vector Machine Results Maximum accuracy: 94.24%  2nd-order polynomial kernel  Box constraint of 1e-1  kernel scale (gamma) of 1  Sensitivity: 96.58%  Specificity: 91.43% Second most accurate configurations, accuracy: 94.13%  Radial Basis Function kernel, box constraint of 1, kernel scale set to auto,  Polynomial degree 3 with box constraint of 1e-1 and kernel scale set to auto.  Sensitivity: 96.35%  Specificity 91.02%

Conclusion  None of the ANN layer configurations with 3 perceptrons achieved a maximum accuracy of 97.4%.  ANN configurations with 3+ layers and perceptrons do not generalize well, and overfit the data.  Classification of breast cancer data using an ANN should be kept to one or two hidden layers of about 5 perceptrons in order to achieve the highest classification accuracy.  In comparison to the SVM, the neural network achieved higher accuracy by 3%, sensitivity by 2% and specificity by 5%.  A neural network appears to be a better algorithm for classifying the data.

Discussion  Other configurations for the ANN and SVM which were not analyzed.  More fine tuning of parameters.  Artificial Neural Network  Learning rates, transfer functions, more/less layers/perceptrons  Support Vector Machines  More Kernels  Different cost function  Removing “bare nuclei” feature and using all 699 samples.