Breast Cancer Diagnosis via Linear Hyper-plane Classifier Presented by Joseph Maalouf December 14, 2001 December 14, 2001.

Slides:



Advertisements
Similar presentations
ECG Signal processing (2)
Advertisements

Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Classification / Regression Support Vector Machines
CHAPTER 10: Linear Discrimination
An Introduction of Support Vector Machine
Support Vector Machines
Support vector machine
The Disputed Federalist Papers : SVM Feature Selection via Concave Minimization Glenn Fung and Olvi L. Mangasarian CSNA 2002 June 13-16, 2002 Madison,
Y.-J. Lee, O. L. Mangasarian & W.H. Wolberg
Support Vector Machines
Support Vector Machines Joseph Gonzalez TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A AA A AA.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
1 Classification: Definition Given a collection of records (training set ) Each record contains a set of attributes, one of the attributes is the class.
Rutgers CS440, Fall 2003 Support vector machines Reading: Ch. 20, Sec. 6, AIMA 2 nd Ed.
Dual Problem of Linear Program subject to Primal LP Dual LP subject to ※ All duality theorems hold and work perfectly!
Tutorial 12 Linear programming Quadratic programming.
Active Learning with Support Vector Machines
Classification Problem 2-Category Linearly Separable Case A- A+ Malignant Benign.
Support Vector Machines and Kernel Methods
Unconstrained Optimization Problem
Lecture outline Support vector machines. Support Vector Machines Find a linear hyperplane (decision boundary) that will separate the data.
SVM (Support Vector Machines) Base on statistical learning theory choose the kernel before the learning process.
Survival-Time Classification of Breast Cancer Patients DIMACS Workshop on Data Mining and Scalable Algorithms August 22-24, Rutgers University Y.-J.
Breast Cancer Diagnosis A discussion of methods Meena Vairavan.
Survival-Time Classification of Breast Cancer Patients DIMACS Workshop on Data Mining and Scalable Algorithms August 22-24, Rutgers University Y.-J.
Mathematical Programming in Support Vector Machines
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Medical Diagnosis via Genetic Programming Project #2 Artificial Intelligence: Biointelligence Computational Neuroscience Connectionist Modeling of Cognitive.
Support Vector Machine & Image Classification Applications
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
Support Vector Machine (SVM) Based on Nello Cristianini presentation
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Breast Cancer Diagnosis via Neural Network Classification Jing Jiang May 10, 2000.
CS 478 – Tools for Machine Learning and Data Mining SVM.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
SUPPORT VECTOR MACHINES. Intresting Statistics: Vladmir Vapnik invented Support Vector Machines in SVM have been developed in the framework of Statistical.
An Introduction to Support Vector Machine (SVM)
Neural Network Classification versus Linear Programming Classification in breast cancer diagnosis Denny Wibisono December 10, 2001.
Prognostic Prediction of Breast Cancer Using C5 Sakina Begum May 1, 2001.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
CSE4334/5334 DATA MINING CSE4334/5334 Data Mining, Fall 2014 Department of Computer Science and Engineering, University of Texas at Arlington Chengkai.
Support Vector Machine Debapriyo Majumdar Data Mining – Fall 2014 Indian Statistical Institute Kolkata November 3, 2014.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Support Vector Machines. Notation Assume a binary classification problem. –Instances are represented by vector x   n. –Training examples: x = (x 1,
Text Classification using Support Vector Machine Debapriyo Majumdar Information Retrieval – Spring 2015 Indian Statistical Institute Kolkata.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Survival-Time Classification of Breast Cancer Patients DIMACS Workshop on Data Mining and Scalable Algorithms August 22-24, Rutgers University Y.-J.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
SVMs in a Nutshell.
Survival-Time Classification of Breast Cancer Patients and Chemotherapy Yuh-Jye Lee, Olvi Mangasarian & W. H. Wolberg UW Madison & UCSD La Jolla Computational.
Support Vector Machine: An Introduction. (C) by Yu Hen Hu 2 Linear Hyper-plane Classifier For x in the side of o : w T x + b  0; d = +1; For.
CSSE463: Image Recognition Day 14 Lab due Weds. Lab due Weds. These solutions assume that you don't threshold the shapes.ppt image: Shape1: elongation.
Generalization Error of pac Model  Let be a set of training examples chosen i.i.d. according to  Treat the generalization error as a r.v. depending on.
Machine Learning and Data Mining: A Math Programming- Based Approach Glenn Fung CS412 April 10, 2003 Madison, Wisconsin.
Classification of Breast Cancer Cells Using Artificial Neural Networks and Support Vector Machines Emmanuel Contreras Guzman.
Support Vector Machines Reading: Textbook, Chapter 5 Ben-Hur and Weston, A User’s Guide to Support Vector Machines (linked from class web page)
Mammogram Analysis – Tumor classification
Glenn Fung, Murat Dundar, Bharat Rao and Jinbo Bi
Geometrical intuition behind the dual problem
Support Vector Machines
Support Vector Machines (SVM)
An Introduction to Support Vector Machines
Machine Learning Week 3.
Somi Jacob and Christian Bach
Other Classification Models: Support Vector Machine (SVM)
COSC 4368 Machine Learning Organization
Adapted by Dr. Sarah from a talk by Dr. Catherine A. Gorini
Support Vector Machines
Presentation transcript:

Breast Cancer Diagnosis via Linear Hyper-plane Classifier Presented by Joseph Maalouf December 14, 2001 December 14, 2001

Problem Description Breast Cancer is second only to lung cancer as a tumor-related cause of death in women. Although there exists reasonable agreement on the criteria for benign/malignant (Benign means that the lump or other problem was not cancer, and Malignant means that the tissue does contain cancer cells) diagnoses using fine needle aspirate (FNA) and mammogram data, the application of these criteria are often quite subjective and a time consuming task for the physician. The idea of the project is to come up with a discriminant function (a separating plane in this case) to determine if an unknown sample is benign or malignant. The project will use the Wisconsin Diagnosis Breast Cancer Database (WDBC) made publicly available by Dr.William H.Wolberg of the Department of Surgery of the University of Wisconsin Medical School. There are 569 samples (357 benign, 212 malignant) with 32 attributes (patient ID, diagnosis type, 30 real-valued input features). The most effective pair of the attributes in determining a correct diagnosis will be determined and used to plot all the testing set points on a two dimensional figure.

Solution Methods Formally, given two sets B and M in the 30-dimensional real space R 30, we wish to construct a discriminant function f, from R 30 into R, such that: f (x) > 0 => x  M, f (x)  0 => x  B, Two approaches will be used to find the optimal hyper-plane: Linear Optimization Quadratic Optimization Linear Optimization Problem Formulation: The discriminant function f can be given by : f(x) = w’x - , determining a plane w’x =  that separates, to the extent possible, malignant points from benign ones in R 30. It remains to show how to determine w  R 30 and   R from the training data. If we let the sets of m points, M, be represented by a matrix M  R mxn and the set of k points, B, be represented by a matrix B  R kxn,

Solution Methods (continued) then the problem becomes one of choosing w and  to: min w,  (1/m)*|| (-M*w + e*  +e) + || 1 + (1/k)*|| (B*w - e *  +e) + || 1 This can be implemented using matlab optimization tool box lp. Quadratic Optimization Problem Formulation: The function f is given by : f(x) = w’x + b. For x i being a support vector, For x i  M  d i = 1, f(x i ) = w ’ * x i + b =  |w|  w o ’ * x i + b o = 1 For x i  B  d i =  1, f(x i ) = w ’ * x i + b =  |w|  w o ’ * x i + b o =  1 The objective is to find w and b such that  (w o ) = (w o ’ * w o )/2 is minimized subject to the constraints: d i * (w o ’ * x i + b o )  1, where 1  i  N (number of support vectors) To implement the quadratic optimization solution, I ’ ll use the OSU support vector machine.

Results The linear optimization method determined a correct diagnosis with a success rate of 97.2 %.. For the quadratic optimization method, I tried different types of kernels with different parameters. oIt seems as far, the SVM classifier with polynomial kernel of order 4 gives a better result, with a success rate of 91.4 %. oTraining a SVM classifier involves a time consuming. The following plot shows a figure of one of the most effective pair of the attributes in determining a correct diagnosis.