Prediction of Malignancy of Ovarian Tumors Using Least Squares Support Vector Machines C. Lu 1, T. Van Gestel 1, J. A. K. Suykens 1, S. Van Huffel 1, I.

Slides:



Advertisements
Similar presentations
SISTA seminar Feb 28, 2002 Preoperative Prediction of Malignancy of Ovarian Tumors Using Least Squares Support Vector Machines C. Lu 1, T. Van Gestel 1,
Advertisements

AIME03, Oct 21, 2003 Classification of Ovarian Tumors Using Bayesian Least Squares Support Vector Machines C. Lu 1, T. Van Gestel 1, J. A. K. Suykens.
ECG Signal processing (2)
Brief introduction on Logistic Regression
Learning Algorithm Evaluation
1 Statistical Modeling  To develop predictive Models by using sophisticated statistical techniques on large databases.
Supervised Learning Recap
Model Assessment, Selection and Averaging
The Center for Signal & Image Processing Georgia Institute of Technology Kernel-Based Detectors and Fusion of Phonological Attributes Brett Matthews Mark.
Chapter 4: Linear Models for Classification
PhD Hearing (Oct 15, 2003) Predictive Computer Models for Medical Classification Problems Predictive Computer Models for Medical Classification Problems.
Relational Learning with Gaussian Processes By Wei Chu, Vikas Sindhwani, Zoubin Ghahramani, S.Sathiya Keerthi (Columbia, Chicago, Cambridge, Yahoo!) Presented.
Introduction to Predictive Learning
Classification and risk prediction Usman Roshan. Disease risk prediction What is the best method to predict disease risk? –We looked at the maximum likelihood.
Reduced Support Vector Machine
Data mining and statistical learning, lecture 5 Outline  Summary of regressions on correlated inputs  Ridge regression  PCR (principal components regression)
Machine Learning CMPT 726 Simon Fraser University
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
Support Vector Machines Classification
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Study of Sparse Online Gaussian Process for Regression EE645 Final Project May 2005 Eric Saint Georges.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Today Evaluation Measures Accuracy Significance Testing
PhD defense C. LU 25/01/ Probabilistic Machine Learning Approaches to Medical Classification Problems Probabilistic Machine Learning Approaches to.
UOG Journal Club: January 2013
Efficient Model Selection for Support Vector Machines
沈致远. Test error(generalization error): the expected prediction error over an independent test sample Training error: the average loss over the training.
Topics on Final Perceptrons SVMs Precision/Recall/ROC Decision Trees Naive Bayes Bayesian networks Adaboost Genetic algorithms Q learning Not on the final:
Introduction to variable selection I Qi Yu. 2 Problems due to poor variable selection: Input dimension is too large; the curse of dimensionality problem.
by B. Zadrozny and C. Elkan
Alignment and classification of time series gene expression in clinical studies Tien-ho Lin, Naftali Kaminski and Ziv Bar-Joseph.
A Comparative Study on Variable Selection for Nonlinear Classifiers C. Lu 1, T. Van Gestel 1, J. A. K. Suykens 1, S. Van Huffel 1, I. Vergote 2, D. Timmerman.
Reduced the 4-class classification problem into 6 pairwise binary classification problems, which yielded the conditional pairwise probability estimates.
Statistical Modeling with SAS/STAT Cheng Lei Department of Electrical and Computer Engineering University of Victoria April 9, 2015.
A New Subspace Approach for Supervised Hyperspectral Image Classification Jun Li 1,2, José M. Bioucas-Dias 2 and Antonio Plaza 1 1 Hyperspectral Computing.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
NFL Play Predictions Will Burton, NCSU Industrial Engineering 2015
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
EMBC2001 Using Artificial Neural Networks to Predict Malignancy of Ovarian Tumors C. Lu 1, J. De Brabanter 1, S. Van Huffel 1, I. Vergote 2, D. Timmerman.
Today Ensemble Methods. Recap of the course. Classifier Fusion
An Ensemble of Three Classifiers for KDD Cup 2009: Expanded Linear Model, Heterogeneous Boosting, and Selective Naive Bayes Members: Hung-Yi Lo, Kai-Wei.
Evaluating Results of Learning Blaž Zupan
Some Aspects of Bayesian Approach to Model Selection Vetrov Dmitry Dorodnicyn Computing Centre of RAS, Moscow.
Support Vector Machines in Marketing Georgi Nalbantov MICC, Maastricht University.
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Reservoir Uncertainty Assessment Using Machine Learning Techniques Authors: Jincong He Department of Energy Resources Engineering AbstractIntroduction.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Machine Learning 5. Parametric Methods.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
Gaussian Process and Prediction. (C) 2001 SNU CSE Artificial Intelligence Lab (SCAI)2 Outline Gaussian Process and Bayesian Regression  Bayesian regression.
Blackbox classifiers for preoperative discrimination between malignant and benign ovarian tumors C. Lu 1, T. Van Gestel 1, J. A. K. Suykens 1, S. Van Huffel.
Introduction Background Medical decision support systems based on patient data and expert knowledge A need to analyze the collected data in order to draw.
Support Vector Machines Optimization objective Machine Learning.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
A distributed PSO – SVM hybrid system with feature selection and parameter optimization Cheng-Lung Huang & Jian-Fan Dun Soft Computing 2008.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Day 17: Duality and Nonlinear SVM Kristin P. Bennett Mathematical Sciences Department Rensselaer Polytechnic Institute.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Hybrid Ant Colony Optimization-Support Vector Machine using Weighted Ranking for Feature Selection and Classification.
CS Statistical Machine learning Lecture 7 Yuan (Alan) Qi Purdue CS Sept Acknowledgement: Sargur Srihari’s slides.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Machine Learning – Classification David Fenyő
In Search of the Optimal Set of Indicators when Classifying Histopathological Images Catalin Stoean University of Craiova, Romania
Evaluating Results of Learning
Pattern Recognition and Machine Learning
MAS 622J Course Project Classification of Affective States - GP Semi-Supervised Learning, SVM and kNN Hyungil Ahn
Introduction to Machine learning
Presentation transcript:

Prediction of Malignancy of Ovarian Tumors Using Least Squares Support Vector Machines C. Lu 1, T. Van Gestel 1, J. A. K. Suykens 1, S. Van Huffel 1, I. Vergote 2, D. Timmerman 2 1 Department of Electrical Engineering, Katholieke Universiteit Leuven, Leuven, Belgium, 2 Department of Obstetrics and Gynecology, University Hospitals Leuven, Leuven, Belgium address: Demographic, serum marker, color Doppler imaging and morphologic variables 1. Introduction Ovarian masses is a common problem in gynecology. A reliable test for preoperative discrimination between benign and malignant ovarian tumors is of considerable help for clinicians in choosing appropriate treatments for patients. In this work, we develop and evaluate several LS-SVM models within Bayesian evidence framework, to preoperatively predict malignancy of ovarian tumors. The analysis includes exploratory data analysis, optimal input variable selection, parameter estimation, performance evaluation via Receiver Operating Characteristic (ROC) curve analysis. 2. Methods o: benign case x: malignant case Visualizing the correlation between the variables and the relations between the variables and clusters. Biplot of Ovarian Tumor Data Patient Data Unv. Hospitals Leuven 1994~ records, 25 features 32% malignant Univariate Analysis Preprocessing Multivariate Analysis PCA, Factor analysis Stepwise logistic regression Model Building Bayesian LS-SVM Classifier (RBF, Linear) Logistic Regression Model Evaluation ROC analysis: AUC Cross validation (Hold out, K-fold CV) Descriptive statistics Histograms Input Selection Data Exploration Model Development Procedure of developing models to predict the malignancy of ovarian tumors Bayesian LS-SVM (RBF, Linear) Forward Selection (Max Evidence) LS-SVM Classifier within Bayesian Evidence Framework Level 1: infer w,b Level 2: Infer hyperparameter Level 3: Compare models Positive definite kernel K(.,.) RBF: Linear: Mercer’s theorem Posterior class probability Model evidence Input variable selection Given a certain type of kernel, Performs forward selection  Initial: 0 variables,  Add: variable which gives the greatest increase in the current model evidence at each iteration.  Stop: when the adding of any remaining variable can no longer increase the model evidence. solved in dual space y(x), p(y=1|x,D,H) D train Bayesian LS-SVM Classifier kernel type (rbf/linear) model evidence x  *,  *,  *,  *, , b initial set of {  j } for rbf kernels trainingtest 10 variables were selected using an RBF kernel. l_ca125, pap, sol, colsc3, bilat, meno, asc, shadows, colsc4, irreg Blackbox of Bayesian LS-SVM Classifier 3. Experimental Results RMI: risk of malignancy index = score morph × score meno × CA125 2) Results from randomized cross-validation (30 runs)  Training set : data from the first treated 265 patients  Test set : data from the latest treated 160 patients 1) Results from Temporal validation -- LSSVMrbf -- LSSVMlin -- LR -- RMI ROC curve on test set Performance on Test set Averaged Performance on 30 runs of validations  randomly separating training set (n=265) and test set (n=160)  Stratified, #malignant : #benign ~ 2:1 for each training and test set.  Repeat 30 times Expected ROC curve on validation Goal:  High sensitivity for malignancy low false positive rate.  Providing probability of malignancy for individuals. 4. Conclusions Within the Bayesian evidence framework, the hyperparameter tuning, input variable selection and computation of posterior class probability can be done in a unified way, without the need of selecting additional validation set. A forward input selection procedure which tries to maximize the model evidence can be used to identify the subset of important variables for model building. LS-SVMs have the potential to give reliable preoperative prediction of malignancy of ovarian tumors. Future work LS-SVMs are blackbox models. Hybrid methodology, e.g. combine the Bayesian network with the learning of LS-SVM, might be promising A larger scale validation is needed.  Conditional class probabilities computed using Gaussian distributions  Posterior class probability  The probability of tumor being malignant p(y=+1|x,D,H) will be used for final classification (by thresholding). References 1. C. Lu, T. Van Gestel, J. A. K. Suykens, S. Van Huffel, I. Vergote, D. Timmerman, Prediction of malignancy of ovarian tumors using Least Squares Support Vector Machines, Artificial Intelligence in Medicine, vol. 28, no. 3, Jul. 2003, pp J.A.K. Suykens, T. Van Gestel, J. De Brabanter, B. De Moor, J. Vandewalle. Least Squares Support Vector Machines. World Scientific, Singapore: 2002.