Proteomic Mass Spectrometry

Slides:



Advertisements
Similar presentations
Discrimination amongst k populations. We want to determine if an observation vector comes from one of the k populations For this purpose we need to partition.
Advertisements

Face Recognition and Biometric Systems Eigenfaces (2)
Data Mining Feature Selection. Data reduction: Obtain a reduced representation of the data set that is much smaller in volume but yet produces the same.
Feature Grouping-Based Fuzzy-Rough Feature Selection Richard Jensen Neil Mac Parthaláin Chris Cornelis.
Comparing Twitter Summarization Algorithms for Multiple Post Summaries David Inouye and Jugal K. Kalita SocialCom May 10 Hyewon Lim.
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
COMPUTER AIDED DIAGNOSIS: FEATURE SELECTION Prof. Yasser Mostafa Kadah –
S-SENCE Signal processing for chemical sensors Martin Holmberg S-SENCE Applied Physics, Department of Physics and Measurement Technology (IFM) Linköping.
A segmented principal component analysis applied to calorimetry information at ATLAS ACAT May 22-27, Zeuthen, Germany H. P. Lima Jr, J. M. de Seixas.
Multivariate Methods Pattern Recognition and Hypothesis Testing.
Exploratory Data Mining and Data Preparation
Dimension reduction : PCA and Clustering Agnieszka S. Juncker Slides: Christopher Workman and Agnieszka S. Juncker Center for Biological Sequence Analysis.
L15:Microarray analysis (Classification) The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
Diagnosis of Ovarian Cancer Based on Mass Spectra of Blood Samples Hong Tang Yelena Mukomel Eugene Fink.
4 th NETTAB Workshop Camerino, 5 th -7 th September 2004 Alberto Bertoni, Raffaella Folgieri, Giorgio Valentini
ACM SAC’06, DM Track Dijon, France “The Impact of Sample Reduction on PCA-based Feature Extraction for Supervised Learning” by M. Pechenizkiy,
Multidimensional Analysis If you are comparing more than two conditions (for example 10 types of cancer) or if you are looking at a time series (cell cycle.
7/17/2002 Greg Grudic: Nonparametric Modeling 1 High Dimensional Nonparametric Modeling Using Two-Dimensional Polynomial Cascades Greg Grudic University.
Feature Subset Selection using Minimum Cost Spanning Trees Mike Farah Supervisor: Dr. Sid Ray.
Microarray analysis 2 Golan Yona. 2) Analysis of co-expression Search for similarly expressed genes experiment1 experiment2 experiment3 ……….. Gene i:
CIBB-WIRN 2004 Perugia, 14 th -17 th September 2004 Alberto Bertoni, Raffaella Folgieri, Giorgio Valentini Feature.
Feature Selection and Its Application in Genomic Data Analysis March 9, 2004 Lei Yu Arizona State University.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Diagnosis of Ovarian Cancer Based on Mass Spectrum of Blood Samples Committee: Eugene Fink Lihua Li Dmitry B. Goldgof Hong Tang.
Rotation Forest: A New Classifier Ensemble Method 交通大學 電子所 蕭晴駿 Juan J. Rodríguez and Ludmila I. Kuncheva.
Selective Sampling on Probabilistic Labels Peng Peng, Raymond Chi-Wing Wong CSE, HKUST 1.
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
A Multivariate Biomarker for Parkinson’s Disease M. Coakley, G. Crocetti, P. Dressner, W. Kellum, T. Lamin The Michael L. Gargano 12 th Annual Research.
Presented by Tienwei Tsai July, 2005
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
Application of Data Mining Algorithms in Atmospheric Neutrino Analyses with IceCube Tim Ruhe, TU Dortmund.
Feature selection LING 572 Fei Xia Week 4: 1/29/08 1.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Reanalysis of Petricoin et al. Ovarian Cancer Data Set 3 Russ Wolfinger and Geoff Mann SAS Institute Inc. NISS Proteomics Workshop March 6, 2003.
Skewing: An Efficient Alternative to Lookahead for Decision Tree Induction David PageSoumya Ray Department of Biostatistics and Medical Informatics Department.
1 Effective Feature Selection Framework for Cluster Analysis of Microarray Data Gouchol Pok Computer Science Dept. Yanbian University China Keun Ho Ryu.
Laxman Yetukuri T : Modeling of Proteomics Data
Reduction of Training Noises for Text Classifiers Rey-Long Liu Dept. of Medical Informatics Tzu Chi University Taiwan.
CLASSIFICATION: Ensemble Methods
Evolutionary Algorithms for Finding Optimal Gene Sets in Micro array Prediction. J. M. Deutsch Presented by: Shruti Sharma.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Chapter 6 Classification and Prediction Dr. Bernard Chen Ph.D. University of Central Arkansas.
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Stable Feature Selection for Biomarker Discovery Name: Goutham Reddy Bakaram Student Id: Instructor Name: Dr. Dongchul Kim Review Article by Zengyou.
NTU & MSRA Ming-Feng Tsai
Feature Selection and Extraction Michael J. Watts
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
2D-LDA: A statistical linear discriminant analysis for image matrix
Dr. Gheith Abandah 1.  Feature selection is typically a search problem for finding an optimal or suboptimal subset of m features out of original M features.
THE SCIENTIFIC METHOD: It’s the method you use to study a question scientifically.
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
On the relevance of facial expressions for biometric recognition Marcos Faundez-Zanuy, Joan Fabregas Escola Universitària Politècnica de Mataró (Barcelona.
Automatic Classification for Pathological Prostate Images Based on Fractal Analysis Source: IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 28, NO. 7, JULY.
Predictive Automatic Relevance Determination by Expectation Propagation Y. Qi T.P. Minka R.W. Picard Z. Ghahramani.
School of Computer Science & Engineering
Chapter 6 Classification and Prediction
Alan Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani
Project 2 k-NN 2018/11/10.
Object Modeling with Layers
Historic Document Image De-Noising using Principal Component Analysis (PCA) and Local Pixel Grouping (LPG) Han-Yang Tang1, Azah Kamilah Muda1, Yun-Huoy.
Design of Hierarchical Classifiers for Efficient and Accurate Pattern Classification M N S S K Pavan Kumar Advisor : Dr. C. V. Jawahar.
Data Preprocessing Copyright, 1996 © Dale Carnegie & Associates, Inc.
Dimension reduction : PCA and Clustering
Machine Learning for Visual Scene Classification with EEG Data
Chapter 7: Transformations
Data Preprocessing Copyright, 1996 © Dale Carnegie & Associates, Inc.
Feature Selection Methods
Feature Extraction (I)
Lecture 16. Classification (II): Practical Considerations
Can Genetic Programming Do Manifold Learning Too?
Presentation transcript:

Proteomic Mass Spectrometry

Outline Previous Research Project Goals Data and Algorithms Experimental Results Conclusions To Do List

Motivation MS spectra has high dimension Dimensionality Reduction (DR) Most ML algorithms are incapable of handling such high dimensional data Dimensionality Reduction (DR) Preserve as much information as possible, while reducing the dimensionality. Feature Extraction (FE) Removal of irrelevant and/or redundant features (information)

Previous research Usually applies DR then FE Does Order matter ? DR: Down Sampling, PCA, Wavelets FE: T-Test, Random Forests, Manual Peak Extraction In [conrads03] show that high resolution MS spectra produces better classification accuracy. Most previous research down samples spectra CONJECTURE: Down Sampling detrimental to performance.

Project Goals Test Down Sampling Conjecture Compare FE algorithms (NOTE: Optimal FE is NP-hard !) Use a simple but fast classifier to test a number of FE approaches Test across different data sets Are there any clearly superior FE algorithms ?

Three Data Sets Heart/Kidney (100/100) Ovarian Cancer (91/162) 164,168 features, 2 classes Ovarian Cancer (91/162) 15,154 features, 2 classes Prostate Cancer (63/190/26/43) 15,154 features, 4 classes Normal, Benign, Stage 1, Stage 2 Cancer Transformed into Normal/Benign Vs Cancer (1&2)

Algorithms Centroid Classifier given class means P, Q and sample point s C = argmin (d(P,s), d(Q,s)) P Q d(Q,s) d(P,s)

Algorithms T-test – do the means of 2 distr. Differ ? KS-test – do the cdf differ ? Composite – (T-test)*(KS-test) IFE - Individual Feature Evaluation using the centroid classifier DPCA – discriminative principle component analysis

Preliminary Experiments Compare normalization approaches Compare similarity metrics Cross correlation (-L1) Angular Across 3 data sets => 27 configurations

Preliminary Experiments (cont) No single norm/metric clearly superior on all data sets 2-5% increase in performance if suitable normalization and similarity metric chosen (can be up to 10% increase) L1-norm with angle similarity metric worked well on Heart/Kidney and Ovarian Cancer sets (easy sets) L1-norm and L1-metric best on Prostate 2-class problem (hard set).

Down Sampling

Statistical Tests T-test, KS-test, Composite Ranks features in terms of relevance SFS – Sequential Forward Selection Selects ever increasing feature sets I.e., {1}; {1,2}; {1,2,3}; {1,2,3,4}

Heart/Kidney

Ovarian Cancer

Prostate Cancer

Single Feature Classification Use each feature to classify test samples Rank features in terms of performance SFS

Performance Comparison

Performance Comparison

Performance Comparison

Summary For each data set, for each FE algorithm ran 15,000 3-fold cross validation experiments. Total of 810,000 FE experiments ran DE experiments ~ 100,000 experiments Additional 50,000 experiments using DPCA classifier did not produce significantly different results than the centroid classifier

Conclusions HK and Ovarian Data sets considerably easier to classify than Prostate Cancer Feature Extraction (in general) significantly improves performance on all data sets No single technique superior on all data sets. Best Performance using SFS with feature weighting Smallest feature set with T-test of KS-Test Composite test inferior to all others. Down Sampling appears to be detrimental What about other Dim. Red. Techniques ? E.g. PCA and Wavelets

Conclusions Down Sampling appears to be detrimental What about other Dim. Red. Techniques ? E.g. PCA and Wavelets What about FE after Down Sampling ? On Prostate data performance appears to drop w.r.t. to best single feature.

To Do List Check PCA, Wavelets and other DR techniques Use other (better) classifiers General Hypothesis Use a simple fast classifier together with FE techniques to extract a good feature set Replace classifier with a more effective one. Need to verify that other classifiers respond well to the extracted features.