Co-Occurrence and Morphological Analysis for Colon Tissue Biopsy Classification Khalid Masood, Nasir Rajpoot, Kashif Rajpoot*, Hammad Qureshi Signal and.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

ECG Signal processing (2)
Component Analysis (Review)
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Input Space versus Feature Space in Kernel- Based Methods Scholkopf, Mika, Burges, Knirsch, Muller, Ratsch, Smola presented by: Joe Drish Department of.
SVM—Support Vector Machines
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
An Overview of Machine Learning
Face Recognition & Biometric Systems, 2005/2006 Face recognition process.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
© 2003 by Davi GeigerComputer Vision September 2003 L1.1 Face Recognition Recognized Person Face Recognition.
Principal Component Analysis
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Feature Screening Concept: A greedy feature selection method. Rank features and discard those whose ranking criterions are below the threshold. Problem:
Computer Vision I Instructor: Prof. Ko Nishino. Today How do we recognize objects in images?
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Biomedical Image Analysis and Machine Learning BMI 731 Winter 2005 Kun Huang Department of Biomedical Informatics Ohio State University.
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Recognition Part II Ali Farhadi CSE 455.
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
A New Subspace Approach for Supervised Hyperspectral Image Classification Jun Li 1,2, José M. Bioucas-Dias 2 and Antonio Plaza 1 1 Hyperspectral Computing.
BACKGROUND LEARNING AND LETTER DETECTION USING TEXTURE WITH PRINCIPAL COMPONENT ANALYSIS (PCA) CIS 601 PROJECT SUMIT BASU FALL 2004.
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Learning a Kernel Matrix for Nonlinear Dimensionality Reduction By K. Weinberger, F. Sha, and L. Saul Presented by Michael Barnathan.
Line detection Assume there is a binary image, we use F(ά,X)=0 as the parametric equation of a curve with a vector of parameters ά=[α 1, …, α m ] and X=[x.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Jan 18, 2008 Ju Han, Hang Chang, Mary Helen Barcellos-Hoff, and Bahram Parvin Lawrence Berkeley National Laboratory Multivariate.
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
CSE 185 Introduction to Computer Vision Face Recognition.
Levels of Image Data Representation 4.2. Traditional Image Data Structures 4.3. Hierarchical Data Structures Chapter 4 – Data structures for.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Reduces time complexity: Less computation Reduces space complexity: Less parameters Simpler models are more robust on small datasets More interpretable;
Supervisor: Nakhmani Arie Semester: Winter 2007 Target Recognition Harmatz Isca.
CZ5225: Modeling and Simulation in Biology Lecture 7, Microarray Class Classification by Machine learning Methods Prof. Chen Yu Zong Tel:
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
2D-LDA: A statistical linear discriminant analysis for image matrix
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
In Search of the Optimal Set of Indicators when Classifying Histopathological Images Catalin Stoean University of Craiova, Romania
University of Ioannina
LECTURE 10: DISCRIMINANT ANALYSIS
Face Recognition and Feature Subspaces
Recognition: Face Recognition
Machine Learning Basics
Brain Hemorrhage Detection and Classification Steps
In summary C1={skin} C2={~skin} Given x=[R,G,B], is it skin or ~skin?
REMOTE SENSING Multispectral Image Classification
REMOTE SENSING Multispectral Image Classification
Midterm Exam Closed book, notes, computer Format:
X.1 Principal component analysis
CS4670: Intro to Computer Vision
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
LECTURE 09: DISCRIMINANT ANALYSIS
Announcements Project 2 artifacts Project 3 due Thursday night
Where are we? We have covered: Project 1b was due today
Feature Selection Methods
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
The “Margaret Thatcher Illusion”, by Peter Thompson
Presentation transcript:

Co-Occurrence and Morphological Analysis for Colon Tissue Biopsy Classification Khalid Masood, Nasir Rajpoot, Kashif Rajpoot*, Hammad Qureshi Signal and Image Processing Group, University of Warwick (UK) * Wolfson Medical Vision Lab, Oxford University (UK)

2 Problem Definition Given hyperspectral image cube of a patient’s colon tissue biopsy sample, automatically label the sample as Benign or Malignant Our approach is based on the idea that malignancy of a tumor alters the macro-architecture of the tissue glands: – Nice tubular structure of the glands for benign tumors – No such structure for malignant tumors BenignMalignant

3 Motivation for Colon Biopsy Classification Useful for screening of the colon cancer Visual assessment by pathologists is very subjective Significant intra- and inter-observational variation between pathologists Quantitative histopathological analysis techniques offer objective, reliable, accurate, and reproducible assessment Source: NIH

4 Hyperspectral Imaging Ordinary cameras only capture reflections from RGB colors Hyperspectral cameras capture reflections from a range of visible wavelengths +

5 Hyperspectral Imaging (HSI) HSI is a fast and reliable means of characterizing the histochemistry of tissues – HSI is also used extensively in remote sensing, satellite imaging, and defence (target detection etc.) applications The Nuance multispectral imaging system can acquire 20 subbands in visible wavelength range of nm Each hyperspectral image is a 3D data cube with a spectral coordinate in the z direction representing 20 subbands

6 Our Classification Algorithm Based on a spectral-spatial analysis of the input data cube, our algorithm consists of three stages: – Stage I: Dimensionality reduction, followed by segmentation – Stage II: Morphological/textural analysis of the segmented results – Stage III: Classification using Subspace Projection methods and Support Vector Machines (SVM)

7 Stage I: Segmentation Dimensionality reduction (from 20 to 4 bands of the multispectral image cube) is achieved using independent component analysis (ICA) and k-means clustering Nuclei, cytoplasm; gland secretions; stroma of the lamina propria

8 Stage II: Feature Extraction Four binary images are extracted from each segmented image Two sets of features are calculated: morphological and co-occurrence matrix features: Morphological Features: – These describe the shape, size, orientation and other attributes of the cellular components – These features are calculated on patches (blocks) of the segmented image Co-Occurrence Features: – These describe the textural properties of a given neighborhood

9 Morphological Features Morphological features describe the shape and texture of the image Feature vector consists of five to ten morphological features Discriminant morphological features are: – Euler Number : number of contiguous parts – Convex Area : number of pixels in convex image – Extent : the proportion of pixels in the bounding box – Solidity : the proportion of pixels in the convex hull – Area : the actual number of pixels in the patch – EquivDiameter : the diameter of a circle with the same area as of the patch

10 Co-occurrence Features The co-occurrence matrix is constructed by analysing the gray levels of neighboring pixels The (i,j) th element of a co-occurrence matrix for a particular angle and distance is given by the joint conditional pdf: Three attributes used are:

11 Stage III : Classification Subspace Projection methods – Principle Component Analysis (PCA) – Linear Discriminant Analysis (LDA) – Kernel PCA – Kernel LDA Support Vector Machines (SVM) – Polynomial kernel – Gaussian kernel

12 Principal Component Analysis (PCA) Eigenvectors of the data in the embedding space can be used to detect directions of maximum variance. The principal components can be computed by solving the eigenvalue problem: The coefficients of projection along a few top principal directions can be used as features Nearest-neighbor classifier is used for assigning label to a biopsy sample

13 Linear Discriminant Analysis (LDA) Limitations of the PCA As opposed to PCA, which maximises the overall scatter, LDA maximises the ratio of between- class scatter S b to within-class scatter S w The w i can be computed by solving the generalised eigenvalue problem:

14 Linear Boundary Assumption In most real-world problems, separating boundaries are not necessarily linear! Consider, for instance, the following example: Will a linear classifier work?

15 The Kernel Trick Kernel machines (eg, SVMs) transform non-linear decision boundaries to linear ones in higher dimensional feature space Two dimensional features (let us say) are mapped to a three dimensional feature space through a non-linear transform Non-linear ellipsoidal decision boundary is replaced to linear boundary in higher dimensional space The trick is to replace dot products in F with a kernel function in the input space R so that the non-linear mapping is performed implicitly in R

16 SVM Kernel Functions A few commonly used kernel functions are: Classifier’s performance highly sensitive to parameter values Best kernel parameter values are searched

17 Experimentation Two sets of Experiments: Mixed testing and Leave one out (LOO) For mixed testing: – 4096 patches (blocks) per image of 16x16 dimensions per patch – Morphological features – Training set contains one quarter of the patches while remaining three quarters make the test set – PCA and modular LDA are used in mixed testing Leave one out testing is done on gray level co-occurrence features – 16x16 patches – Support Vector Machines (polynomial kernel and gaussian kernel are used) – Kernel parameters are optimized Classification label is assigned to a slide according to the class of the majority of the patches

18 Experimental Results

19 Experimental Results AUCH of the ROC curve for LDA goes up to 0.92 for 5 features.

20 Conclusions and Future Work Conclusions – Tissue segmentation affects the performance – Implementation of LDA saves the computational cost and the performance achieved by it is encouraging – Gaussian kernel SVM gives no false alarm Future Work – More effective segmentation – Combination of classifiers – Shape modelling of nuclei glands

21 Acknowledgements Prof David Rimm, Department of Pathology, Yale University School of Medicine (USA) Prof Gustave Davis, Department of Pathology, Yale University School of Medicine (USA) Prof Ronald Coifman, Department of Applied Mathematics, Yale University (USA)

22 Thanks for your attention Any Questions?