Classification of tissues and samples 指導老師:藍清隆 演講者:張許恩、王人禾.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Perceptron Lecture 4.
COMPUTER AIDED DIAGNOSIS: CLASSIFICATION Prof. Yasser Mostafa Kadah –
VC theory, Support vectors and Hedged prediction technology.
G53MLE | Machine Learning | Dr Guoping Qiu
Machine Learning Instance Based Learning & Case Based Reasoning Exercise Solutions.
Vector Operations in R 3 Section 6.7. Standard Unit Vectors in R 3 The standard unit vectors, i(1,0,0), j(0,1,0) and k(0,0,1) can be used to form any.
1 Lecture 5 Support Vector Machines Large-margin linear classifier Non-separable case The Kernel trick.
Achim Tresch Computational Biology ‘Omics’ - Analysis of high dimensional Data.
Region labelling Giving a region a name. Image Processing and Computer Vision: 62 Introduction Region detection isolated regions Region description properties.
Chapter 5: Linear Discriminant Functions
Classification: Support Vector Machine 10/10/07. What hyperplane (line) can separate the two classes of data?
Classification of Microarray Data. Sample Preparation Hybridization Array design Probe design Question Experimental Design Buy Chip/Array Statistical.
L15:Microarray analysis (Classification). The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Discriminant Functions Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Classification of Microarray Data. Sample Preparation Hybridization Array design Probe design Question Experimental Design Buy Chip/Array Statistical.
Basics of discriminant analysis
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
DIMACS Workshop on Machine Learning Techniques in Bioinformatics 1 Cancer Classification with Data-dependent Kernels Anne Ya Zhang (with Xue-wen.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Optimization Theory Primal Optimization Problem subject to: Primal Optimal Value:
Applications of Data Mining in Microarray Data Analysis Yen-Jen Oyang Dept. of Computer Science and Information Engineering.
JM - 1 Introduction to Bioinformatics: Lecture VIII Classification and Supervised Learning Jarek Meller Jarek Meller Division.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Classification of multiple cancer types by multicategory support vector machines using gene expression data.
Outline Separating Hyperplanes – Separable Case
Classification (Supervised Clustering) Naomi Altman Nov '06.
More on Microarrays Chitta Baral Arizona State University.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
PCA, Clustering and Classification by Agnieszka S. Juncker Part of the slides is adapted from Chris Workman.
Classification of microarray samples Tim Beißbarth Mini-Group Meeting
Breast Cancer Diagnosis via Neural Network Classification Jing Jiang May 10, 2000.
Evolutionary Algorithms for Finding Optimal Gene Sets in Micro array Prediction. J. M. Deutsch Presented by: Shruti Sharma.
Non-Bayes classifiers. Linear discriminants, neural networks.
Kernels in Pattern Recognition. A Langur - Baboon Binary Problem m/2006/ /himplu s4.jpg … HA HA HA …
An Introduction to Support Vector Machine (SVM)
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Applications of Supervised Learning in Bioinformatics Yen-Jen Oyang Dept. of Computer Science and Information Engineering.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
CZ5225: Modeling and Simulation in Biology Lecture 7, Microarray Class Classification by Machine learning Methods Prof. Chen Yu Zong Tel:
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
Case Study: Characterizing Diseased States from Expression/Regulation Data Tuck et al., BMC Bioinformatics, 2006.
Support-Vector Networks C Cortes and V Vapnik (Tue) Computational Models of Intelligence Joon Shik Kim.
CSE182 L14 Mass Spec Quantitation MS applications Microarray analysis.
1 Learning Bias & Clustering Louis Oliphant CS based on slides by Burr H. Settles.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
SUPPORT VECTOR MACHINES Presented by: Naman Fatehpuria Sumana Venkatesh.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Linear Models & Clustering Presented by Kwak, Nam-ju 1.
Classifiers!!! BCH364C/391L Systems Biology / Bioinformatics – Spring 2015 Edward Marcotte, Univ of Texas at Austin.
Automatic Classification of Audio Data by Carlos H. L. Costa, Jaime D. Valle, Ro L. Koerich IEEE International Conference on Systems, Man, and Cybernetics.
SUPPORT VECTOR MACHINES
Gene Expression Classification
Molecular Classification of Cancer
Support Vector Machines
PCA, Clustering and Classification by Agnieszka S. Juncker
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
The Naïve Bayes (NB) Classifier
Other Classification Models: Support Vector Machine (SVM)
Parametric Methods Berlin Chen, 2005 References:
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Support Vector Machines 2
Presentation transcript:

Classification of tissues and samples 指導老師:藍清隆 演講者:張許恩、王人禾

introduction The use of microarrays to find group of genes that can be used diagnostically to determine the disease that an individual is suffering from, or prognostically to predict the success of a course of therapy or results of an experiment. Purpose: find a small number of genes that can predict to which group each individual belongs

Data set 9A Bone marrow samples are taken from 27 patients suffering from acute lymphoblastic leukemia(ALL) and 11 patients suffering from acute myeloid leukemia(AML)

Method of classification Describe methods that allow you to predict the class to which an individual belongs, based on gene expression measurements Assume that we have already selected a small number of genes whose expression measurements we use, and are not using all the genes on the microarray

Two concept central to classification Separability and linearity

Separability Separable: the different groups to which the samples belong occupy different regions of the gene expression space Non-Separable: the different groups to which the samples belong are mixed together in the same region of gene expression space

linearity Linearly separable: possible to partition the space between the two (or more) group using straight lines Non- Linearly separable: separable, but are not possible to partition the groups using straight lines

Method of classification 1.K-nearest neighbours 2.Nearest centroid 3.Linear discriminant analysis 4.Neural networks 5.Support vector machines

K-nearest neighbours Steps: 1.We look at the gene expression measurements for the sample we are trying to classify 2.Find the nearest of the known samples as measured by an appropriate distance measure 3.The class of the sample is the class of the nearest sample

K-nearest neighbours 2 parameter: k & l

Centroid classification Steps: 1.For each class, calculate the center of mass of the points of the representative samples 2.Calculate the distance between the position of the sample to be classified and each of the centers of mass of the classes using an appropriate distance measure 3.Assign the sample to the class whose center of mass is nearest to it

Centroid classification

Linear discriminant analysis Steps: 1.Calculate a straight line (in two dimensions) or hyperplane (in more than two dimensions) that separates two known classes so as to minimise the within class variance on either side of the line and maximise the between class variance(figure 9.4) 2.The class of the unknown sample is determined by the side of the hyperplane on which the sample lies.

Linear discriminant analysis

Neural networks Steps: 1.Train the neural network using the samples with known classes 2.Apply the neural network to the new individual to determine it’s class

Neural networks

Support vector machines Steps: 1.Project the data from the known classes into a suitable high-dimensional space 2.Identify a hyperplane that separates two classes 3.The class of the new individual is determined by the side of hyperplane on which the sample lines

Support vector machines

謝謝大家 的支持!