MicroArray Data Analysis Candice Quadros & Amol Kothari.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Salvatore giorgi Ece 8110 machine learning 5/12/2014
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Modular Neural Networks Approach to Chemical Content Analysis of Vegetation 1 N. Kussul, 1 V. Yatsenko, 2 A. Sachenko, 3 G. Markowsky, 1 A. Sydorenko,
COMP 328: Final Review Spring 2010 Nevin L. Zhang Department of Computer Science & Engineering The Hong Kong University of Science & Technology
Assuming normally distributed data! Naïve Bayes Classifier.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
SocalBSI 2008: Clustering Microarray Datasets Sagar Damle, Ph.D. Candidate, Caltech  Distance Metrics: Measuring similarity using the Euclidean and Correlation.
Statistical Concepts (part III) Concepts to cover or review today: –Population parameter –Sample statistics –Mean –Standard deviation –Coefficient of variation.
Chapter 5: Linear Discriminant Functions
Radial Basis Functions
Classification: Support Vector Machine 10/10/07. What hyperplane (line) can separate the two classes of data?
Speaker Adaptation for Vowel Classification
Discrimination Methods As Used In Gene Array Analysis.
Multidimensional Analysis If you are comparing more than two conditions (for example 10 types of cancer) or if you are looking at a time series (cell cycle.
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
1 lBayesian Estimation (BE) l Bayesian Parameter Estimation: Gaussian Case l Bayesian Parameter Estimation: General Estimation l Problems of Dimensionality.
1 Automated Feature Abstraction of the fMRI Signal using Neural Network Clustering Techniques Stefan Niculescu and Tom Mitchell Siemens Medical Solutions,
Microarray analysis 2 Golan Yona. 2) Analysis of co-expression Search for similarly expressed genes experiment1 experiment2 experiment3 ……….. Gene i:
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
Bayesian Estimation (BE) Bayesian Parameter Estimation: Gaussian Case
Projection methods in chemistry Autumn 2011 By: Atefe Malek.khatabi M. Daszykowski, B. Walczak, D.L. Massart* Chemometrics and Intelligent Laboratory.
Ranga Rodrigo April 5, 2014 Most of the sides are from the Matlab tutorial. 1.
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
Matlab Matlab Sigmoid Sigmoid Perceptron Perceptron Linear Linear Training Training Small, Round Blue-Cell Tumor Classification Example Small, Round Blue-Cell.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 21 Oct 28, 2005 Nanjing University of Science & Technology.
Microarrays.
The Perceptron. Perceptron Pattern Classification One of the purposes that neural networks are used for is pattern classification. Once the neural network.
1/15 Strengthening I-ReGEC classifier G. Attratto, D. Feminiano, and M.R. Guarracino High Performance Computing and Networking Institute Italian National.
A comparison of the ability of artificial neural network and polynomial fitting was carried out in order to model the horizontal deformation field. It.
Ch 4. Linear Models for Classification (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized and revised by Hee-Woong Lim.
Evolutionary Algorithms for Finding Optimal Gene Sets in Micro array Prediction. J. M. Deutsch Presented by: Shruti Sharma.
Non-Bayes classifiers. Linear discriminants, neural networks.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Speech Communication Lab, State University of New York at Binghamton Dimensionality Reduction Methods for HMM Phonetic Recognition Hongbing Hu, Stephen.
Clustering Instructor: Max Welling ICS 178 Machine Learning & Data Mining.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
CSE182 L14 Mass Spec Quantitation MS applications Microarray analysis.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Principal Components Analysis ( PCA)
Classification of tissues and samples 指導老師:藍清隆 演講者:張許恩、王人禾.
Big data classification using neural network
Neural Networks for Quantum Simulation
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
第 3 章 神经网络.
Information Retrieval
Chapter 3: Maximum-Likelihood and Bayesian Parameter Estimation (part 2)
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 20
Neural Networks Advantages Criticism
CSE572, CBS598: Data Mining by H. Liu
Neuro-Computing Lecture 4 Radial Basis Function Network
Neural Networks and Their Application in the Fields of Coporate Finance By Eric Séverin Hanna Viinikainen.
network of simple neuron-like computing elements
Word Embedding Word2Vec.
CSE572, CBS572: Data Mining by H. Liu
Dimension reduction : PCA and Clustering
CSSE463: Image Recognition Day 18
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 18
CSSE463: Image Recognition Day 18
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 18
CSE572: Data Mining by H. Liu
Chapter 3: Maximum-Likelihood and Bayesian Parameter Estimation (part 2)
Using Clustering to Make Prediction Intervals For Neural Networks
Batch Normalization.
Pattern Recognition: Statistical and Neural
Presentation transcript:

MicroArray Data Analysis Candice Quadros & Amol Kothari

Harnessing the power of a neural network for classifying samples. Neural Network for classification

Reduce the no. of genes :  We have to reduce the data dimensionality, i.e. reduce the no. of genes to consider.  PCA can be used to select most informative genes, but it is computationally expensive to obtain the Eigen vectors for high dimensional data.  Use the method suggested by Golub et al. to obtain the informative genes. Neural Network for classification

Steps in classification  Obtain the informative genes using Golub’s method.  Normalize the genes by shifting them to the mean & dividing by the standard deviation.  Train the neural network by using the training data & targets, and get the weights.  Classify the test data using the weights obtained above. Neural Network for classification

Results obtained: Inform. Genes No. of Hidden Units NN: Accuracy Golub: Accuracy Neural Network for classification

Hierarchical Merging: When to stop? Question: When to stop the merging? Suggested Solutions:  Diameter(C)  MaxD  Avg(sim(O i,O j )) ≥  (O i,O j  C) Difficult to estimate the parameters in high dimensions.

Another solution: When m clusters are present, stop merging. Problem: The m clusters might contain single point clusters. Use the concept of MinPts (from DBScan). A set of points is a significant cluster only if the set has MinPts. When there are m significant clusters, then stop. Hierarchical Merging: When to stop?

No. of iterations No. of Significant Clusters

Visualization of data: Vizstruct

Equation used: How do weigh each dimension, i.e. how do we select λ? Default value = 0.5 Use the Eigen Values of each dimension to obtain the value of λ. Visualization of data: Vizstruct

Steps for visualization:  Project the data into Eigen space.  The Eigen values of each dimension i = λi  Now use the same formulae for calculating the 2D point: Where λ i = Eigen value of the i th dimension Visualization of data: Vizstruct

Results:  The visualization obtained by this method is more representative of the data, compared to Vizstruct.  Demo Visualization of data: Vizstruct

Thank You