A Plot for Visualizing Multivariate Data Rida E. A. Moustafa George Mason University ADM Group,AAL

Slides:



Advertisements
Similar presentations
Augmented Negative Selection Algorithm with Variable-Coverage Detectors Zhou Ji, Zhou Ji, St. Jude Childrens Research Hospital Dipankar Dasgupta, Dipankar.
Advertisements

Lecture 11: Algorithms and Time Complexity I Discrete Mathematical Structures: Theory and Applications.
A pre-Weekend Talk on Online Learning TGIF Talk Series Purushottam Kar.
Spoofing State Estimation
CHAPTER 13: Alpaydin: Kernel Machines
LOGO Regression Analysis Lecturer: Dr. Bo Yuan
Step three: statistical analyses to test biological hypotheses General protocol continued.
Feb. 17, 2011 Midterm overview Real life examples of built chips
Math Review with Matlab:
Motion in Two and Three Dimensions
Test B, 100 Subtraction Facts
Calculus 5e Early Transcendentals Multivariable James Stewart Chapter 16 Vector Calculus.
ENGINEERING MECHANICS CHAPTER 7
STOR 892 Object Oriented Data Analysis Radial Distance Weighted Discrimination Jie Xiong Advised by Prof. J.S. Marron Department of Statistics and Operations.
Covariance Matrix Applications
Dimension reduction (1)
Manifold Learning Dimensionality Reduction. Outline Introduction Dim. Reduction Manifold Isomap Overall procedure Approximating geodesic dist. Dijkstra’s.
1 Multivariate Statistics ESM 206, 5/17/05. 2 WHAT IS MULTIVARIATE STATISTICS? A collection of techniques to help us understand patterns in and make predictions.
Multivariate Methods Pattern Recognition and Hypothesis Testing.
Principal Component Analysis
DNA Microarray Bioinformatics - #27611 Program Normalization exercise (from last week) Dimension reduction theory (PCA/Clustering) Dimension reduction.
Manifold Learning: ISOMAP Alan O'Connor April 29, 2008.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Section 1.1 The Distance and Midpoint Formulas. x axis y axis origin Rectangular or Cartesian Coordinate System.
Comparing Kernel-based Learning Methods for Face Recognition Zhiguo Li
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Microarray Gene Expression Data Analysis A.Venkatesh CBBL Functional Genomics Chapter: 07.
First Al-Khawarezmi Conference: Qatar, December 6-8, 2010 Ali Hadi 0 0 The Effects of Centering and Scaling the Rows of Multidimensional Data on Their.
Summarized by Soo-Jin Kim
This week: overview on pattern recognition (related to machine learning)
O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY RobustMap: A Fast and Robust Algorithm for Dimension Reduction and Clustering Lionel F.
LHE 11.2 Three-Dimensional Coordinate in Space Calculus III Berkley High School September 14, 2009.
Retraining of the SAMANN Network Viktor Medvedev Viktor Medvedev, Gintautas Dzemyda {Viktor.m, Institute of Mathematics and Informatics.
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
Enhancing Interactive Visual Data Analysis by Statistical Functionality Jürgen Platzer VRVis Research Center Vienna, Austria.
Using Interactive Evolution for Exploratory Data Analysis Tomáš Řehořek Czech Technical University in Prague.
Available at Chapter 13 Multivariate Analysis BCB 702: Biostatistics
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
Dimensionality Reduction Motivation I: Data Compression Machine Learning.
Tokyo Research Laboratory © Copyright IBM Corporation 2005SDM 05 | 2005/04/21 | IBM Research, Tokyo Research Lab Tsuyoshi Ide Knowledge Discovery from.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
Principle Component Analysis and its use in MA clustering Lecture 12.
Principal Component Analysis (PCA)
11.2 Space coordinates and vectors in Space. 3 dimensional coordinate plane.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
Principal Component Analysis Zelin Jia Shengbin Lin 10/20/2015.
Scikit-Learn Intro to Data Science Presented by: Vishnu Karnam A
Linear Classifiers Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
1 UNC, Stat & OR Hailuoto Workshop Object Oriented Data Analysis, I J. S. Marron Dept. of Statistics and Operations Research, University of North Carolina.
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
Principal Components Analysis ( PCA)
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Out of sample extension of PCA, Kernel PCA, and MDS WILSON A. FLORERO-SALINAS DAN LI MATH 285, FALL
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Principal Component Analysis (PCA)
Dimensionality Reduction
LECTURE 11: Advanced Discriminant Analysis
Dimension reduction : PCA and Clustering by Agnieszka S. Juncker
Principal Component Analysis (PCA)
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
Principal Component Analysis
Descriptive Statistics vs. Factor Analysis
Chapter 3 Graphs and Functions.
Principal Components Analysis
Unit 1 Representing Real Numbers
INTRODUCTION TO Machine Learning
Lecture 16. Classification (II): Practical Considerations
Dimensionality Reduction Part 1 of 2
Presentation transcript:

A Plot for Visualizing Multivariate Data Rida E. A. Moustafa George Mason University ADM Group,AAL

Talk Outline The Theory of MV-Plot. Detecting Linear Structures with MV-plot. Detecting Non-Linear Structures with MV-plot. Comparisons with other methods and application on real data.

MV-Plot Theory Given an observation x=(x 1,x 2,…,x d ) We define m and v as follows: Computing m and v for every observation produces vector of m and v. What is the relationship between m and v?

MV-Relationship in 2-d Normalizing the data in range (0,1) avoid the abs-value in computing m. Close to the PC in 2-d

MV- detects linear structure(s) If the data is linear in the original space  It will be linear in the MV-space!!

MV- detects linear structure(s)

Detecting Linear structure(s) Example I

Detecting Linear structure(s) Example II

Detecting Linear structure(s) Example III

Detecting nonlinear data with MV-plot MV- plot can detect nonlinear structure in the data set without any changes in the equations.

Detecting nonlinear structure

Detecting Sphere(s) Case I: The sphere radius R The sphere center is the origin

Detecting Sphere(s) Case II: The sphere radius R The sphere center is not the origin

Detecting Sphere(s)

Fisher’s IRIS data (150x4) 3-classes of( 50 point each) Process control data (600x60) 6-classes of (100 points each) Pollen data (3,848x5) (Wegman’s data) 2-classes (linear and nonlinear) Application on Real data

Multidimensional Scaling Fisher Discriminate Analysis Principal Component Related Dimensional Reduction Methods

IRIS (R. A. Fisher) Dataset 1 50-cases in 4-dim

Time Series Dataset 600-cases in 60-dim

Pollen dataset 3,848-points in 5-dim Other methods: Require more storage and speed. Even if it work, we expect bad results on this particular data. (Wegman2002)

Pollen dataset Linear and Nonlinear mixed structures.

The linear structure in the Pollen data set =98 Linear, 3750 nonlinear

Summary MV-algorithm can discover the linear and nonlinear pattern at the same time. MV-algorithm can discover symmetric data. MV-algorithm deals with large multivariate data.