1 Data Analysis  Data Matrix Variables ObjectsX1X1 X2X2 X3X3 …XPXP 1 2 3. n.

Slides:



Advertisements
Similar presentations
Data Analysis Lecture 8 Tijl De Bie. Dimensionality reduction How to deal with high-dimensional data? How to visualize it? How to explore it? Dimensionality.
Advertisements

Self-Organizing Maps Projection of p dimensional observations to a two (or one) dimensional grid space Constraint version of K-means clustering –Prototypes.
Covariance Matrix Applications
Pattern Recognition for the Natural Sciences Explorative Data Analysis Principal Component Analysis (PCA) Lutgarde Buydens, IMM, Analytical Chemistry.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
Terminology species data = the measured variables we want to explain (response or dependent variables) environmental data = the variables we use for explaining.
Maths for Computer Graphics
Multivariate Methods Pattern Recognition and Hypothesis Testing.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Lecture 6 Ordination Ordination contains a number of techniques to classify data according to predefined standards. The simplest ordination technique is.
Lecture 19 Quadratic Shapes and Symmetric Positive Definite Matrices Shang-Hua Teng.
Localization from Mere Connectivity Yi Shang (University of Missouri - Columbia); Wheeler Ruml (Palo Alto Research Center ); Ying Zhang; Markus Fromherz.
CHAPTER 19 Correspondence Analysis From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon.
Now Playing: My Mathematical Mind Spoon From Gimme Fiction Released May 10, 2005.
Manifold Learning: ISOMAP Alan O'Connor April 29, 2008.
Face Recognition Jeremy Wyatt.
Mathematics Operations Research Syllabus – Update 1.
SVD and PCA COS 323. Dimensionality Reduction Map points in high-dimensional space to lower number of dimensionsMap points in high-dimensional space to.
Introduction Given a Matrix of distances D, (which contains zeros in the main diagonal and is squared and symmetric), find variables which could be able,
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
SVD and PCA COS 323, Spring 05. SVD and PCA Principal Components Analysis (PCA): approximating a high-dimensional data set with a lower-dimensional subspacePrincipal.
Clustering and MDS Exploratory Data Analysis. Outline What may be hoped for by clustering What may be hoped for by clustering Representing differences.
Agenda Dimension reduction Principal component analysis (PCA)
Separate multivariate observations
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
The Tutorial of Principal Component Analysis, Hierarchical Clustering, and Multidimensional Scaling Wenshan Wang.
Summarized by Soo-Jin Kim
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Инвестиционный паспорт Муниципального образования «Целинский район»
CPSC 491 Xin Liu November 19, Norm Norm is a generalization of Euclidean length in vector space Definition 2.
(x – 8) (x + 8) = 0 x – 8 = 0 x + 8 = x = 8 x = (x + 5) (x + 2) = 0 x + 5 = 0 x + 2 = x = - 5 x = - 2.
Principal Component Analysis (PCA). Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite)
CpSc 881: Machine Learning PCA and MDS. 2 Copy Right Notice Most slides in this presentation are adopted from slides of text book and various sources.
Matrix Notation for Representing Vectors
Visualizing and Exploring Data 1. Outline 1.Introduction 2.Summarizing Data: Some Simple Examples 3.Tools for Displaying Single Variable 4.Tools for Displaying.
Analyzing Expression Data: Clustering and Stats Chapter 16.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Introduction to Multivariate Analysis and Multivariate Distances Hal Whitehead BIOL4062/5062.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
Factor & Cluster Analyses. Factor Analysis Goals Data Process Results.
Chapter 14: Multidimensional Scaling: British Water Voles and Voting in US Congress By: Laila, Rozie, Vimal.
Factor Analysis Basics. Why Factor? Combine similar variables into more meaningful factors. Reduce the number of variables dramatically while retaining.
Matrix Multiplication The Introduction. Look at the matrix sizes.
MAT 4725 Numerical Analysis Section 7.1 Part I Norms of Vectors and Matrices
Description of Multivariate Data. Multivariate Analysis The analysis of many variables.
= the matrix for T relative to the standard basis is a basis for R 2. B is the matrix for T relative to To find B, complete:
Principal Components Analysis ( PCA)
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Out of sample extension of PCA, Kernel PCA, and MDS WILSON A. FLORERO-SALINAS DAN LI MATH 285, FALL
Unsupervised Learning II Feature Extraction
照片档案整理 一、照片档案的含义 二、照片档案的归档范围 三、 卷内照片的分类、组卷、排序与编号 四、填写照片档案说明 五、照片档案编目及封面、备考填写 六、数码照片整理方法 七、照片档案的保管与保护.
공무원연금관리공단 광주지부 공무원대부등 공적연금 연계제도 공무원연금관리공단 광주지부. 공적연금 연계제도 국민연금과 직역연금 ( 공무원 / 사학 / 군인 / 별정우체국 ) 간의 연계가 이루어지지 않고 있 어 공적연금의 사각지대가 발생해 노후생활안정 달성 미흡 연계제도 시행전.
Жюль Верн ( ). Я мальчиком мечтал, читая Жюля Верна, Что тени вымысла плоть обретут для нас; Что поплывет судно громадней «Грейт Истерна»; Что.
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
Dimensionality Reduction
Vectors and Matrices I.
1.3 Vector Equations.
Principal Component Analysis (PCA)
Multidimensional array
Multidimensional Scaling
(A) Refined profile principal component analysis loading weights plot was used to derive insight into possible association of biomarkers. (A) Refined profile.
Unsupervised Learning and Clustering
Multidimensional Space,
Factor Analysis (Principal Components) Output
Comparison of distal gut metabolite profiles between Egyptian and U. S
Principal Component Analysis
Presentation transcript:

1 Data Analysis  Data Matrix Variables ObjectsX1X1 X2X2 X3X3 …XPXP n

2 Two Approaches  R-techniques -Comparing columns of the data matrix, i.e., the variables -Principal Component Analysis, Factor Analysis

3  Q-techniques -Comparing the rows of the data matrix, i.e., the different objects -Discriminant Analysis, Cluster Analysis, Multidimensional Scaling

4 Geometrical Ideas  R-techniques -The columns can be viewed as p points in a n- dimensional space, called Objects Space (or R space). -The interpretation of the correlation matrix r ij in the object space for the centered data matrix Y.

5  Q-techniques -The n rows can be viewed as n points in p- dimensional Q-space or variable space. -Look at the Euclidean distance between two rows, and