Next. A Big Thanks Again Prof. Jason Bohland Quantitative Neuroscience Laboratory Boston University.

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

3D Geometry for Computer Graphics
Chapter 28 – Part II Matrix Operations. Gaussian elimination Gaussian elimination LU factorization LU factorization Gaussian elimination with partial.
Tensors and Component Analysis Musawir Ali. Tensor: Generalization of an n-dimensional array Vector: order-1 tensor Matrix: order-2 tensor Order-3 tensor.
Lecture 19 Singular Value Decomposition
Slides by Olga Sorkine, Tel Aviv University. 2 The plan today Singular Value Decomposition  Basic intuition  Formal definition  Applications.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Computer Graphics Recitation 5.
3D Geometry for Computer Graphics
Unsupervised Learning - PCA The neural approach->PCA; SVD; kernel PCA Hertz chapter 8 Presentation based on Touretzky + various additions.
Computer Graphics Recitation 6. 2 Last week - eigendecomposition A We want to learn how the transformation A works:
Singular Value Decomposition (SVD) (see Appendix A.6, Trucco & Verri) CS485/685 Computer Vision Prof. George Bebis.
Face Recognition Jeremy Wyatt.
TFIDF-space  An obvious way to combine TF-IDF: the coordinate of document in axis is given by  General form of consists of three parts: Local weight.
Singular Value Decomposition
SVD and PCA COS 323. Dimensionality Reduction Map points in high-dimensional space to lower number of dimensionsMap points in high-dimensional space to.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
Microarray analysis Algorithms in Computational Biology Spring 2006 Written by Itai Sharon.
3D Geometry for Computer Graphics
3D Geometry for Computer Graphics
Lecture 20 SVD and Its Applications Shang-Hua Teng.
Boot Camp in Linear Algebra Joel Barajas Karla L Caballero University of California Silicon Valley Center October 8th, 2008.
1cs542g-term Notes  Extra class next week (Oct 12, not this Friday)  To submit your assignment: me the URL of a page containing (links to)
Matrices CS485/685 Computer Vision Dr. George Bebis.
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka Virginia de Sa (UCSD) Cogsci 108F Linear.
SVD(Singular Value Decomposition) and Its Applications
Summarized by Soo-Jin Kim
Linear Least Squares Approximation. 2 Definition (point set case) Given a point set x 1, x 2, …, x n  R d, linear least squares fitting amounts to find.
Chapter 2 Dimensionality Reduction. Linear Methods
Presented By Wanchen Lu 2/25/2013
Linear Algebra Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
CS246 Topic-Based Models. Motivation  Q: For query “car”, will a document with the word “automobile” be returned as a result under the TF-IDF vector.
Returning Back …. A Big Thanks Again Prof. Jason Bohland Quantitative Neuroscience Laboratory Boston University Prof. Matt Hibbs Jackson Labs.
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
D. van Alphen1 ECE 455 – Lecture 12 Orthogonal Matrices Singular Value Decomposition (SVD) SVD for Image Compression Recall: If vectors a and b have the.
SVD: Singular Value Decomposition
SINGULAR VALUE DECOMPOSITION (SVD)
CpSc 881: Machine Learning PCA and MDS. 2 Copy Right Notice Most slides in this presentation are adopted from slides of text book and various sources.
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Instructor: Mircea Nicolescu Lecture 8 CS 485 / 685 Computer Vision.
Oceanography 569 Oceanographic Data Analysis Laboratory Kathie Kelly Applied Physics Laboratory 515 Ben Hall IR Bldg class web site: faculty.washington.edu/kellyapl/classes/ocean569_.
Principal Components Analysis ( PCA)
Central limit theorem - go to web applet. Correlation maps vs. regression maps PNA is a time series of fluctuations in 500 mb heights PNA = 0.25 *
Unsupervised Learning II Feature Extraction
Boot Camp in Linear Algebra TIM 209 Prof. Ram Akella.
CS246 Linear Algebra Review. A Brief Review of Linear Algebra Vector and a list of numbers Addition Scalar multiplication Dot product Dot product as a.
Introduction to Vectors and Matrices
Principal Component Analysis (PCA)
Review of Linear Algebra
CS479/679 Pattern Recognition Dr. George Bebis
Lecture 8:Eigenfaces and Shared Features
Lecture: Face Recognition and Feature Reduction
Multivariate Analysis: Theory and Geometric Interpretation
SVD: Physical Interpretation and Applications
CS485/685 Computer Vision Dr. George Bebis
Recitation: SVD and dimensionality reduction
SVD, PCA, AND THE NFL By: Andrew Zachary.
Outline Singular Value Decomposition Example of PCA: Eigenfaces.
Feature space tansformation methods
Symmetric Matrices and Quadratic Forms
Principal Components What matters most?.
Lecture 13: Singular Value Decomposition (SVD)
Introduction to Vectors and Matrices
Subject :- Applied Mathematics
Symmetric Matrices and Quadratic Forms
Presentation transcript:

Next

A Big Thanks Again Prof. Jason Bohland Quantitative Neuroscience Laboratory Boston University

Expression energy for each gene (M=4,376) and for each voxel (N=51,533) Spatial Transcriptome

Clusters of Correlated Gene Expression

Gene Finder User navigates to voxel-of-interest in reference atlas volume and a fixed threshold AGEA correlation map appears Get a gene list from ABA is returned.

Large-scale data analysis How much structure is present across space and across genes? How would brain segment using ISH gene expression patterns ? Is there structure in patterns of expression of localized genes ? What do we learn from expression of genes in disorders?

 Combine gene volumes into a large matrix  Decompose the voxel x gene matrix M using the singular value decomposition (SVD)  Any m x n matrix M can be decomposed: where u n (x) are the spatial modes, and v n (g) are the gene modes, n 2 is the energy in the n th mode Large-scale Correlation

 Quality control → set of 3041 genes  Combine gene volumes into a large matrix  Decompose the voxel x gene matrix using singular value decomposition (SVD) voxels modes xx genes s.v.’s M ≈M ≈ “weight” spatial pattern gene pattern Large-scale Correlation

10 SVD A = USV T A (m by n) is any rectangular matrix (m rows and n columns) U (m by n) is an “orthogonal” matrix S (n by n) is a diagonal matrix V (n by n) is another orthogonal matrix Such decomposition always exists All matrices are real; m ≥ n

Why SVD ? Product of three orthogonal matrices Each voxel is a vector in 3041 dimensional space Looking for directions in vector space with most variability Each “mode” has 3 items – spatial pattern – voxels that are correlated gene pattern – weighing how much a gene expresses this mode a weight – how much of overall variability in data does this mode account for SVD modes are ordered – Each successive mode accounts for as much of the variability remaining in the data as is possible

N=271 before we get to 90% of the variance N=67 before we get to 80% of the variance Principal modes (SVD) Cerebral cortex Olfactory areas Hippocampus Retrohippocampal Striatum Pallidum Thalamus Hypothalamus Midbrain Pons Medulla Cerebellum All LH brain voxels plotted as projections on first 3 modes

Singular Value Decomposition

14 SVD for microarray data (Alter et al, PNAS 2000)

15 A = USV T A is any rectangular matrix (m ≥ n) Row space: vector subspace generated by the row vectors of A Column space: vector subspace generated by the column vectors of A The dimension of the row & column space is rank of the matrix A: r (≤ n) A is a linear transformation that maps a vector x in row space into vector Ax in column space

16 A = USV T U is an “orthogonal” matrix (m ≥ n) Column vectors of U form orthonormal basis for column space of A: U T U=I u 1, …, u n in U are eigenvectors of AA T – AA T =USV T VSU T =US 2 U T “Left singular vectors”

17 A = USV T V is an orthogonal matrix (n by n) Column vectors of V form an orthonormal basis for the row space of A: V T V=VV T =I v 1, …, v n in V are eigenvectors of A T A – A T A =VSU T USV T =VS 2 V T “Right singular vectors”

18 A = USV T S is diagonal matrix of non-negative singular values Sorted from largest to smallest Singular values are non-negative square root of corresponding eigenvalues of A T A and AA T

19 AV = US Means each Av i = s i u i Note A is linear map from row to column space A maps orthonormal basis {v i } in row space into orthonormal basis {u i } in column space Each component of u i is projection of a row onto vector v i

20 SVD of A (m by n): recap A = USV T = (big-"orthogonal")(diagonal)(sq- orthogonal) u 1, …, u m in U are eigenvectors of AA T v 1, …, v n in V are eigenvectors of A T A s 1, …, s n in S are nonnegative singular values of A AV = US means each Av i = s i u i “Every A is diagonalized by 2 orthogonal matrices”

21 Recap

22 Eigengenes

23 Copyright ©2000 by the National Academy of Sciences Alter, Orly et al. (2000) Proc. Natl. Acad. Sci. USA 97, Genes Ranking – Correlation of Top 2 Eigengenes

24 Copyright ©2000 by the National Academy of Sciences Alter, Orly et al. (2000) Proc. Natl. Acad. Sci. USA 97, Normalized elutriation expression in the subspace associated with the cell cycle

25 SVD in Row Space x y v1v1 x y This line segment that goes through origin approximates the original data set The projected data set approximates the original data set x y s1u1v1Ts1u1v1T A

26 SVD = PCA? Centered data x y x’ y’

27 x y x’ y’ y’’ x’’ Translation is not a linear operation, as it moves the origin ! PCA SVD SVD = PCA?

Returning Back …

 Quality control → set of 3041 genes  Combine gene volumes into a large matrix  Decompose the voxel x gene matrix using singular value decomposition (SVD) voxels modes xx genes s.v.’s M ≈M ≈ “weight” spatial pattern gene pattern Large-scale Correlation

Interpreting gene modes Spatial modes are easily visualized. Attempt to annotate eigenmodes using Gene Ontology (GO) annotations: Each GO term partitions gene list into two subsets: IN genes: Genes annotated by that GO term OUT genes: Genes not annotated by that GO term Each singular vector associates each subset above with a set of amplitudes Compare these amplitudes, asking whether ‘IN’ genes have larger magnitudes than ‘OUT’ genes use K-S test to test whether the amplitude distributions are different

Component annotations