Download presentation
Presentation is loading. Please wait.
1
Part 2: Graph Mining Tools - SVD and ranking
(C) C. Faloutsos, 2017 Part 2: Graph Mining Tools - SVD and ranking Christos Faloutsos CMU
2
Outline Introduction – Motivation Node importance (HITS, PageRank)
(C) C. Faloutsos, 2017 Outline Introduction – Motivation Node importance (HITS, PageRank) SVD definitions and properties HITS PageRank Conclusions Tepper, CMU, April 4 (C) C. Faloutsos, 2017
3
Node importance - Motivation:
Given a graph (eg., web pages containing the desirable query word) Q: Which node is the most important? Tepper, CMU, April 4 (C) C. Faloutsos, 2017
4
Node importance - Motivation:
Given a graph (eg., web pages containing the desirable query word) Q: Which node is the most important? A1: HITS (SVD = Singular Value Decomposition) A2: eigenvector (PageRank) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
5
Node importance - motivation
SVD and eigenvector analysis: very closely related Tepper, CMU, April 4 (C) C. Faloutsos, 2017
6
SVD - Detailed outline Motivation Definition - properties
Interpretation Complexity Case studies Tepper, CMU, April 4 (C) C. Faloutsos, 2017
7
SVD - Motivation problem #1: text - LSI: find ‘concepts’
problem #2: compression / dim. reduction Tepper, CMU, April 4 (C) C. Faloutsos, 2017
8
SVD - Motivation problem #1: text - LSI: find ‘concepts’
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
9
SVD - Motivation Customer-product, for recommendation system:
lettuce tomatos chicken bread beef vegetarians meat eaters Tepper, CMU, April 4 (C) C. Faloutsos, 2017
10
SVD - Motivation details problem #2: compress / reduce dimensionality
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
11
Problem - specs details ~10**6 rows; ~10**3 columns; no updates;
random access to any cell(s) ; small error: OK Tepper, CMU, April 4 (C) C. Faloutsos, 2017
12
details SVD - Motivation Tepper, CMU, April 4 (C) C. Faloutsos, 2017
13
details SVD - Motivation Tepper, CMU, April 4 (C) C. Faloutsos, 2017
14
SVD - Detailed outline Motivation Definition - properties
Interpretation Complexity Case studies Additional properties Tepper, CMU, April 4 (C) C. Faloutsos, 2017
15
SVD - Definition (reminder: matrix multiplication x = 3 x 2 2 x 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
16
SVD - Definition (reminder: matrix multiplication x = 3 x 2 2 x 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
17
SVD - Definition (reminder: matrix multiplication x = 3 x 2 2 x 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
18
SVD - Definition (reminder: matrix multiplication x = 3 x 2 2 x 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
19
SVD - Definition (reminder: matrix multiplication x =
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
20
A[n x m] = U[n x r] L [ r x r] (V[m x r])T
SVD - Definition A[n x m] = U[n x r] L [ r x r] (V[m x r])T A: n x m matrix (eg., n documents, m terms) U: n x r matrix (n documents, r concepts) L: r x r diagonal matrix (strength of each ‘concept’) (r : rank of the matrix) V: m x r matrix (m terms, r concepts) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
21
SVD - Definition A = U L VT - example: Tepper, CMU, April 4
(C) C. Faloutsos, 2017
22
SVD - Properties THEOREM [Press+92]: always possible to decompose matrix A into A = U L VT , where U, L, V: unique (*) U, V: column orthonormal (ie., columns are unit vectors, orthogonal to each other) UT U = I; VT V = I (I: identity matrix) L: singular are positive, and sorted in decreasing order Tepper, CMU, April 4 (C) C. Faloutsos, 2017
23
SVD - Example A = U L VT - example: CS x x = MD retrieval inf. lung
brain data CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017
24
SVD - Example A = U L VT - example: CS-concept MD-concept CS x x = MD
retrieval CS-concept inf. lung MD-concept brain data CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017
25
SVD - Example A = U L VT - example: doc-to-concept similarity matrix
retrieval CS-concept inf. lung MD-concept brain data CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017
26
‘strength’ of CS-concept
SVD - Example A = U L VT - example: retrieval ‘strength’ of CS-concept inf. lung brain data CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017
27
SVD - Example A = U L VT - example: term-to-concept similarity matrix
retrieval inf. lung brain data CS-concept CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017
28
SVD - Example A = U L VT - example: term-to-concept similarity matrix
retrieval inf. lung brain data CS-concept CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017
29
SVD - Detailed outline Motivation Definition - properties
Interpretation Complexity Case studies Additional properties Tepper, CMU, April 4 (C) C. Faloutsos, 2017
30
SVD - Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’:
U: document-to-concept similarity matrix V: term-to-concept sim. matrix L: its diagonal elements: ‘strength’ of each concept Tepper, CMU, April 4 (C) C. Faloutsos, 2017
31
SVD – Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’:
(C) C. Faloutsos, 2017 Faloutsos, Tong SVD – Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’: Q: if A is the document-to-term matrix, what is AT A? A: Q: A AT ? Tepper, CMU, April 4 (C) C. Faloutsos, 2017 31
32
SVD – Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’:
Faloutsos, Tong (C) C. Faloutsos, 2017 SVD – Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’: Q: if A is the document-to-term matrix, what is AT A? A: term-to-term ([m x m]) similarity matrix Q: A AT ? A: document-to-document ([n x n]) similarity matrix Tepper, CMU, April 4 (C) C. Faloutsos, 2017 -32 32
33
Copyright: Faloutsos, Tong (2009)
(C) C. Faloutsos, 2017 SVD properties V are the eigenvectors of the covariance matrix ATA U are the eigenvectors of the Gram (inner-product) matrix AAT Further reading: 1. Ian T. Jolliffe, Principal Component Analysis (2nd ed), Springer, 2002. 2. Gilbert Strang, Linear Algebra and Its Applications (4th ed), Brooks Cole, 2005. Tepper, CMU, April 4 Copyright: Faloutsos, Tong (2009) (C) C. Faloutsos, 2017 2-33 33
34
SVD - Interpretation #2 details
best axis to project on: (‘best’ = min sum of squares of projection errors) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
35
details SVD - Motivation Tepper, CMU, April 4 (C) C. Faloutsos, 2017
36
SVD - interpretation #2 details SVD: gives best axis to project
first singular vector v1 minimum RMS error Tepper, CMU, April 4 (C) C. Faloutsos, 2017
37
SVD - Interpretation #2 details Tepper, CMU, April 4
(C) C. Faloutsos, 2017
38
SVD - Interpretation #2 details A = U L VT - example: x = v1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
39
variance (‘spread’) on the v1 axis
details SVD - Interpretation #2 A = U L VT - example: variance (‘spread’) on the v1 axis x x = Tepper, CMU, April 4 (C) C. Faloutsos, 2017
40
SVD - Interpretation #2 details A = U L VT - example:
U L gives the coordinates of the points in the projection axis x x = Tepper, CMU, April 4 (C) C. Faloutsos, 2017
41
SVD - Interpretation #2 details More details
Q: how exactly is dim. reduction done? = x Tepper, CMU, April 4 (C) C. Faloutsos, 2017
42
SVD - Interpretation #2 details More details
Q: how exactly is dim. reduction done? A: set the smallest singular values to zero: x x = Tepper, CMU, April 4 (C) C. Faloutsos, 2017
43
SVD - Interpretation #2 details x x ~ Tepper, CMU, April 4
(C) C. Faloutsos, 2017
44
SVD - Interpretation #2 details x x ~ Tepper, CMU, April 4
(C) C. Faloutsos, 2017
45
SVD - Interpretation #2 details x x ~ Tepper, CMU, April 4
(C) C. Faloutsos, 2017
46
SVD - Interpretation #2 details ~ Tepper, CMU, April 4
(C) C. Faloutsos, 2017
47
SVD - Interpretation #2 details Exactly equivalent:
‘spectral decomposition’ of the matrix: x x = Tepper, CMU, April 4 (C) C. Faloutsos, 2017
48
SVD - Interpretation #2 details Exactly equivalent:
‘spectral decomposition’ of the matrix: l1 x x = u1 u2 l2 v1 v2 Tepper, CMU, April 4 (C) C. Faloutsos, 2017
49
SVD - Interpretation #2 details Exactly equivalent:
‘spectral decomposition’ of the matrix: m = u1 l1 vT1 + u2 l2 vT2 +... n Tepper, CMU, April 4 (C) C. Faloutsos, 2017
50
SVD - Interpretation #2 details Exactly equivalent:
‘spectral decomposition’ of the matrix: m r terms = u1 l1 vT1 + u2 l2 vT2 +... n n x 1 1 x m Tepper, CMU, April 4 (C) C. Faloutsos, 2017
51
SVD - Interpretation #2 details approximation / dim. reduction:
by keeping the first few terms (Q: how many?) m = u1 l1 vT1 + u2 l2 vT2 +... n assume: l1 >= l2 >= ... Tepper, CMU, April 4 (C) C. Faloutsos, 2017
52
SVD - Interpretation #2 details
A (heuristic - [Fukunaga]): keep 80-90% of ‘energy’ (= sum of squares of li ’s) m = u1 l1 vT1 + u2 l2 vT2 +... n assume: l1 >= l2 >= ... Tepper, CMU, April 4 (C) C. Faloutsos, 2017
53
SVD - Detailed outline Motivation Definition - properties
Interpretation #1: documents/terms/concepts #2: dim. reduction #3: picking non-zero, rectangular ‘blobs’ Complexity Case studies Additional properties Tepper, CMU, April 4 (C) C. Faloutsos, 2017
54
SVD - Interpretation #3 finds non-zero ‘blobs’ in a data matrix x x =
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
55
SVD - Interpretation #3 finds non-zero ‘blobs’ in a data matrix x x =
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
56
SVD - Interpretation #3 finds non-zero ‘blobs’ in a data matrix =
‘communities’ (bi-partite cores, here) Row 1 Row 4 Col 1 Col 3 Col 4 Row 5 Row 7 Tepper, CMU, April 4 (C) C. Faloutsos, 2017
57
SVD - Detailed outline Motivation Definition - properties
Interpretation Complexity Case studies Additional properties Tepper, CMU, April 4 (C) C. Faloutsos, 2017
58
SVD - Complexity O( n * m * m) or O( n * n * m) (whichever is less)
less work, if we just want singular values or if we want first k singular vectors or if the matrix is sparse [Berry] Implemented: in any linear algebra package (LINPACK, matlab, Splus, mathematica ...) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
59
SVD - conclusions so far
SVD: A= U L VT : unique (*) U: document-to-concept similarities V: term-to-concept similarities L: strength of each concept dim. reduction: keep the first few strongest singular values (80-90% of ‘energy’) SVD: picks up linear correlations SVD: picks up non-zero ‘blobs’ Tepper, CMU, April 4 (C) C. Faloutsos, 2017
60
SVD - detailed outline ... SVD properties case studies Conclusions
eigenSpokes – visualization Kleinberg’s algorithm Google’s algorithm Conclusions Tepper, CMU, April 4 (C) C. Faloutsos, 2017
61
(C) C. Faloutsos, 2017 EigenSpokes B. Aditya Prakash, Mukund Seshadri, Ashwin Sridharan, Sridhar Machiraju and Christos Faloutsos: EigenSpokes: Surprising Patterns and Scalable Community Chipping in Large Graphs, PAKDD 2010, Hyderabad, India, June 2010. Tepper, CMU, April 4 (C) C. Faloutsos, 2017
62
EigenSpokes Given a big graph (1M x 1M) Say something about it
A = U \Sigma U^T Tepper, CMU, April 4 (C) C. Faloutsos, 2017
63
EigenSpokes Eigenvectors of adjacency matrix
related to singular vectors (symmetric, undirected graph) A = U \Sigma U^T Tepper, CMU, April 4 (C) C. Faloutsos, 2017
64
EigenSpokes Eigenvectors of adjacency matrix
details EigenSpokes Eigenvectors of adjacency matrix related to singular vectors (symmetric, undirected graph) A = U \Sigma U^T \vec{u}_1 \vec{u}_i N N Tepper, CMU, April 4 (C) C. Faloutsos, 2017
65
EigenSpokes EE plot: Scatter plot of scores of u1 vs u2
(C) C. Faloutsos, 2017 EigenSpokes 2nd Principal component EE plot: Scatter plot of scores of u1 vs u2 One would expect Many origin A few scattered ~randomly u2 u1 1st Principal component Tepper, CMU, April 4 (C) C. Faloutsos, 2017
66
EigenSpokes EE plot: Scatter plot of scores of u1 vs u2
(C) C. Faloutsos, 2017 EigenSpokes EE plot: Scatter plot of scores of u1 vs u2 One would expect Many origin A few scattered ~randomly u1 u2 90o Tepper, CMU, April 4 (C) C. Faloutsos, 2017
67
EigenSpokes - pervasiveness
Present in mobile social graph across time and space Patent citation graph Tepper, CMU, April 4 (C) C. Faloutsos, 2017
68
EigenSpokes - explanation
Near-cliques, or near-bipartite-cores, loosely connected Tepper, CMU, April 4 (C) C. Faloutsos, 2017
69
EigenSpokes - explanation
Near-cliques, or near-bipartite-cores, loosely connected Tepper, CMU, April 4 (C) C. Faloutsos, 2017
70
EigenSpokes - explanation
Near-cliques, or near-bipartite-cores, loosely connected So what? Extract nodes with high scores high connectivity Good “communities” spy plot of top 20 nodes Tepper, CMU, April 4 (C) C. Faloutsos, 2017
71
Bipartite Communities!
patents from same inventor(s) cut-and-paste bibliography! magnified bipartite community Tepper, CMU, April 4 (C) C. Faloutsos, 2017
72
SVD - detailed outline ... SVD properties case studies Conclusions
Kleinberg’s algorithm Google’s algorithm Conclusions Tepper, CMU, April 4 (C) C. Faloutsos, 2017
73
Kleinberg’s algo (HITS)
Kleinberg, Jon (1998). Authoritative sources in a hyperlinked environment. Proc. 9th ACM-SIAM Symposium on Discrete Algorithms. Tepper, CMU, April 4 (C) C. Faloutsos, 2017
74
Recall: problem dfn Given a graph (eg., web pages containing the desirable query word) Q: Which node is the most important? Tepper, CMU, April 4 (C) C. Faloutsos, 2017
75
Kleinberg’s algorithm
Problem dfn: given the web and a query find the most ‘authoritative’ web pages for this query Step 0: find all pages containing the query terms Step 1: expand by one move forward and backward Tepper, CMU, April 4 (C) C. Faloutsos, 2017
76
Kleinberg’s algorithm
Step 1: expand by one move forward and backward Tepper, CMU, April 4 (C) C. Faloutsos, 2017
77
Kleinberg’s algorithm
on the resulting graph, give high score (= ‘authorities’) to nodes that many important nodes point to give high importance score (‘hubs’) to nodes that point to good ‘authorities’) hubs authorities Tepper, CMU, April 4 (C) C. Faloutsos, 2017
78
Kleinberg’s algorithm
observations recursive definition! each node (say, ‘i’-th node) has both an authoritativeness score ai and a hubness score hi Tepper, CMU, April 4 (C) C. Faloutsos, 2017
79
Kleinberg’s algorithm
Let E be the set of edges and A be the adjacency matrix: the (i,j) is 1 if the edge from i to j exists Let h and a be [n x 1] vectors with the ‘hubness’ and ‘authoritativiness’ scores. Then: Tepper, CMU, April 4 (C) C. Faloutsos, 2017
80
Kleinberg’s algorithm
Then: ai = hk + hl + hm that is ai = Sum (hj) over all j that (j,i) edge exists or a = AT h k i l m Tepper, CMU, April 4 (C) C. Faloutsos, 2017
81
Kleinberg’s algorithm
symmetrically, for the ‘hubness’: hi = an + ap + aq that is hi = Sum (qj) over all j that (i,j) edge exists or h = A a i n p q Tepper, CMU, April 4 (C) C. Faloutsos, 2017
82
Kleinberg’s algorithm
In conclusion, we want vectors h and a such that: h = A a a = AT h Recall properties: C(2): A [n x m] v1 [m x 1] = l1 u1 [n x 1] C(3): u1T A = l1 v1T = Converges, after a few iterations, to first left- and right- singular vectors Tepper, CMU, April 4 (C) C. Faloutsos, 2017
83
Kleinberg’s algorithm
In short, the solutions to h = A a a = AT h are the left- and right- singular-vectors of the adjacency matrix A. Starting from random a’ and iterating, we’ll eventually converge (Q: to which of all the singular-vectors? why?) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
84
Kleinberg’s algorithm
(Q: to which of all the singular-vectors? why?) A: to the ones of the strongest singular-value, because of property : (AT A ) k v’ ~ (constant) v1 Tepper, CMU, April 4 (C) C. Faloutsos, 2017
85
Kleinberg’s algorithm - results
Eg., for the query ‘java’: 0.251 java.sun.com (“the java developer”) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
86
Kleinberg’s algorithm - discussion
‘authority’ score can be used to find ‘similar pages’ (how?) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
87
SVD - detailed outline ... Complexity SVD properties Case studies
Kleinberg’s algorithm (HITS) Google’s algorithm Conclusions Tepper, CMU, April 4 (C) C. Faloutsos, 2017
88
PageRank (google) Brin, Sergey and Lawrence Page (1998). Anatomy of a Large-Scale Hypertextual Web Search Engine. 7th Intl World Wide Web Conf. Larry Page Sergey Brin Tepper, CMU, April 4 (C) C. Faloutsos, 2017
89
(C) C. Faloutsos, 2017 Problem: PageRank Given a directed graph, find its most interesting/central node A node is important, if it is connected with important nodes (recursive, but OK!) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
90
Problem: PageRank - solution
(C) C. Faloutsos, 2017 Problem: PageRank - solution Given a directed graph, find its most interesting/central node Proposed solution: Random walk; spot most ‘popular’ node (-> steady state prob. (ssp)) A node has high ssp, if it is connected with high ssp nodes (recursive, but OK!) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
91
(Simplified) PageRank algorithm
Let A be the adjacency matrix; let B be the transition matrix: transpose, column-normalized - then From B To 1 2 3 4 5 = Tepper, CMU, April 4 (C) C. Faloutsos, 2017
92
(Simplified) PageRank algorithm
B p = p B p = p 1 2 3 4 5 = Tepper, CMU, April 4 (C) C. Faloutsos, 2017
93
Definitions A Adjacency matrix (from-to)
D Degree matrix = (diag ( d1, d2, …, dn) ) B Transition matrix: to-from, column normalized B = AT D-1 Tepper, CMU, April 4 (C) C. Faloutsos, 2017
94
(Simplified) PageRank algorithm
(C) C. Faloutsos, 2017 (Simplified) PageRank algorithm B p = 1 * p thus, p is the eigenvector that corresponds to the highest eigenvalue (=1, since the matrix is column-normalized) Why does such a p exist? p exists if B is nxn, nonnegative, irreducible [Perron–Frobenius theorem] Tepper, CMU, April 4 (C) C. Faloutsos, 2017
95
(Simplified) PageRank algorithm
(C) C. Faloutsos, 2017 (Simplified) PageRank algorithm In short: imagine a particle randomly moving along the edges compute its steady-state probabilities (ssp) Full version of algo: with occasional random jumps Why? To make the matrix irreducible Tepper, CMU, April 4 (C) C. Faloutsos, 2017
96
Full Algorithm With probability 1-c, fly-out to a random node
(C) C. Faloutsos, 2017 Full Algorithm With probability 1-c, fly-out to a random node Then, we have p = c B p + (1-c)/n 1 => p = (1-c)/n [I - c B] -1 1 Tepper, CMU, April 4 (C) C. Faloutsos, 2017
97
Alternative notation M Modified transition matrix
M = c B + (1-c)/n 1 1T Then p = M p That is: the steady state probabilities = PageRank scores form the first eigenvector of the ‘modified transition matrix’ Tepper, CMU, April 4 (C) C. Faloutsos, 2017
98
Parenthesis: intuition behind eigenvectors
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
99
Formal definition If A is a (n x n) square matrix
(l , x) is an eigenvalue/eigenvector pair of A if A x = l x CLOSELY related to singular values: Tepper, CMU, April 4 (C) C. Faloutsos, 2017
100
Property #1: Eigen- vs singular-values
if B[n x m] = U[n x r] L [ r x r] (V[m x r])T then A = (BTB) is symmetric and C(4): BT B vi = li2 vi ie, v1 , v2 , ...: eigenvectors of A = (BTB) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
101
Property #2 If A[nxn] is a real, symmetric matrix
Then it has n real eigenvalues (if A is not symmetric, some eigenvalues may be complex) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
102
Intuition A as vector transformation x’ A x x’ = x 1 3 2 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
103
Intuition By defn., eigenvectors remain parallel to themselves (‘fixed points’) l1 v1 A v1 3.62 * = Tepper, CMU, April 4 (C) C. Faloutsos, 2017
104
Convergence Usually, fast: Tepper, CMU, April 4 (C) C. Faloutsos, 2017
105
Convergence Usually, fast: Tepper, CMU, April 4 (C) C. Faloutsos, 2017
106
Convergence Usually, fast: depends on ratio l2 l1 l1 : l2
Tepper, CMU, April 4 (C) C. Faloutsos, 2017
107
Kleinberg/google - conclusions
SVD helps in graph analysis: hub/authority scores: strongest left- and right- singular-vectors of the adjacency matrix random walk on a graph: steady state probabilities are given by the strongest eigenvector of the (modified) transition matrix Tepper, CMU, April 4 (C) C. Faloutsos, 2017
108
Conclusions SVD: a valuable tool
given a document-term matrix, it finds ‘concepts’ (LSI) ... and can find fixed-points or steady-state probabilities (google/ Kleinberg/ Markov Chains) … and can find blocks (‘eigenSpokes’) … and help visualize the graph Tepper, CMU, April 4 (C) C. Faloutsos, 2017
109
Conclusions cont’d (We didn’t discuss/elaborate, but, SVD
... can reduce dimensionality (KL) ... and can solve optimally over- and under-constraint linear systems (least squares / query feedbacks) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
110
PART2 END Tepper, CMU, April 4 (C) C. Faloutsos, 2017
111
Overall Conclusions Real graphs exhibit surprising patterns (power laws, shrinking diameter etc) SVD: a powerful tool (HITS, PageRank, eigenSpokes, dim. reduction) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
112
Our goal: Open source system for mining huge graphs:
(C) C. Faloutsos, 2017 11/18/2018 Our goal: Open source system for mining huge graphs: PEGASUS project (PEta GrAph mining System) code and papers Tepper, CMU, April 4 (C) C. Faloutsos, 2017
113
(C) C. Faloutsos, 2017 References D. Chakrabarti, C. Faloutsos: Graph Mining – Laws, Tools and Case Studies, Morgan Claypool 2012 Tepper, CMU, April 4 (C) C. Faloutsos, 2017
114
Project info Thanks to: NSF IIS-0705359, IIS-0534205,
(C) C. Faloutsos, 2017 11/18/2018 Project info Chau, Polo McGlohon, Mary Tsourakakis, Babis Akoglu, Leman Prakash, Aditya Tong, Hanghang Kang, U Thanks to: NSF IIS , IIS , CTA-INARC; Yahoo (M45), LLNL, IBM, SPRINT, INTEL, HP Tepper, CMU, April 4 (C) C. Faloutsos, 2017
115
References Berry, Michael: http://www.cs.utk.edu/~lsi/
(C) C. Faloutsos, 2017 References Berry, Michael: Brin, S. and L. Page (1998). Anatomy of a Large-Scale Hypertextual Web Search Engine. 7th Intl World Wide Web Conf. Tepper, CMU, April 4 (C) C. Faloutsos, 2017
116
References Christos Faloutsos, Searching Multimedia Databases by Content, Springer, (App. D) Fukunaga, K. (1990). Introduction to Statistical Pattern Recognition, Academic Press. I.T. Jolliffe Principal Component Analysis Springer, 2002 (2nd ed.) Tepper, CMU, April 4 (C) C. Faloutsos, 2017
117
References cont’d Kleinberg, J. (1998). Authoritative sources in a hyperlinked environment. Proc. 9th ACM-SIAM Symposium on Discrete Algorithms. Press, W. H., S. A. Teukolsky, et al. (1992). Numerical Recipes in C, Cambridge University Press. Tepper, CMU, April 4 (C) C. Faloutsos, 2017
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.