Presentation is loading. Please wait.

Presentation is loading. Please wait.

Part 2: Graph Mining Tools - SVD and ranking

Similar presentations


Presentation on theme: "Part 2: Graph Mining Tools - SVD and ranking"— Presentation transcript:

1 Part 2: Graph Mining Tools - SVD and ranking
(C) C. Faloutsos, 2017 Part 2: Graph Mining Tools - SVD and ranking Christos Faloutsos CMU

2 Outline Introduction – Motivation Node importance (HITS, PageRank)
(C) C. Faloutsos, 2017 Outline Introduction – Motivation Node importance (HITS, PageRank) SVD definitions and properties HITS PageRank Conclusions Tepper, CMU, April 4 (C) C. Faloutsos, 2017

3 Node importance - Motivation:
Given a graph (eg., web pages containing the desirable query word) Q: Which node is the most important? Tepper, CMU, April 4 (C) C. Faloutsos, 2017

4 Node importance - Motivation:
Given a graph (eg., web pages containing the desirable query word) Q: Which node is the most important? A1: HITS (SVD = Singular Value Decomposition) A2: eigenvector (PageRank) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

5 Node importance - motivation
SVD and eigenvector analysis: very closely related Tepper, CMU, April 4 (C) C. Faloutsos, 2017

6 SVD - Detailed outline Motivation Definition - properties
Interpretation Complexity Case studies Tepper, CMU, April 4 (C) C. Faloutsos, 2017

7 SVD - Motivation problem #1: text - LSI: find ‘concepts’
problem #2: compression / dim. reduction Tepper, CMU, April 4 (C) C. Faloutsos, 2017

8 SVD - Motivation problem #1: text - LSI: find ‘concepts’
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

9 SVD - Motivation Customer-product, for recommendation system:
lettuce tomatos chicken bread beef vegetarians meat eaters Tepper, CMU, April 4 (C) C. Faloutsos, 2017

10 SVD - Motivation details problem #2: compress / reduce dimensionality
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

11 Problem - specs details ~10**6 rows; ~10**3 columns; no updates;
random access to any cell(s) ; small error: OK Tepper, CMU, April 4 (C) C. Faloutsos, 2017

12 details SVD - Motivation Tepper, CMU, April 4 (C) C. Faloutsos, 2017

13 details SVD - Motivation Tepper, CMU, April 4 (C) C. Faloutsos, 2017

14 SVD - Detailed outline Motivation Definition - properties
Interpretation Complexity Case studies Additional properties Tepper, CMU, April 4 (C) C. Faloutsos, 2017

15 SVD - Definition (reminder: matrix multiplication x = 3 x 2 2 x 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

16 SVD - Definition (reminder: matrix multiplication x = 3 x 2 2 x 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

17 SVD - Definition (reminder: matrix multiplication x = 3 x 2 2 x 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

18 SVD - Definition (reminder: matrix multiplication x = 3 x 2 2 x 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

19 SVD - Definition (reminder: matrix multiplication x =
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

20 A[n x m] = U[n x r] L [ r x r] (V[m x r])T
SVD - Definition A[n x m] = U[n x r] L [ r x r] (V[m x r])T A: n x m matrix (eg., n documents, m terms) U: n x r matrix (n documents, r concepts) L: r x r diagonal matrix (strength of each ‘concept’) (r : rank of the matrix) V: m x r matrix (m terms, r concepts) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

21 SVD - Definition A = U L VT - example: Tepper, CMU, April 4
(C) C. Faloutsos, 2017

22 SVD - Properties THEOREM [Press+92]: always possible to decompose matrix A into A = U L VT , where U, L, V: unique (*) U, V: column orthonormal (ie., columns are unit vectors, orthogonal to each other) UT U = I; VT V = I (I: identity matrix) L: singular are positive, and sorted in decreasing order Tepper, CMU, April 4 (C) C. Faloutsos, 2017

23 SVD - Example A = U L VT - example: CS x x = MD retrieval inf. lung
brain data CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017

24 SVD - Example A = U L VT - example: CS-concept MD-concept CS x x = MD
retrieval CS-concept inf. lung MD-concept brain data CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017

25 SVD - Example A = U L VT - example: doc-to-concept similarity matrix
retrieval CS-concept inf. lung MD-concept brain data CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017

26 ‘strength’ of CS-concept
SVD - Example A = U L VT - example: retrieval ‘strength’ of CS-concept inf. lung brain data CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017

27 SVD - Example A = U L VT - example: term-to-concept similarity matrix
retrieval inf. lung brain data CS-concept CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017

28 SVD - Example A = U L VT - example: term-to-concept similarity matrix
retrieval inf. lung brain data CS-concept CS x x = MD Tepper, CMU, April 4 (C) C. Faloutsos, 2017

29 SVD - Detailed outline Motivation Definition - properties
Interpretation Complexity Case studies Additional properties Tepper, CMU, April 4 (C) C. Faloutsos, 2017

30 SVD - Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’:
U: document-to-concept similarity matrix V: term-to-concept sim. matrix L: its diagonal elements: ‘strength’ of each concept Tepper, CMU, April 4 (C) C. Faloutsos, 2017

31 SVD – Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’:
(C) C. Faloutsos, 2017 Faloutsos, Tong SVD – Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’: Q: if A is the document-to-term matrix, what is AT A? A: Q: A AT ? Tepper, CMU, April 4 (C) C. Faloutsos, 2017 31

32 SVD – Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’:
Faloutsos, Tong (C) C. Faloutsos, 2017 SVD – Interpretation #1 ‘documents’, ‘terms’ and ‘concepts’: Q: if A is the document-to-term matrix, what is AT A? A: term-to-term ([m x m]) similarity matrix Q: A AT ? A: document-to-document ([n x n]) similarity matrix Tepper, CMU, April 4 (C) C. Faloutsos, 2017 -32 32

33 Copyright: Faloutsos, Tong (2009)
(C) C. Faloutsos, 2017 SVD properties V are the eigenvectors of the covariance matrix ATA U are the eigenvectors of the Gram (inner-product) matrix AAT Further reading: 1. Ian T. Jolliffe, Principal Component Analysis (2nd ed), Springer, 2002. 2. Gilbert Strang, Linear Algebra and Its Applications (4th ed), Brooks Cole, 2005. Tepper, CMU, April 4 Copyright: Faloutsos, Tong (2009) (C) C. Faloutsos, 2017 2-33 33

34 SVD - Interpretation #2 details
best axis to project on: (‘best’ = min sum of squares of projection errors) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

35 details SVD - Motivation Tepper, CMU, April 4 (C) C. Faloutsos, 2017

36 SVD - interpretation #2 details SVD: gives best axis to project
first singular vector v1 minimum RMS error Tepper, CMU, April 4 (C) C. Faloutsos, 2017

37 SVD - Interpretation #2 details Tepper, CMU, April 4
(C) C. Faloutsos, 2017

38 SVD - Interpretation #2 details A = U L VT - example: x = v1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

39 variance (‘spread’) on the v1 axis
details SVD - Interpretation #2 A = U L VT - example: variance (‘spread’) on the v1 axis x x = Tepper, CMU, April 4 (C) C. Faloutsos, 2017

40 SVD - Interpretation #2 details A = U L VT - example:
U L gives the coordinates of the points in the projection axis x x = Tepper, CMU, April 4 (C) C. Faloutsos, 2017

41 SVD - Interpretation #2 details More details
Q: how exactly is dim. reduction done? = x Tepper, CMU, April 4 (C) C. Faloutsos, 2017

42 SVD - Interpretation #2 details More details
Q: how exactly is dim. reduction done? A: set the smallest singular values to zero: x x = Tepper, CMU, April 4 (C) C. Faloutsos, 2017

43 SVD - Interpretation #2 details x x ~ Tepper, CMU, April 4
(C) C. Faloutsos, 2017

44 SVD - Interpretation #2 details x x ~ Tepper, CMU, April 4
(C) C. Faloutsos, 2017

45 SVD - Interpretation #2 details x x ~ Tepper, CMU, April 4
(C) C. Faloutsos, 2017

46 SVD - Interpretation #2 details ~ Tepper, CMU, April 4
(C) C. Faloutsos, 2017

47 SVD - Interpretation #2 details Exactly equivalent:
‘spectral decomposition’ of the matrix: x x = Tepper, CMU, April 4 (C) C. Faloutsos, 2017

48 SVD - Interpretation #2 details Exactly equivalent:
‘spectral decomposition’ of the matrix: l1 x x = u1 u2 l2 v1 v2 Tepper, CMU, April 4 (C) C. Faloutsos, 2017

49 SVD - Interpretation #2 details Exactly equivalent:
‘spectral decomposition’ of the matrix: m = u1 l1 vT1 + u2 l2 vT2 +... n Tepper, CMU, April 4 (C) C. Faloutsos, 2017

50 SVD - Interpretation #2 details Exactly equivalent:
‘spectral decomposition’ of the matrix: m r terms = u1 l1 vT1 + u2 l2 vT2 +... n n x 1 1 x m Tepper, CMU, April 4 (C) C. Faloutsos, 2017

51 SVD - Interpretation #2 details approximation / dim. reduction:
by keeping the first few terms (Q: how many?) m = u1 l1 vT1 + u2 l2 vT2 +... n assume: l1 >= l2 >= ... Tepper, CMU, April 4 (C) C. Faloutsos, 2017

52 SVD - Interpretation #2 details
A (heuristic - [Fukunaga]): keep 80-90% of ‘energy’ (= sum of squares of li ’s) m = u1 l1 vT1 + u2 l2 vT2 +... n assume: l1 >= l2 >= ... Tepper, CMU, April 4 (C) C. Faloutsos, 2017

53 SVD - Detailed outline Motivation Definition - properties
Interpretation #1: documents/terms/concepts #2: dim. reduction #3: picking non-zero, rectangular ‘blobs’ Complexity Case studies Additional properties Tepper, CMU, April 4 (C) C. Faloutsos, 2017

54 SVD - Interpretation #3 finds non-zero ‘blobs’ in a data matrix x x =
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

55 SVD - Interpretation #3 finds non-zero ‘blobs’ in a data matrix x x =
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

56 SVD - Interpretation #3 finds non-zero ‘blobs’ in a data matrix =
‘communities’ (bi-partite cores, here) Row 1 Row 4 Col 1 Col 3 Col 4 Row 5 Row 7 Tepper, CMU, April 4 (C) C. Faloutsos, 2017

57 SVD - Detailed outline Motivation Definition - properties
Interpretation Complexity Case studies Additional properties Tepper, CMU, April 4 (C) C. Faloutsos, 2017

58 SVD - Complexity O( n * m * m) or O( n * n * m) (whichever is less)
less work, if we just want singular values or if we want first k singular vectors or if the matrix is sparse [Berry] Implemented: in any linear algebra package (LINPACK, matlab, Splus, mathematica ...) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

59 SVD - conclusions so far
SVD: A= U L VT : unique (*) U: document-to-concept similarities V: term-to-concept similarities L: strength of each concept dim. reduction: keep the first few strongest singular values (80-90% of ‘energy’) SVD: picks up linear correlations SVD: picks up non-zero ‘blobs’ Tepper, CMU, April 4 (C) C. Faloutsos, 2017

60 SVD - detailed outline ... SVD properties case studies Conclusions
eigenSpokes – visualization Kleinberg’s algorithm Google’s algorithm Conclusions Tepper, CMU, April 4 (C) C. Faloutsos, 2017

61 (C) C. Faloutsos, 2017 EigenSpokes B. Aditya Prakash, Mukund Seshadri, Ashwin Sridharan, Sridhar Machiraju and Christos Faloutsos: EigenSpokes: Surprising Patterns and Scalable Community Chipping in Large Graphs, PAKDD 2010, Hyderabad, India, June 2010. Tepper, CMU, April 4 (C) C. Faloutsos, 2017

62 EigenSpokes Given a big graph (1M x 1M) Say something about it
A = U \Sigma U^T Tepper, CMU, April 4 (C) C. Faloutsos, 2017

63 EigenSpokes Eigenvectors of adjacency matrix
related to singular vectors (symmetric, undirected graph) A = U \Sigma U^T Tepper, CMU, April 4 (C) C. Faloutsos, 2017

64 EigenSpokes Eigenvectors of adjacency matrix
details EigenSpokes Eigenvectors of adjacency matrix related to singular vectors (symmetric, undirected graph) A = U \Sigma U^T \vec{u}_1 \vec{u}_i N N Tepper, CMU, April 4 (C) C. Faloutsos, 2017

65 EigenSpokes EE plot: Scatter plot of scores of u1 vs u2
(C) C. Faloutsos, 2017 EigenSpokes 2nd Principal component EE plot: Scatter plot of scores of u1 vs u2 One would expect Many origin A few scattered ~randomly u2 u1 1st Principal component Tepper, CMU, April 4 (C) C. Faloutsos, 2017

66 EigenSpokes EE plot: Scatter plot of scores of u1 vs u2
(C) C. Faloutsos, 2017 EigenSpokes EE plot: Scatter plot of scores of u1 vs u2 One would expect Many origin A few scattered ~randomly u1 u2 90o Tepper, CMU, April 4 (C) C. Faloutsos, 2017

67 EigenSpokes - pervasiveness
Present in mobile social graph across time and space Patent citation graph Tepper, CMU, April 4 (C) C. Faloutsos, 2017

68 EigenSpokes - explanation
Near-cliques, or near-bipartite-cores, loosely connected Tepper, CMU, April 4 (C) C. Faloutsos, 2017

69 EigenSpokes - explanation
Near-cliques, or near-bipartite-cores, loosely connected Tepper, CMU, April 4 (C) C. Faloutsos, 2017

70 EigenSpokes - explanation
Near-cliques, or near-bipartite-cores, loosely connected So what? Extract nodes with high scores high connectivity Good “communities” spy plot of top 20 nodes Tepper, CMU, April 4 (C) C. Faloutsos, 2017

71 Bipartite Communities!
patents from same inventor(s) cut-and-paste bibliography! magnified bipartite community Tepper, CMU, April 4 (C) C. Faloutsos, 2017

72 SVD - detailed outline ... SVD properties case studies Conclusions
Kleinberg’s algorithm Google’s algorithm Conclusions Tepper, CMU, April 4 (C) C. Faloutsos, 2017

73 Kleinberg’s algo (HITS)
Kleinberg, Jon (1998). Authoritative sources in a hyperlinked environment. Proc. 9th ACM-SIAM Symposium on Discrete Algorithms. Tepper, CMU, April 4 (C) C. Faloutsos, 2017

74 Recall: problem dfn Given a graph (eg., web pages containing the desirable query word) Q: Which node is the most important? Tepper, CMU, April 4 (C) C. Faloutsos, 2017

75 Kleinberg’s algorithm
Problem dfn: given the web and a query find the most ‘authoritative’ web pages for this query Step 0: find all pages containing the query terms Step 1: expand by one move forward and backward Tepper, CMU, April 4 (C) C. Faloutsos, 2017

76 Kleinberg’s algorithm
Step 1: expand by one move forward and backward Tepper, CMU, April 4 (C) C. Faloutsos, 2017

77 Kleinberg’s algorithm
on the resulting graph, give high score (= ‘authorities’) to nodes that many important nodes point to give high importance score (‘hubs’) to nodes that point to good ‘authorities’) hubs authorities Tepper, CMU, April 4 (C) C. Faloutsos, 2017

78 Kleinberg’s algorithm
observations recursive definition! each node (say, ‘i’-th node) has both an authoritativeness score ai and a hubness score hi Tepper, CMU, April 4 (C) C. Faloutsos, 2017

79 Kleinberg’s algorithm
Let E be the set of edges and A be the adjacency matrix: the (i,j) is 1 if the edge from i to j exists Let h and a be [n x 1] vectors with the ‘hubness’ and ‘authoritativiness’ scores. Then: Tepper, CMU, April 4 (C) C. Faloutsos, 2017

80 Kleinberg’s algorithm
Then: ai = hk + hl + hm that is ai = Sum (hj) over all j that (j,i) edge exists or a = AT h k i l m Tepper, CMU, April 4 (C) C. Faloutsos, 2017

81 Kleinberg’s algorithm
symmetrically, for the ‘hubness’: hi = an + ap + aq that is hi = Sum (qj) over all j that (i,j) edge exists or h = A a i n p q Tepper, CMU, April 4 (C) C. Faloutsos, 2017

82 Kleinberg’s algorithm
In conclusion, we want vectors h and a such that: h = A a a = AT h Recall properties: C(2): A [n x m] v1 [m x 1] = l1 u1 [n x 1] C(3): u1T A = l1 v1T = Converges, after a few iterations, to first left- and right- singular vectors Tepper, CMU, April 4 (C) C. Faloutsos, 2017

83 Kleinberg’s algorithm
In short, the solutions to h = A a a = AT h are the left- and right- singular-vectors of the adjacency matrix A. Starting from random a’ and iterating, we’ll eventually converge (Q: to which of all the singular-vectors? why?) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

84 Kleinberg’s algorithm
(Q: to which of all the singular-vectors? why?) A: to the ones of the strongest singular-value, because of property : (AT A ) k v’ ~ (constant) v1 Tepper, CMU, April 4 (C) C. Faloutsos, 2017

85 Kleinberg’s algorithm - results
Eg., for the query ‘java’: 0.251 java.sun.com (“the java developer”) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

86 Kleinberg’s algorithm - discussion
‘authority’ score can be used to find ‘similar pages’ (how?) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

87 SVD - detailed outline ... Complexity SVD properties Case studies
Kleinberg’s algorithm (HITS) Google’s algorithm Conclusions Tepper, CMU, April 4 (C) C. Faloutsos, 2017

88 PageRank (google) Brin, Sergey and Lawrence Page (1998). Anatomy of a Large-Scale Hypertextual Web Search Engine. 7th Intl World Wide Web Conf. Larry Page Sergey Brin Tepper, CMU, April 4 (C) C. Faloutsos, 2017

89 (C) C. Faloutsos, 2017 Problem: PageRank Given a directed graph, find its most interesting/central node A node is important, if it is connected with important nodes (recursive, but OK!) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

90 Problem: PageRank - solution
(C) C. Faloutsos, 2017 Problem: PageRank - solution Given a directed graph, find its most interesting/central node Proposed solution: Random walk; spot most ‘popular’ node (-> steady state prob. (ssp)) A node has high ssp, if it is connected with high ssp nodes (recursive, but OK!) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

91 (Simplified) PageRank algorithm
Let A be the adjacency matrix; let B be the transition matrix: transpose, column-normalized - then From B To 1 2 3 4 5 = Tepper, CMU, April 4 (C) C. Faloutsos, 2017

92 (Simplified) PageRank algorithm
B p = p B p = p 1 2 3 4 5 = Tepper, CMU, April 4 (C) C. Faloutsos, 2017

93 Definitions A Adjacency matrix (from-to)
D Degree matrix = (diag ( d1, d2, …, dn) ) B Transition matrix: to-from, column normalized B = AT D-1 Tepper, CMU, April 4 (C) C. Faloutsos, 2017

94 (Simplified) PageRank algorithm
(C) C. Faloutsos, 2017 (Simplified) PageRank algorithm B p = 1 * p thus, p is the eigenvector that corresponds to the highest eigenvalue (=1, since the matrix is column-normalized) Why does such a p exist? p exists if B is nxn, nonnegative, irreducible [Perron–Frobenius theorem] Tepper, CMU, April 4 (C) C. Faloutsos, 2017

95 (Simplified) PageRank algorithm
(C) C. Faloutsos, 2017 (Simplified) PageRank algorithm In short: imagine a particle randomly moving along the edges compute its steady-state probabilities (ssp) Full version of algo: with occasional random jumps Why? To make the matrix irreducible Tepper, CMU, April 4 (C) C. Faloutsos, 2017

96 Full Algorithm With probability 1-c, fly-out to a random node
(C) C. Faloutsos, 2017 Full Algorithm With probability 1-c, fly-out to a random node Then, we have p = c B p + (1-c)/n 1 => p = (1-c)/n [I - c B] -1 1 Tepper, CMU, April 4 (C) C. Faloutsos, 2017

97 Alternative notation M Modified transition matrix
M = c B + (1-c)/n 1 1T Then p = M p That is: the steady state probabilities = PageRank scores form the first eigenvector of the ‘modified transition matrix’ Tepper, CMU, April 4 (C) C. Faloutsos, 2017

98 Parenthesis: intuition behind eigenvectors
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

99 Formal definition If A is a (n x n) square matrix
(l , x) is an eigenvalue/eigenvector pair of A if A x = l x CLOSELY related to singular values: Tepper, CMU, April 4 (C) C. Faloutsos, 2017

100 Property #1: Eigen- vs singular-values
if B[n x m] = U[n x r] L [ r x r] (V[m x r])T then A = (BTB) is symmetric and C(4): BT B vi = li2 vi ie, v1 , v2 , ...: eigenvectors of A = (BTB) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

101 Property #2 If A[nxn] is a real, symmetric matrix
Then it has n real eigenvalues (if A is not symmetric, some eigenvalues may be complex) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

102 Intuition A as vector transformation x’ A x x’ = x 1 3 2 1
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

103 Intuition By defn., eigenvectors remain parallel to themselves (‘fixed points’) l1 v1 A v1 3.62 * = Tepper, CMU, April 4 (C) C. Faloutsos, 2017

104 Convergence Usually, fast: Tepper, CMU, April 4 (C) C. Faloutsos, 2017

105 Convergence Usually, fast: Tepper, CMU, April 4 (C) C. Faloutsos, 2017

106 Convergence Usually, fast: depends on ratio l2 l1 l1 : l2
Tepper, CMU, April 4 (C) C. Faloutsos, 2017

107 Kleinberg/google - conclusions
SVD helps in graph analysis: hub/authority scores: strongest left- and right- singular-vectors of the adjacency matrix random walk on a graph: steady state probabilities are given by the strongest eigenvector of the (modified) transition matrix Tepper, CMU, April 4 (C) C. Faloutsos, 2017

108 Conclusions SVD: a valuable tool
given a document-term matrix, it finds ‘concepts’ (LSI) ... and can find fixed-points or steady-state probabilities (google/ Kleinberg/ Markov Chains) … and can find blocks (‘eigenSpokes’) … and help visualize the graph Tepper, CMU, April 4 (C) C. Faloutsos, 2017

109 Conclusions cont’d (We didn’t discuss/elaborate, but, SVD
... can reduce dimensionality (KL) ... and can solve optimally over- and under-constraint linear systems (least squares / query feedbacks) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

110 PART2 END Tepper, CMU, April 4 (C) C. Faloutsos, 2017

111 Overall Conclusions Real graphs exhibit surprising patterns (power laws, shrinking diameter etc) SVD: a powerful tool (HITS, PageRank, eigenSpokes, dim. reduction) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

112 Our goal: Open source system for mining huge graphs:
(C) C. Faloutsos, 2017 11/18/2018 Our goal: Open source system for mining huge graphs: PEGASUS project (PEta GrAph mining System) code and papers Tepper, CMU, April 4 (C) C. Faloutsos, 2017

113 (C) C. Faloutsos, 2017 References D. Chakrabarti, C. Faloutsos: Graph Mining – Laws, Tools and Case Studies, Morgan Claypool 2012 Tepper, CMU, April 4 (C) C. Faloutsos, 2017

114 Project info Thanks to: NSF IIS-0705359, IIS-0534205,
(C) C. Faloutsos, 2017 11/18/2018 Project info Chau, Polo McGlohon, Mary Tsourakakis, Babis Akoglu, Leman Prakash, Aditya Tong, Hanghang Kang, U Thanks to: NSF IIS , IIS , CTA-INARC; Yahoo (M45), LLNL, IBM, SPRINT, INTEL, HP Tepper, CMU, April 4 (C) C. Faloutsos, 2017

115 References Berry, Michael: http://www.cs.utk.edu/~lsi/
(C) C. Faloutsos, 2017 References Berry, Michael: Brin, S. and L. Page (1998). Anatomy of a Large-Scale Hypertextual Web Search Engine. 7th Intl World Wide Web Conf. Tepper, CMU, April 4 (C) C. Faloutsos, 2017

116 References Christos Faloutsos, Searching Multimedia Databases by Content, Springer, (App. D) Fukunaga, K. (1990). Introduction to Statistical Pattern Recognition, Academic Press. I.T. Jolliffe Principal Component Analysis Springer, 2002 (2nd ed.) Tepper, CMU, April 4 (C) C. Faloutsos, 2017

117 References cont’d Kleinberg, J. (1998). Authoritative sources in a hyperlinked environment. Proc. 9th ACM-SIAM Symposium on Discrete Algorithms. Press, W. H., S. A. Teukolsky, et al. (1992). Numerical Recipes in C, Cambridge University Press. Tepper, CMU, April 4 (C) C. Faloutsos, 2017


Download ppt "Part 2: Graph Mining Tools - SVD and ranking"

Similar presentations


Ads by Google