Eigenvectors of random graphs: nodal domains James R. Lee University of Washington Yael Dekel and Nati Linial Hebrew University TexPoint fonts used in.

Slides:



Advertisements
Similar presentations
Matroid Bases and Matrix Concentration
Advertisements

Inapproximability of MAX-CUT Khot,Kindler,Mossel and O ’ Donnell Moshe Ben Nehemia June 05.
TexPoint fonts used in EMF.
Gibbs sampler - simple properties It’s not hard to show that this MC chain is aperiodic. Often is reversible distribution. If in addition the chain is.
Week 4 – Random Graphs Dr. Anthony Bonato Ryerson University AM8002 Fall 2014.
Some Graph Problems. LINIAL’S CONJECTURE Backgound: In a partially ordered set we have Dilworth’s Theorem; The largest size of an independent set (completely.
Matrix Concentration Nick Harvey University of British Columbia TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Online Social Networks and Media. Graph partitioning The general problem – Input: a graph G=(V,E) edge (u,v) denotes similarity between u and v weighted.
Iterative methods TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAA A A A A A A A.
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Analysis of Network Diffusion and Distributed Network Algorithms Rajmohan Rajaraman Northeastern University, Boston May 2012 Chennai Network Optimization.
Complexity 16-1 Complexity Andrei Bulatov Non-Approximability.
1 List Coloring and Euclidean Ramsey Theory TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A Noga Alon, Tel Aviv.
Graph Clustering. Why graph clustering is useful? Distance matrices are graphs  as useful as any other clustering Identification of communities in social.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey University of Waterloo Department of Combinatorics and Optimization Joint.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey U. Waterloo C&O Joint work with Isaac Fung TexPoint fonts used in EMF. Read.
CS 584. Review n Systems of equations and finite element methods are related.
Computability and Complexity 15-1 Computability and Complexity Andrei Bulatov NP-Completeness.
Segmentation Graph-Theoretic Clustering.
Job Scheduling Lecture 19: March 19. Job Scheduling: Unrelated Multiple Machines There are n jobs, each job has: a processing time p(i,j) (the time to.
Expanders Eliyahu Kiperwasser. What is it? Expanders are graphs with no small cuts. The later gives several unique traits to such graph, such as: – High.
Small Subgraphs in Random Graphs and the Power of Multiple Choices The Online Case Torsten Mütze, ETH Zürich Joint work with Reto Spöhel and Henning Thomas.
Ramanujan Graphs of Every Degree Adam Marcus (Crisply, Yale) Daniel Spielman (Yale) Nikhil Srivastava (MSR India)
Maximal Independent Set Distributed Algorithms for Multi-Agent Networks Instructor: K. Sinan YILDIRIM.
Finding a maximum independent set in a sparse random graph Uriel Feige and Eran Ofek.
cover times, blanket times, and majorizing measures Jian Ding U. C. Berkeley James R. Lee University of Washington Yuval Peres Microsoft Research TexPoint.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University COT 5410 – Spring 2004.
The Theory of NP-Completeness 1. Nondeterministic algorithms A nondeterminstic algorithm consists of phase 1: guessing phase 2: checking If the checking.
Graph Sparsifiers Nick Harvey University of British Columbia Based on joint work with Isaac Fung, and independent work of Ramesh Hariharan & Debmalya Panigrahi.
Small subgraphs in the Achlioptas process Reto Spöhel, ETH Zürich Joint work with Torsten Mütze and Henning Thomas TexPoint fonts used in EMF. Read the.
Nattee Niparnan. Easy & Hard Problem What is “difficulty” of problem? Difficult for computer scientist to derive algorithm for the problem? Difficult.
Ran El-Yaniv and Dmitry Pechyony Technion – Israel Institute of Technology, Haifa, Israel Transductive Rademacher Complexity and its Applications.
Graph Sparsifiers Nick Harvey Joint work with Isaac Fung TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
1 Burning a graph as a model of social contagion Anthony Bonato Ryerson University Institute of Software Chinese Academy of Sciences.
PODC Distributed Computation of the Mode Fabian Kuhn Thomas Locher ETH Zurich, Switzerland Stefan Schmid TU Munich, Germany TexPoint fonts used in.
Vector Norms and the related Matrix Norms. Properties of a Vector Norm: Euclidean Vector Norm: Riemannian metric:
Yuval Peled, HUJI Joint work with Nati Linial, Benny Sudakov, Hao Huang and Humberto Naves.
Artur Czumaj DIMAP DIMAP (Centre for Discrete Maths and it Applications) Computer Science & Department of Computer Science University of Warwick Testing.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University.
Miniconference on the Mathematics of Computation
1 How to burn a graph Anthony Bonato Ryerson University GRASCan 2015.
Graphs, Vectors, and Matrices Daniel A. Spielman Yale University AMS Josiah Willard Gibbs Lecture January 6, 2016.
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
Introduction to Graph Theory
 In the previews parts we have seen some kind of segmentation method.  In this lecture we will see graph cut, which is a another segmentation method.
Network Theory: Community Detection Dr. Henry Hexmoor Department of Computer Science Southern Illinois University Carbondale.
Multi-way spectral partitioning and higher-order Cheeger inequalities University of Washington James R. Lee Stanford University Luca Trevisan Shayan Oveis.
Presented by Alon Levin
A Tutorial on Spectral Clustering Ulrike von Luxburg Max Planck Institute for Biological Cybernetics Statistics and Computing, Dec. 2007, Vol. 17, No.
Expanders and Ramanujan Graphs Mike Krebs, Cal State LA For slideshow: click “Research and Talks” from
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
Laplacian Matrices of Graphs: Algorithms and Applications ICML, June 21, 2016 Daniel A. Spielman.
Document Clustering with Prior Knowledge Xiang Ji et al. Document Clustering with Prior Knowledge. SIGIR 2006 Presenter: Suhan Yu.
CHAPTER SIX T HE P ROBABILISTIC M ETHOD M1 Zhang Cong 2011/Nov/28.
Laplacian Matrices of Graphs: Algorithms and Applications ICML, June 21, 2016 Daniel A. Spielman.
Random Walk for Similarity Testing in Complex Networks
Dimension reduction for finite trees in L1
Spectral Clustering.
Computability and Complexity
Structural Properties of Low Threshold Rank Graphs
Segmentation Graph-Theoretic Clustering.
Grouping.
4.a The random waves model
On the effect of randomness on planted 3-coloring models
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
Discrete Mathematics and its Applications Lecture 3 – ILT model
Modelling and Searching Networks Lecture 5 – Random graphs
Discrete Mathematics and its Applications Lecture 5 – Random graphs
On Solving Linear Systems in Sublinear Time
Presentation transcript:

eigenvectors of random graphs: nodal domains James R. Lee University of Washington Yael Dekel and Nati Linial Hebrew University TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A A A

preliminaries Random graphs edges present with probability prandom d-regular graph Adjacency matrix Eigenvectors A (non-zero) function f : V  R is an eigenvector of G if there exists an (eigenvalue) for which for every x 2 V, where  (x) is the set of neighbors of x.

eigenvaules of random graphs/discrete matrices - Let’s arrange the eigenvalues of matrices so that Much is known about the large eigenvalues of random graphs, e.g. G(n,½) Wigner semi-circle law Füredi-Komlos more recently, the small (magnitude) eigenvalues of random non-symmetric discrete matrices Rudelson 06, Tao-Vu 06, Rudelson-Vershynin 07 (Littlewood-Offord estimates) T RACE M ETHOD Litvak-Pajor-Rudelson-Tomczak-Jaegermann 05 for singular values of rectangular matrices … but significantly less is understood about the eigenvectors.

spectral analysis In many areas such as machine learning and computer vision, eigenvectors of graphs are the primary tools for tasks like partitioning and clustering. [Shi and Malik (image segmentation); Coifman et. al (PDE, machine learning); Pothen, Simon and Lou (matrix sparsification)] Heuristics for random instances of NP-hard problems, e.g. - Refuting random 3-SAT above the threshold - Planted cliques, bisections, assignments, colorings

are eigenvectors uniform on the sphere? If we scale the (non-first) eigenvectors of G(n,½) so they lie on S n-1, do they behave like random vectors on the sphere? ? For example, do we (almost surely) have… or X open problem: Discrete version of “quantum chaos” (?)

are eigenvectors uniform on the sphere? Nodal domains If f : V  R is an eigenvector of a graph G, then f partitions G into maximal connected components on which f has constant sign (say, positive vs. non-positive). Graph with positive and non-positive nodes marked. So this graph/eigenvector pair has 6 domains. Our question: What is the nodal domain structure of the eigenvectors of G(n,p)? If we scale the (non-first) eigenvectors of G(n,½) so they lie on S n-1, do they behave like random vectors on the sphere? Observation: If we choose a random vector on S n-1 and a random graph, then almost surely the number of domains is precisely 2.

nodal domains - If f k is the k th eigenvector of G, then a discrete version [Davies-Leydold-Stadler] of Courant’s nodal domain theorem (from Riemannian geometry) says that f k has at most k nodal domains. observations: - If G has 2N nodal domains, then it has an independent set of size N, hence N = O(log n)/p. theorem: Almost surely, every eigenvector of G(n,p) has 2 primary nodal domains, along with (possibly) O( 1 /p) exceptional vertices. Experiments suggest that there are at most 2 nodal domains even for:

nodal domains theorem: Almost surely, every eigenvector of G(n,p) has 2 primary nodal domains, along with (possibly) O( 1 /p) exceptional vertices. Experiments suggest that there are at most 2 nodal domains even for: probability of exceptional vertex number of nodes can be a delicate issue: In the combinatorial Laplacian of G(n,½), exceptional vertices can occur (it’s always the vertex of max degree in the largest eigenvalue)

nodal domains theorem: main lemma (2-norm can’t vanish on large subsets): Almost surely, for every (non-first) eigenvector f of G(n,p) and every subset of vertices S With | S | ¸ ( 1 -  )n, we have || f | S || 2 ¸  p (  ) where  p (  ) ! 1 as  ! 0 and  p (  ) > 0 for  < (The point is that  p (  ) is independent of n.) follows from… x Almost surely, every eigenvector of G(n,p) has 2 primary nodal domains, along with (possibly) O( 1 /p) exceptional vertices.

the main lemma and LPRT main lemma: Almost surely, for every (non-first) eigenvector f of G(n,p) and every subset of vertices S With | S | ¸ ( 1 -  )n, we have || f | S || 2 ¸  p (  ) where  p (  ) ! 1 as  ! 0 and  p (  ) > 0 for  < (The point is that  p (  ) is independent of n.) Consider p = ½ and | S | =0.99n. S z VnSVnS

the main lemma and LPRT main lemma: Almost surely, for every (non-first) eigenvector f of G(n,p) and every subset of vertices S With | S | ¸ ( 1 -  )n, we have || f | S || 2 ¸  p (  ) where  p (  ) ! 1 as  ! 0 and  p (  ) > 0 for  < (The point is that  p (  ) is independent of n.) Consider p = ½ and | S | =0.99n. S z VnSVnS B is i.i.d. The above inequality yields but this almost surely impossible (even taking a union bound over all S’s)

lower bounding singular values Want to show that for a ( 1 +  ) n £ n random sign matrix B, Want to argue that is often large for i.i.d. signs {  1, …,  n } The vectors and have very different behaviors. As   0, need a very good understanding of “bad” vectors. For eps> 1, this is easy (Payley-Zygmund, Chernoff, union bound over a net) For 0<eps< 1, this requires also a quantitative CLT (for the “spread” vectors) [LPRT] For eps=0, requires a deeper understanding of the additive structure of the coordinates Tao-Vu 06 showed that this is related to the additive structure of the coordinates, e.g. whether (rescaled) coordinates lie in arithmetic progression. (See Rudelson-Vershynin for state of the art)

only the beginning… - Tightening nodal domain structure (e.g. no exceptional vertices), e.g. prove: - We’re missing something big (as experiments show) The case of G n,d : E.g. is the adjacency matrix of a random 3-regular graph almost surely non-singular? d=3d=4 d=5