On Solving Linear Systems in Sublinear Time

Slides:



Advertisements
Similar presentations
The Data Stream Space Complexity of Cascaded Norms T.S. Jayram David Woodruff IBM Almaden.
Advertisements

Efficient classification for metric data Lee-Ad GottliebWeizmann Institute Aryeh KontorovichBen Gurion U. Robert KrauthgamerWeizmann Institute TexPoint.
Numerical Linear Algebra in the Streaming Model Ken Clarkson - IBM David Woodruff - IBM.
Lower Bounds for Additive Spanners, Emulators, and More David P. Woodruff MIT and Tsinghua University To appear in FOCS, 2006.
Vertex sparsifiers: New results from old techniques (and some open questions) Robert Krauthgamer (Weizmann Institute) Joint work with Matthias Englert,
05/11/2005 Carnegie Mellon School of Computer Science Aladdin Lamps 05 Combinatorial and algebraic tools for multigrid Yiannis Koutis Computer Science.
A Nonlinear Approach to Dimension Reduction Robert Krauthgamer Weizmann Institute of Science Joint work with Lee-Ad Gottlieb TexPoint fonts used in EMF.
Matroid Bases and Matrix Concentration
Fast Algorithms For Hierarchical Range Histogram Constructions
An Efficient Parallel Solver for SDD Linear Systems Richard Peng M.I.T. Joint work with Dan Spielman (Yale)
Metric Embeddings As Computational Primitives Robert Krauthgamer Weizmann Institute of Science [Based on joint work with Alex Andoni]
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
Iterative methods TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAA A A A A A A A.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey U. Waterloo Department of Combinatorics and Optimization Joint work with Isaac.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey University of Waterloo Department of Combinatorics and Optimization Joint.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey U. Waterloo C&O Joint work with Isaac Fung TexPoint fonts used in EMF. Read.
On Sketching Quadratic Forms Robert Krauthgamer, Weizmann Institute of Science Joint with: Alex Andoni, Jiecao Chen, Bo Qin, David Woodruff and Qin Zhang.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 8 May 4, 2005
Proximity algorithms for nearly-doubling spaces Lee-Ad Gottlieb Robert Krauthgamer Weizmann Institute TexPoint fonts used in EMF. Read the TexPoint manual.
Sublinear time algorithms Ronitt Rubinfeld Blavatnik School of Computer Science Tel Aviv University TexPoint fonts used in EMF. Read the TexPoint manual.
Matrix sparsification and the sparse null space problem Lee-Ad GottliebWeizmann Institute Tyler NeylonBynomial Inc. TexPoint fonts used in EMF. Read the.
1 On the Benefits of Adaptivity in Property Testing of Dense Graphs Joint work with Mira Gonen Dana Ron Tel-Aviv University.
Visual Recognition Tutorial
NEW APPROACH TO CALCULATION OF RANGE OF POLYNOMIALS USING BERNSTEIN FORMS.
Sketching and Embedding are Equivalent for Norms Alexandr Andoni (Simons Inst. / Columbia) Robert Krauthgamer (Weizmann Inst.) Ilya Razenshteyn (MIT, now.
Eigenvectors of random graphs: nodal domains James R. Lee University of Washington Yael Dekel and Nati Linial Hebrew University TexPoint fonts used in.
Algorithms on negatively curved spaces James R. Lee University of Washington Robert Krauthgamer IBM Research (Almaden) TexPoint fonts used in EMF. Read.
Yiannis Koutis , U of Puerto Rico, Rio Piedras
Ch 8.1 Numerical Methods: The Euler or Tangent Line Method
Approximating the MST Weight in Sublinear Time Bernard Chazelle (Princeton) Ronitt Rubinfeld (NEC) Luca Trevisan (U.C. Berkeley)
Quantum Computing MAS 725 Hartmut Klauck NTU TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A.
Ran El-Yaniv and Dmitry Pechyony Technion – Israel Institute of Technology, Haifa, Israel Transductive Rademacher Complexity and its Applications.
Expanders via Random Spanning Trees R 許榮財 R 黃佳婷 R 黃怡嘉.
Graph Sparsifiers Nick Harvey Joint work with Isaac Fung TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Spectrally Thin Trees Nick Harvey University of British Columbia Joint work with Neil Olver (MIT  Vrije Universiteit) TexPoint fonts used in EMF. Read.
Sublinear Algorithms via Precision Sampling Alexandr Andoni (Microsoft Research) joint work with: Robert Krauthgamer (Weizmann Inst.) Krzysztof Onak (CMU)
Numerical Methods.
C&O 355 Mathematical Programming Fall 2010 Lecture 16 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
C&O 355 Lecture 24 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A A A A A A A.
Quantum Computing MAS 725 Hartmut Klauck NTU
Forrelation: A Problem that Optimally Separates Quantum from Classical Computing.
Monte Carlo Linear Algebra Techniques and Their Parallelization Ashok Srinivasan Computer Science Florida State University
Sketching complexity of graph cuts Alexandr Andoni joint work with: Robi Krauthgamer, David Woodruff.
Generating Random Spanning Trees via Fast Matrix Multiplication Keyulu Xu University of British Columbia Joint work with Nick Harvey TexPoint fonts used.
Laplacian Matrices of Graphs: Algorithms and Applications ICML, June 21, 2016 Daniel A. Spielman.
Monte Carlo Linear Algebra Techniques and Their Parallelization Ashok Srinivasan Computer Science Florida State University
Laplacian Matrices of Graphs: Algorithms and Applications ICML, June 21, 2016 Daniel A. Spielman.
Stochastic Streams: Sample Complexity vs. Space Complexity
Visual Recognition Tutorial
New Characterizations in Turnstile Streams with Applications
Hans Bodlaender, Marek Cygan and Stefan Kratsch
Approximating the MST Weight in Sublinear Time
Solving Linear Systems Ax=b
Parallel Algorithm Design using Spectral Graph Theory
Path Coupling And Approximate Counting
Chapter 2 Fundamentals of the Analysis of Algorithm Efficiency
Density Independent Algorithms for Sparsifying
Sketching and Embedding are Equivalent for Norms
Nuclear Norm Heuristic for Rank Minimization
Matrix Martingales in Randomized Numerical Linear Algebra
Linear sketching over
On the effect of randomness on planted 3-coloring models
Linear sketching with parities
Streaming Symmetric Norms via Measure Concentration
Lecture 6: Counting triangles Dynamic graphs & sampling
A Numerical Analysis Approach to Convex Optimization
Lecture 15: Least Square Regression Metric Embeddings
Optimization on Graphs
Presentation transcript:

On Solving Linear Systems in Sublinear Time Alexandr Andoni, Columbia University Robert Krauthgamer, Weizmann Institute Yosef Pogrow, Weizmann Institute  Google WOLA 2019 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAA

Solving Linear Systems Input: 𝐴∈ ℝ 𝑛×𝑛 and 𝑏∈ ℝ 𝑛 Output: vector 𝑥 that solves 𝐴𝑥=𝑏 Many algorithms, different variants: Matrix 𝐴 is sparse, Laplacian, PSD etc. Bounded precision (solution 𝑥 is approximate) vs. exact arithmetic Significant progress: Linear system in Laplacian matrix 𝐿 𝐺 can be solved approximately in near-linear time 𝑂 (nnz 𝐿 𝐺 ⋅log 1 𝜖 ) [Spielman-Teng’04, …, Cohen-Kyng-Miller-Pachocky-Peng-Rao-Xu’14] Our focus: Sublinear running time On Solving Linear Systems in Sublinear Time

Sublinear-Time Solver Input: 𝐴∈ ℝ 𝑛×𝑛 , 𝑏∈ ℝ 𝑛 (also 𝜖>0) and 𝑖∈[𝑛] Output: approximate coordinate 𝑥 𝑖 from (any) solution 𝑥 ∗ to 𝐴𝑥=𝑏 Accuracy bound 𝑥 − 𝑥 ∗ ∞ ≤𝜖 𝑥 ∗ ∞ Formal requirement: There is a solution 𝑥 ∗ to the system, such that ∀𝑖∈ 𝑛 , Pr 𝑥 𝑖 − 𝑥 𝑖 ∗ ≤ 𝜖 𝑥 ∗ ∞ ≥ 3 4 Follows framework of Local Computation Algorithms (LCA), previously used for graph problems [Rubinfeld-Tamir-Vardi-Xie’10] On Solving Linear Systems in Sublinear Time

On Solving Linear Systems in Sublinear Time Motivation Fast quantum algorithms for solving linear systems and for machine learning problems [Harrow-Hassidim-Lloyd’09, …] Can we match their performance classically? Recent success story: quantum  classical algorithm [Tang’18] New direction in sublinear-time algorithms “Local” computation in numerical problems Compare computational models (representation, preprocessing), accuracy guarantees, input families (e.g., Laplacian vs. PSD) Known quantum algorithms have modeling requirements (e.g., quantum encoding of 𝑏) On Solving Linear Systems in Sublinear Time

Algorithm for Laplacians Informally: Can solve Laplacian systems of bounded-degree expander in polylog(n) time Key limitations: sparsity and condition number Notation: 𝐿 𝐺 =𝐷−𝐴 is the Laplacian matrix of graph 𝐺 𝐿 𝐺 + is its Moore-Penrose pseudo-inverse Theorem 1: Suppose the input is a 𝑑-regular 𝑛-vertex graph 𝐺, together with its condition number 𝜅>0, 𝑏∈ ℝ 𝑛 , 𝑢∈ 𝑛 and 𝜖>0. Our algorithm computes 𝑥 𝑢 ∈ℝ such that for 𝑥 ∗ = 𝐿 𝐺 + 𝑏, ∀𝑢∈ 𝑛 , Pr 𝑥 𝑢 − 𝑥 𝑢 ∗ ≤ 𝜖 𝑥 ∗ ∞ ≥ 3 4 , and runs in time 𝑂 (𝑑 𝜖 −2 𝑠 3 ) for 𝑠= 𝑂 (𝜅 log 𝑛 ). More inputs? Faster? On Solving Linear Systems in Sublinear Time

On Solving Linear Systems in Sublinear Time Some Extensions Can replace 𝑛 with 𝑏 0 Example: Effective resistance can be approximate (in expanders) in constant running time! 𝑅 eff (𝑢,𝑣)= 𝑒 𝑢 − 𝑒 𝑣 𝑇 𝐿 𝐺 + ( 𝑒 𝑢 − 𝑒 𝑣 ) Improved running time if Graph 𝐺 is preprocessed One can sample a neighbor in 𝐺, or Extends to Symmetric Diagonally Dominant (SDD) matrix 𝑆 𝜅 is condition number of 𝐷 −1/2 𝑆 𝐷 −1/2 On Solving Linear Systems in Sublinear Time

Lower Bound for PSD Systems Informally: Solving “similar” PSD systems requires polynomial time Similar = bounded condition number and sparsity Even if the matrix can be preprocessed Theorem 2: For certain invertible PSD matrices 𝑆, with bounded sparsity 𝑑 and condition number 𝜅, every randomized algorithm must query 𝑛 Ω(1/ 𝑑 2 ) coordinates of the input 𝑏. Here, the output is 𝑥 𝑢 ∈ℝ for a fixed 𝑢∈ 𝑛 , required to satisfy ∀𝑢∈ 𝑛 , Pr 𝑥 𝑢 − 𝑥 𝑢 ∗ ≤ 1 5 𝑥 ∗ ∞ ≥ 3 4 , for 𝑥 ∗ = 𝑆 −1 𝑏. In particular, 𝑆 may be preprocessed On Solving Linear Systems in Sublinear Time

Dependence on Condition Number Informally: Quadratic dependence on 𝜅 is necessary Our algorithmic bound O ( 𝜅 3 ) is near-optimal, esp. when matrix 𝑆 can be preprocessed Theorem 3: For certain graphs 𝐺 of maximum degree 4 and any condition number 𝜅>0, every randomized algorithm (for 𝐿 𝐺 ) with accuracy 𝜖= 1 log 𝑛 must probe Ω ( 𝜅 2 ) coordinates of the input 𝑏. Again, the output is 𝑥 𝑢 ∈ℝ for a fixed 𝑢∈ 𝑛 , required to satisfy ∀𝑢∈ 𝑛 , Pr 𝑥 𝑢 − 𝑥 𝑢 ∗ ≤ 1 log 𝑛 𝑥 ∗ ∞ ≥ 3 4 , for 𝑥 ∗ = 𝐿 𝐺 + 𝑏. In particular, 𝐺 may be preprocessed On Solving Linear Systems in Sublinear Time

Algorithmic Techniques Famous Monte-Carlo method of von Neumann and Ulam: Write matrix inverse by power series ∀ 𝑋 <1, 𝐼−𝑋 −1 = 𝑡≥0 𝑋 𝑡 then estimate it by random walks (in 𝑋) with unbiased expectation Inverting a Laplacian 𝐿 𝐺 =𝑑𝐼−𝐴 corresponds to summing walks in 𝐺 For us: view 𝑒 𝑢 𝑇 𝑡≥0 𝐴 𝑡 𝑏 as sum over all walks, estimate it by sampling (random walks) Need to control: number of walks and their length Large powers 𝑡> 𝑡 ∗ contribute relatively little (by condition number) Estimate truncated series (𝑡≤ 𝑡 ∗ ) by short random walks (by Chebyshev’s inequality) On Solving Linear Systems in Sublinear Time

Related Work – All Algorithmic Similar techniques were used before in related contexts but under different assumptions, models and analyses: Probabilistic log-space algorithms for approximating 𝐿 𝐺 + [Doron-Le Gall-Ta-Shma’17] Asks for entire matrix, uses many long random walks (independent of 𝜅) Local solver for Laplacian systems with boundary conditions [Chung-Simpson’15] Solver relies on a different power series and random walks Local solver for PSD systems [Shyamkumar-Banerjee-Lofgren’16] Polynomial time nnz 𝑆 2/3 under assumptions like bounded matrix norm and random 𝑢∈ 𝑛 Local solver for Pagerank [Bressan-Peserico-Pretto’18, Borgs-Brautbar-Chayes-Teng’14] Polynomial time O( 𝑛 2/3 ) and O( nd 1/2 ) for certain matrices (non-symmetric but by definition are diagonally-dominant) On Solving Linear Systems in Sublinear Time

Lower Bound Techniques PSD lower bound: Take Laplacian of 2𝑑-regular expander but with: high girth, edges signed ±1 at random, and 𝑂( 𝑑 ) on the diagonal (PSD but not Laplacian) The graph looks like a tree locally Up to radius Θ log 𝑛 around 𝑢 Set 𝑏 𝑤 =±1 for 𝑤 at distance 𝑟, and 0 otherwise Signs have small bias 𝛿≈ 𝑑 −𝑟/2 Recovering it requires reading Ω( 𝛿 −2 ) entries Using inversion formula, 𝑥 𝑢 ≈ average of 𝑏 𝑤 ‘s Condition number lower bound: Take two 3-regular expanders connected by a matching of size 𝑛/𝜅 Let 𝑏 𝑤 =±1 with slight bias inside each expander 𝑢 𝑟 𝑏 𝑤 =0 𝑏 𝑤 =±1 𝑏 𝑤 =0 On Solving Linear Systems in Sublinear Time

On Solving Linear Systems in Sublinear Time Further Questions Accuracy guarantees Different norms? Condition number of 𝑆 instead of 𝐷 −1/2 𝑆 𝐷 −1/2 ? Other representations (input/output models)? Access the input 𝑏 via random sampling? Sample from the output 𝑥? Other numerical problems? Thank You! On Solving Linear Systems in Sublinear Time