On Solving Linear Systems in Sublinear Time Alexandr Andoni, Columbia University Robert Krauthgamer, Weizmann Institute Yosef Pogrow, Weizmann Institute ๏ Google WOLA 2019
Solving Linear Systems Input: ๐ต โ โ ๐ร๐ and ๐ โ โ ๐ ๏ฎ Output: vector ๐ฆ that solves ๐ต๐ฆ = ๐ ๏ฎ Many algorithms, different variants: ๏ฎ Matrix ๐ต is sparse, Laplacian, PSD etc. ๏ฑ Bounded precision (solution ๐ฆ is approximate) vs. exact arithmetic ๏ฑ Significant progress: Linear system in Laplacian matrix ๐ ๐ป can be ๏ฎ 1 solved approximately in near-linear time เทจ ๐(nnz ๐ ๐ป โ log ๐ ) [Spielman- Teng โ 04, โฆ , Cohen-Kyng-Miller-Pachocky-Peng-Rao-Xu โ 14] Our focus: Sublinear running time On Solving Linear Systems in Sublinear Time 2
Sublinear-Time Solver Input: ๐ต โ โ ๐ร๐ , ๐ โ โ ๐ (also ๐ > 0 ) and ๐ โ [๐] ๏ฎ ๐ฆ ๐ from (any) solution ๐ฆ โ to ๐ต๐ฆ = ๐ Output: approximate coordinate เท ๏ฎ ๐ฆ โ ๐ฆ โ โ โค ๐ ๐ฆ โ โ Accuracy bound เท ๏ฑ Formal requirement: There is a solution ๐ฆ โ to the system, such that ๏ฎ โ โค ๐ ๐ฆ โ โ โฅ 3 โ๐ โ ๐ , Pr เท ๐ฆ ๐ โ ๐ฆ ๐ 4 Follows framework of Local Computation Algorithms (LCA), ๏ฎ previously used for graph problems [Rubinfeld-Tamir-Vardi-Xie โ 10] On Solving Linear Systems in Sublinear Time 3
Motivation Fast quantum algorithms for solving linear systems and for machine ๏ฎ learning problems [Harrow-Hassidim-Lloyd โ 09, โฆ ] Can we match their performance classically? ๏ฑ Recent success story: quantum ๏ classical algorithm [Tang โ 18] ๏ฑ New direction in sublinear-time algorithms ๏ฎ โ Local โ computation in numerical problems ๏ฑ Compare computational models (representation, preprocessing), ๏ฑ accuracy guarantees, input families (e.g., Laplacian vs. PSD) Known quantum algorithms have modeling requirements (e.g., quantum ๏ฑ encoding of ๐ ) On Solving Linear Systems in Sublinear Time 4
Algorithm for Laplacians Informally: Can solve Laplacian systems of bounded-degree ๏ฎ expander in polylog(n) time Key limitations: sparsity and condition number ๏ฑ Notation: ๏ฎ ๐ ๐ป = ๐ธ โ ๐ต is the Laplacian matrix of graph ๐ป ๏ฑ + is its Moore-Penrose pseudo-inverse ๐ ๐ป ๏ฑ Theorem 1: Suppose the input is a ๐ -regular ๐ -vertex graph ๐ป , ๏ฎ together with its condition number ๐ > 0 , ๐ โ โ ๐ , ๐ฃ โ ๐ and ๐ > 0 . ๐ฆ ๐ฃ โ โ such that for ๐ฆ โ = ๐ ๐ป + ๐ , Our algorithm computes เท โ โค ๐ ๐ฆ โ โ โฅ 3 โ๐ฃ โ ๐ , Pr เท ๐ฆ ๐ฃ โ ๐ฆ ๐ฃ 4 , and runs in time เทจ ๐(๐๐ โ2 ๐ก 3 ) for ๐ก = เทจ ๐(๐ log ๐) . More inputs? Faster? On Solving Linear Systems in Sublinear Time 5
Some Extensions Can replace ๐ with ๐ 0 ๏ฎ Example: Effective resistance can be approximate (in expanders) in ๏ฑ constant running time! ๐eff(๐ฃ, ๐ค) = ๐ ๐ฃ โ ๐ ๐ค ๐ ๐ ๐ป + (๐ ๐ฃ โ ๐ ๐ค ) Improved running time if ๏ฎ Graph ๐ป is preprocessed ๏ฑ One can sample a neighbor in ๐ป , or ๏ฑ Extends to Symmetric Diagonally Dominant (SDD) matrix ๐ ๏ฎ ๐ is condition number of ๐ธ โ1/2 ๐๐ธ โ1/2 ๏ฑ On Solving Linear Systems in Sublinear Time 6
Lower Bound for PSD Systems Informally: Solving โ similar โ PSD systems requires polynomial time ๏ฎ Similar = bounded condition number and sparsity ๏ฑ Even if the matrix can be preprocessed ๏ฑ Theorem 2: For certain invertible PSD matrices ๐ , with bounded ๏ฎ sparsity ๐ and condition number ๐ , every randomized algorithm must query ๐ ฮฉ(1/๐ 2 ) coordinates of the input ๐ . Here, the output is เท ๐ฆ ๐ฃ โ โ for a fixed ๐ฃ โ ๐ , required to satisfy ๏ฎ โ โค 1 3 5 ๐ฆ โ โ โฅ โ๐ฃ โ ๐ , Pr ๐ฆ ๐ฃ โ ๐ฆ ๐ฃ เท 4 , for ๐ฆ โ = ๐ โ1 ๐ . In particular, ๐ may be preprocessed ๏ฎ On Solving Linear Systems in Sublinear Time 7
Dependence on Condition Number Informally: Quadratic dependence on ๐ is necessary ๏ฎ Our algorithmic bound เทฉ O(๐ 3 ) is near-optimal, esp. when matrix ๐ can be ๏ฑ preprocessed Theorem 3: For certain graphs ๐ป of maximum degree 4 and any ๏ฎ condition number ๐ > 0 , every randomized algorithm (for ๐ ๐ป ) with 1 log ๐ must probe เทฉ ฮฉ(๐ 2 ) coordinates of the input ๐ . accuracy ๐ = Again, the output is เท ๐ฆ ๐ฃ โ โ for a fixed ๐ฃ โ ๐ , required to satisfy ๏ฎ โ โค 1 3 log ๐ ๐ฆ โ โ โฅ โ๐ฃ โ ๐ , Pr ๐ฆ ๐ฃ โ ๐ฆ ๐ฃ เท 4 , for ๐ฆ โ = ๐ ๐ป + ๐ . In particular, ๐ป may be preprocessed ๏ฎ On Solving Linear Systems in Sublinear Time 8
Algorithmic Techniques Famous Monte-Carlo method of von Neumann and Ulam: ๏ฎ Write matrix inverse by power series ๐ฝ โ ๐ โ1 = ฯ ๐ขโฅ0 ๐ ๐ข โ ๐ < 1 , then estimate it by random walks (in ๐ ) with unbiased expectation Inverting a Laplacian ๐ ๐ป = ๐๐ฝ โ ๐ต corresponds to summing walks in ๐ป ๏ฎ ๐ ฯ ๐ขโฅ0 ๐ต ๐ข ๐ as sum over all walks, estimate it by sampling For us: view ๐ ๐ฃ ๏ฑ (random walks) Need to control: number of walks and their length ๏ฎ Large powers ๐ข > ๐ข โ contribute relatively little (by condition number) ๏ฑ Estimate truncated series ( ๐ข โค ๐ข โ ) by short random walks (by Chebyshev โ s ๏ฑ inequality) On Solving Linear Systems in Sublinear Time 9
Related Work โ All Algorithmic Similar techniques were used before in related contexts but under ๏ฎ different assumptions, models and analyses: + [Doron-Le Gall- Probabilistic log-space algorithms for approximating ๐ ๐ป ๏ฑ Ta-Shma โ 17] Asks for entire matrix, uses many long random walks (independent of ๐ ) ๏ฎ Local solver for Laplacian systems with boundary conditions [Chung- ๏ฑ Simpson โ 15] Solver relies on a different power series and random walks ๏ฎ Local solver for PSD systems [Shyamkumar-Banerjee-Lofgren โ 16] ๏ฑ Polynomial time nnz ๐ 2/3 under assumptions like bounded matrix norm and ๏ฎ random ๐ฃ โ ๐ Local solver for Pagerank [Bressan-Peserico-Pretto โ 18, Borgs-Brautbar- ๏ฑ Chayes-Teng โ 14] Polynomial time O(๐ 2/3 ) and O( nd 1/2 ) for certain matrices (non-symmetric ๏ฎ but by definition are diagonally-dominant) On Solving Linear Systems in Sublinear Time 10
Lower Bound Techniques PSD lower bound: Take Laplacian of 2๐ -regular expander but with: ๏ฎ ๐ฃ high girth, ๏ฑ edges signed ยฑ1 at random, and ๏ฑ ๐( ๐) on the diagonal (PSD but not Laplacian) ๏ฑ ๐ The graph looks like a tree locally ๏ฎ Up to radius ฮ log ๐ around ๐ฃ ๏ฑ ๐ ๐ฅ = 0 Set ๐ ๐ฅ = ยฑ1 for ๐ฅ at distance ๐ , and 0 otherwise ๏ฎ ๐ ๐ฅ = ยฑ1 Signs have small bias ๐ โ ๐ โ๐ /2 ๏ฑ Recovering it requires reading ฮฉ(๐ โ2 ) entries ๏ฑ Using inversion formula, ๐ฆ ๐ฃ โ average of ๐ ๐ฅ โ s ๐ ๐ฅ = 0 ๏ฎ Condition number lower bound: Take two 3-regular expanders ๏ฎ connected by a matching of size ๐/๐ Let ๐ ๐ฅ = ยฑ1 with slight bias inside each expander ๏ฑ On Solving Linear Systems in Sublinear Time 11
Further Questions Accuracy guarantees ๏ฎ Different norms? ๏ฑ Condition number of ๐ instead of ๐ธ โ1/2 ๐๐ธ โ1/2 ? ๏ฑ Other representations (input/output models)? ๏ฎ Access the input ๐ via random sampling? ๏ฑ Sample from the output ๐ฆ ? ๏ฑ Other numerical problems? ๏ฎ Thank You! On Solving Linear Systems in Sublinear Time 12
Recommend
More recommend