orthogonal polynomials and spectral algorithms
play

Orthogonal Polynomials and Spectral Algorithms Nisheeth K. Vishnoi - PowerPoint PPT Presentation

Orthogonal Polynomials and Spectral Algorithms Nisheeth K. Vishnoi 1.0 d=0 d=1 0.5 d=5 1.0 0.5 0.5 1.0 d=4 d=3 0.5 d=2 1.0 FOCS, Oct. 8, 2016 Orthogonal Polynomials -orthogonality Polynomials p ( x ) , q ( x ) are


  1. Orthogonal Polynomials and Spectral Algorithms Nisheeth K. Vishnoi 1.0 d=0 d=1 0.5 d=5 � 1.0 � 0.5 0.5 1.0 d=4 d=3 � 0.5 d=2 � 1.0 FOCS, Oct. 8, 2016

  2. Orthogonal Polynomials µ -orthogonality Polynomials p ( x ) , q ( x ) are µ -orthogonal w.r.t. µ : I → R ≥ 0 if � � p , q � µ := p ( x ) q ( x ) d µ ( x ) = 0 x ∈I µ -orthogonal family Start with 1 , x , x 2 , . . . , x d , . . . and apply Gram-Schmidt orthogonalization w.r.t. �· , ·� µ to obtain a µ -orthogonal family p 0 ( x ) = 1 , p 1 ( x ) , p 2 ( x ) , . . . , p d ( x ) , . . . Examples Legendre: I = [ − 1 , 1] and µ ( x ) = 1. Hermite: I = R and µ ( x ) = e − x 2 / 2 . Laguerre: I = R ≥ 0 and µ ( x ) = e − x . 1 Chebyshev (Type 1): I = [ − 1 , 1] and µ ( x ) = 1 − x 2 . √

  3. Orthogonal Polynomials µ -orthogonality Polynomials p ( x ) , q ( x ) are µ -orthogonal w.r.t. µ : I → R ≥ 0 if � � p , q � µ := p ( x ) q ( x ) d µ ( x ) = 0 x ∈I µ -orthogonal family Start with 1 , x , x 2 , . . . , x d , . . . and apply Gram-Schmidt orthogonalization w.r.t. �· , ·� µ to obtain a µ -orthogonal family p 0 ( x ) = 1 , p 1 ( x ) , p 2 ( x ) , . . . , p d ( x ) , . . . Examples Legendre: I = [ − 1 , 1] and µ ( x ) = 1. Hermite: I = R and µ ( x ) = e − x 2 / 2 . Laguerre: I = R ≥ 0 and µ ( x ) = e − x . 1 Chebyshev (Type 1): I = [ − 1 , 1] and µ ( x ) = 1 − x 2 . √

  4. Orthogonal Polynomials µ -orthogonality Polynomials p ( x ) , q ( x ) are µ -orthogonal w.r.t. µ : I → R ≥ 0 if � � p , q � µ := p ( x ) q ( x ) d µ ( x ) = 0 x ∈I µ -orthogonal family Start with 1 , x , x 2 , . . . , x d , . . . and apply Gram-Schmidt orthogonalization w.r.t. �· , ·� µ to obtain a µ -orthogonal family p 0 ( x ) = 1 , p 1 ( x ) , p 2 ( x ) , . . . , p d ( x ) , . . . Examples Legendre: I = [ − 1 , 1] and µ ( x ) = 1. Hermite: I = R and µ ( x ) = e − x 2 / 2 . Laguerre: I = R ≥ 0 and µ ( x ) = e − x . 1 Chebyshev (Type 1): I = [ − 1 , 1] and µ ( x ) = 1 − x 2 . √

  5. Orthgonal polynomials have many amazing properties Monic µ -orthogonal polynomials satisfy 3-term recurrences p d +1 ( x ) = ( x − α d +1 ) p d + β d p d − 1 for d ≥ 0 with p − 1 = 0.

  6. Orthgonal polynomials have many amazing properties Monic µ -orthogonal polynomials satisfy 3-term recurrences p d +1 ( x ) = ( x − α d +1 ) p d + β d p d − 1 for d ≥ 0 with p − 1 = 0. Proof sketch degree d � �� � p d +1 − xp d = α d +1 p d + β d p d − 1 + � i < d − 1 γ i p i 1

  7. Orthgonal polynomials have many amazing properties Monic µ -orthogonal polynomials satisfy 3-term recurrences p d +1 ( x ) = ( x − α d +1 ) p d + β d p d − 1 for d ≥ 0 with p − 1 = 0. Proof sketch degree d � �� � p d +1 − xp d = α d +1 p d + β d p d − 1 + � i < d − 1 γ i p i 1 2 For i < d − 1 , � xp d , p i � µ = � p d +1 − xp d , p i � µ = γ i � p i , p i � µ but

  8. Orthgonal polynomials have many amazing properties Monic µ -orthogonal polynomials satisfy 3-term recurrences p d +1 ( x ) = ( x − α d +1 ) p d + β d p d − 1 for d ≥ 0 with p − 1 = 0. Proof sketch degree d � �� � p d +1 − xp d = α d +1 p d + β d p d − 1 + � i < d − 1 γ i p i 1 2 For i < d − 1 , � xp d , p i � µ = � p d +1 − xp d , p i � µ = γ i � p i , p i � µ but 3 � xp d , p i � µ = � p d , xp i � µ = 0 as deg( xp i ) < d implying γ i = 0.

  9. Orthgonal polynomials have many amazing properties Monic µ -orthogonal polynomials satisfy 3-term recurrences p d +1 ( x ) = ( x − α d +1 ) p d + β d p d − 1 for d ≥ 0 with p − 1 = 0. Proof sketch degree d � �� � p d +1 − xp d = α d +1 p d + β d p d − 1 + � i < d − 1 γ i p i 1 2 For i < d − 1 , � xp d , p i � µ = � p d +1 − xp d , p i � µ = γ i � p i , p i � µ but 3 � xp d , p i � µ = � p d , xp i � µ = 0 as deg( xp i ) < d implying γ i = 0. Roots (corollaries) If p 0 , p 1 , . . . , p d , . . . are orthogonal w.r.t. µ : [ a , b ] → R ≥ 0 then for each p d , roots are distinct, real and lie in [ a , b ]. Roots of p d and p d +1 also interlace!

  10. Many and growing applications in TCS ... Hermite: I = R and µ ( x ) = e − x 2 / 2 Invariance principles, hardness of approximation a la Mossel, O’Donnell, Oleszkiewicz, ... Laguerre: I = R ≥ 0 and µ ( x ) = e − x Constructing sparsifiers a la Batson, Marcus, Spielman, Srivastava, ... √ 1 − x 2 Chebyshev (Type 2): I = [ − 1 , 1] and µ ( x ) = Nonbacktracking random walks and Ramanujan graphs a la Alon, Boppana, Friedman, Lubotzky, Philips, Sarnak, ... 1 Chebyshev (Type 1): I = [ − 1 , 1] and µ ( x ) = √ 1 − x 2 Spectral algorithms – This talk Extensions to multivariate and matrix polynomials Several examples in this workshop ..

  11. Many and growing applications in TCS ... Hermite: I = R and µ ( x ) = e − x 2 / 2 Invariance principles, hardness of approximation a la Mossel, O’Donnell, Oleszkiewicz, ... Laguerre: I = R ≥ 0 and µ ( x ) = e − x Constructing sparsifiers a la Batson, Marcus, Spielman, Srivastava, ... √ 1 − x 2 Chebyshev (Type 2): I = [ − 1 , 1] and µ ( x ) = Nonbacktracking random walks and Ramanujan graphs a la Alon, Boppana, Friedman, Lubotzky, Philips, Sarnak, ... 1 Chebyshev (Type 1): I = [ − 1 , 1] and µ ( x ) = √ 1 − x 2 Spectral algorithms – This talk Extensions to multivariate and matrix polynomials Several examples in this workshop ..

  12. Many and growing applications in TCS ... Hermite: I = R and µ ( x ) = e − x 2 / 2 Invariance principles, hardness of approximation a la Mossel, O’Donnell, Oleszkiewicz, ... Laguerre: I = R ≥ 0 and µ ( x ) = e − x Constructing sparsifiers a la Batson, Marcus, Spielman, Srivastava, ... √ 1 − x 2 Chebyshev (Type 2): I = [ − 1 , 1] and µ ( x ) = Nonbacktracking random walks and Ramanujan graphs a la Alon, Boppana, Friedman, Lubotzky, Philips, Sarnak, ... 1 Chebyshev (Type 1): I = [ − 1 , 1] and µ ( x ) = √ 1 − x 2 Spectral algorithms – This talk Extensions to multivariate and matrix polynomials Several examples in this workshop ..

  13. Many and growing applications in TCS ... Hermite: I = R and µ ( x ) = e − x 2 / 2 Invariance principles, hardness of approximation a la Mossel, O’Donnell, Oleszkiewicz, ... Laguerre: I = R ≥ 0 and µ ( x ) = e − x Constructing sparsifiers a la Batson, Marcus, Spielman, Srivastava, ... √ 1 − x 2 Chebyshev (Type 2): I = [ − 1 , 1] and µ ( x ) = Nonbacktracking random walks and Ramanujan graphs a la Alon, Boppana, Friedman, Lubotzky, Philips, Sarnak, ... 1 Chebyshev (Type 1): I = [ − 1 , 1] and µ ( x ) = √ 1 − x 2 Spectral algorithms – This talk Extensions to multivariate and matrix polynomials Several examples in this workshop ..

  14. Many and growing applications in TCS ... Hermite: I = R and µ ( x ) = e − x 2 / 2 Invariance principles, hardness of approximation a la Mossel, O’Donnell, Oleszkiewicz, ... Laguerre: I = R ≥ 0 and µ ( x ) = e − x Constructing sparsifiers a la Batson, Marcus, Spielman, Srivastava, ... √ 1 − x 2 Chebyshev (Type 2): I = [ − 1 , 1] and µ ( x ) = Nonbacktracking random walks and Ramanujan graphs a la Alon, Boppana, Friedman, Lubotzky, Philips, Sarnak, ... 1 Chebyshev (Type 1): I = [ − 1 , 1] and µ ( x ) = √ 1 − x 2 Spectral algorithms – This talk Extensions to multivariate and matrix polynomials Several examples in this workshop ..

  15. The goal of today’s talk Many spectral algorithms today rely on ability to quickly compute good approximations to matrix-function-vector products: e.g., A s v , A − 1 v , exp( − A ) v , ... or top few eigenvalues and eigenvectors. Demonstrate How to reduce the problem of computing these primitives to a small number of computations of the form Bu where B is a matrix closely related to A (often A itself) and u is some vector. A key feature: If Av can be computed quickly (e.g., if A is sparse) then Bu can also be computed quickly. Approximation theory provides the right framework to study these questions – Borrows heavily from orthogonal polynomials!

  16. The goal of today’s talk Many spectral algorithms today rely on ability to quickly compute good approximations to matrix-function-vector products: e.g., A s v , A − 1 v , exp( − A ) v , ... or top few eigenvalues and eigenvectors. Demonstrate How to reduce the problem of computing these primitives to a small number of computations of the form Bu where B is a matrix closely related to A (often A itself) and u is some vector. A key feature: If Av can be computed quickly (e.g., if A is sparse) then Bu can also be computed quickly. Approximation theory provides the right framework to study these questions – Borrows heavily from orthogonal polynomials!

  17. The goal of today’s talk Many spectral algorithms today rely on ability to quickly compute good approximations to matrix-function-vector products: e.g., A s v , A − 1 v , exp( − A ) v , ... or top few eigenvalues and eigenvectors. Demonstrate How to reduce the problem of computing these primitives to a small number of computations of the form Bu where B is a matrix closely related to A (often A itself) and u is some vector. A key feature: If Av can be computed quickly (e.g., if A is sparse) then Bu can also be computed quickly. Approximation theory provides the right framework to study these questions – Borrows heavily from orthogonal polynomials!

Recommend


More recommend