polynomials
play

Polynomials Paul Valiant Brown University (Based on joint work - PowerPoint PPT Presentation

Three Perspectives on Orthogonal Polynomials Paul Valiant Brown University (Based on joint work with Gregory Valiant, mostly STOC11: Estimating the Unseen: An n/log(n)-sample Estimator for Entropy and Support Size, Shown Optimal via New


  1. Three Perspectives on Orthogonal Polynomials Paul Valiant Brown University (Based on joint work with Gregory Valiant, mostly STOC’11: “ Estimating the Unseen: An n/log(n)-sample Estimator for Entropy and Support Size, Shown Optimal via New CLTs” )

  2. Structure of this talk: 3 polynomial challenges… and solutions Chebyshev Laguerre Hermite Seems like: “cosine over “cosine times “cosine” Gaussian ” exponential ” “Seems,” madam? Nay, it is . I know not “seems.” – Hamlet

  3. Challenge 1: Poisson bumps  thinnest bumps k k+1 k+2 k-2 k-1 𝑞𝑝𝑗 𝜇, 𝑙 = 𝜇 𝑙 𝑓 −𝑙 𝑙! x ? Linear transform Motivation: Given an event with probability p, 𝑞𝑝𝑗 𝑞𝑜, 𝑙 captures the probability of it occurring exactly k times in x Poi(n) samples. Let F k be total x-resolution number of events that were 1. Thin as possible observed k times. F k captures (with bounded coeffs) probabilities from . Is  there a linear combination of 2. σ =1 F k that captures ? y-resolution Thm: general log n factor improvement in resolution, #samples

  4. Chebyshev Polynomials 𝑈 𝑘 cos 𝑦 = cos(𝑘𝑦) Chebyshev is exactly like cosine, except on distorted x-axis x-resolution 1. Thin as possible Both unchanged under (with bounded coeffs) x-axis distortion! 2. σ j =1 y-resolution New question: thin cosine bumps

  5. Thinnest Cosine Bumps Thinnest linear combination of cos(𝑘𝑦) for 𝑘 < 𝑐 : ≈ 1/𝑐 (Intuition: Fourier transform of degree b gives resolution 1/b) x-resolution 1. Thin as possible Sum of all possible x- (with bounded coeffs) translated bumps is constant 2. σ j =1 (Trig functions are well- y-resolution behaved under x-translation)

  6. Chebyshev Takeaways: (Modulo x-axis distortion) “polynomials are cosines” 𝑞𝑝𝑗 𝜇, 𝑙 = 𝜇 𝑙 𝑓 −𝑙 𝑙! Linear transform Motivation: Given an event with probability p, 𝑞𝑝𝑗 𝑞𝑜, 𝑙 captures the probability of it occurring exactly k times in Poi(n) samples. Let F k be total b times x-resolution number of events that were 1. Thin as possible thinner observed k times. F k captures (with coeffs << n) probabilities from . Is  there a linear combination of 2. σ =1 exp(b) F k that captures ? y-resolution Thm: general log n factor improvement in resolution, #samples Thus: 𝑐 = 𝜄(log 𝑜)

  7. Challenge 2: Exponentially Growing Derivatives Degree j polynomial with roots at  , 2  ; Find: and all remaining roots have much larger derivative, growing exponentially with x Roots close together have small derivatives Pulling a root farther away increases its derivative, but only polynomially Success requires a delicate balancing act!

  8. Orthogonal to Polynomials Motivation: Want to construct a pair of distributions g + ,g - that are, respectively, close to the uniform distributions on T and 2T elements, but where for each (small) k, the expected number of domain elements Fact: If P is a degree j polynomial with distinct real roots {x i }, seen k times from Poi(n) then the signed measure h P having point mass 1/𝑄′(𝑦 𝑗 ) at samples is identical for g + ,g - . each root x i is orthogonal to all polynomials of degree ≤j -2 Essentially: find a signed measure g(x) that is Essentially: find a signed measure ℎ(𝑦) that is 1) Orthogonal to 𝑞𝑝𝑗 𝑦, 𝑙 = 𝑦 𝑙 𝑓 −𝑙 1) Orthogonal to all degree ≤k polynomials for each small k, 2) Has most of its positive mass at 1/T and 𝑙! 𝑕 𝑦 ≜ 𝑓 𝑦 ℎ 𝑦 2) Has most of its positive most of its negative mass at 1/(2T) mass at 1/T and most of its and otherwise decays ≪ 𝑓 −𝑦 • negative mass at 1/(2T) Task: find P such that 𝑄′(𝑦 𝑗 ) grows exponentially in x i

  9. Laguerre Polynomials Defined by 𝑀 𝑜 𝑦 = 𝑓 𝑦 𝑒 𝑜 𝑓 −𝑦 𝑦 𝑜 and 𝑒𝑦 𝑜 𝑜! ∞ 𝑀 𝑜 𝑦 𝑀 𝑛 𝑦 𝑓 −𝑦 𝑒𝑦 = [𝑛 = 𝑜] orthogonal as: ׬ 0 Why should the derivative be so nicely behaved at its roots, in particular, growing exponentially? Transform the Laguerre: 𝑤 = 𝑓 −𝑦 2 /2 𝑦 ⋅ 𝑀 𝑜 (𝑦 2 ) Many differential equations, including 𝑤 ′′ + 4𝑜 + 2 − 𝑦 2 + 1 4𝑦 2 𝑤 = 0 Almost harmonic motion, v → sine Nicely spaced zeros, and max derivative at the zeros

  10. The Construction Recall: We want a signed measure g on the positive reals that: • Is orthogonal to low degree polynomials • Decays exponentially fast • Its positive portion has most of its mass at 2𝜗 • Its negative portion has most of its mass at 𝜗 p + p - Theorem: p + is “close” to U n/2 , and p - is “close” to U n , and p + and p - are indistinguishable via cn/log n samples (Modulo diff-eq distortion) “polynomials are 𝑓 𝑦 sin(𝑦) ”

  11. Challenge 3: exponentially good bump approximations Find a linear combination over j of poi(x,j) that approximates poi(x,k) 2 to Motivation: Previously, constructed lowerbound distributions g + ,g - where within  , using coefficients ≤1/  expectation of every measurement matched. Lower bound? No… until Think of  =1/exp(j) we show variances match too. Aim: show that variances can be approximated as linear combinations of expectations, with moderate coefficients; thus matching means implies matching variances. Since means come from poi(j,x), second moments come from poi(j,x) 2 . These look like Gaussians! 1) What’s the answer for Gaussians? 2) Analyze via Hermite polynomials instead

  12. Approximating “Thin” Gaussians as Linear Combinations of Gaussians What do we convolve a Gaussian with to approximate a thinner Gaussian? (Other direction is easy, since convolving Gaussians adds their variances) “Blurring is easy, unblurring is hard”  can only do it approximately How to analyze? Fourier transform! Convolution becomes multiplication Now: what do we multiply a Gaussian with to approximate a fatter Gaussian? 𝑓 −𝑦 2 ⋅ ? ? ? = 𝑓 −𝑦 2 /2 𝑓 𝑦 2 /2 Problem: blows up Answer: if we want to approximate to within 𝜗 , we only need to approximate out to where 𝑓 −𝑦 2 /2 = 𝜗 . How big is 𝑓 𝑦 2 /2 here? 1/𝜗 Result: Can approximate to within 𝜗 using coefficients no bigger than 1/𝜗

  13. Hermite Polynomials Poissons seem a lot like Gaussians (I was stuck here for about a month) = 𝑧 2𝑘 𝑓 −𝑧 2 𝑦 𝑘 𝑓 −𝑦 𝑒 2𝑘 1 𝑒𝑥 2𝑘 𝑓 −𝑥 2 = 1 𝑔𝑝𝑣𝑠𝑗𝑓𝑠 Idea: x  y 2 𝑘! 𝐼 2𝑘 𝑥 𝑓 −𝑥 2 𝑘! 𝑘! 𝑘! Hermite Polynomials! Sequence of orthogonal polynomials, from which we take the even ones: orthogonal basis for even functions To express any function in this basis, just compute each coefficient as an inner product Which function? Fourier transform of “thin” Poisson, cut off at 𝜗 Proposition: Can approximate Pr[𝑄𝑝𝑗(2𝜇) = 𝑙] to within 𝜗 as a linear combination σ 𝑘 𝛽 𝑙,𝑘 Pr[𝑄𝑝𝑗 𝜇 = 𝑘] with coefficients that 4 𝑙, 24 log 3 1 2 1 sum to σ 𝑘 |𝛽 𝑙,𝑘 | ≤ 𝜗 200max{ 𝜗 }

  14. Structure of this talk: 3 polynomial challenges… and solutions Chebyshev Laguerre Hermite Seems like: “cosine over “cosine times “cosine” Gaussian ” exponential ” “Seems,” madam? Nay, it is . I know not “seems.” – Hamlet

Recommend


More recommend