the interplay of analysis and algorithms
play

The interplay of analysis and algorithms (or, Computational Harmonic - PowerPoint PPT Presentation

The interplay of analysis and algorithms (or, Computational Harmonic Analysis) Anna Gilbert University of Michigan supported by DARPA-ONR, NSF, and Sloan Foundation Two themes Sparse representation Represent or approximate signal, function


  1. The interplay of analysis and algorithms (or, Computational Harmonic Analysis) Anna Gilbert University of Michigan supported by DARPA-ONR, NSF, and Sloan Foundation

  2. Two themes

  3. Sparse representation Represent or approximate signal, function by a linear combination of a few atomic elements

  4. Compressed Sensing Noisy, sparse signals can be approximately reconstructed from a small number of linear measurements

  5. Recovery = find significant entries Sparse representation = signal recovery different input models

  6. How to compute ? Analysis and algorithms are both key components

  7. SPARSE Signal space: dimension d Dictionary: finite collection of unit norm atoms D = { φ ω : ω ∈ Ω } , | Ω | = N > d Representation: linear combination of atoms � s = c λ φ λ λ ∈ Λ Find best -term representation m

  8. Applications Approximation theory Signal/Image compression Scientific computing, numerics Data mining, massive data sets Generalized decoding Modern, hyperspectral imaging systems Medical imaging

  9. SPARSE is NP-HARD SPARSE is NP-COMPLETE

  10. If dictionary is ONB, then SPARSE is easy (in polynomial time)

  11. Incoherent dictionaries (a basic result) -coherent dictionary, = smallest angle µ µ between vectors = number of terms in sparse representation m m < 1 2 µ Algorithm returns -term approx. with error m � 2 µm 2 � x − a m � ≤ 1 + (1 − 2 µm ) 2 � x − a OPT � Two-phase greedy pursuit Joint work with Tropp, Muthukrishnan, and Strauss

  12. Future for sparse approximation Hardness of approximation is related to hardness of SET COVER Approximability of SET COVER well-studied (Feige, etc.) Need insight from previous work in TCS Geometry is critical in sparse approximation Need a way to describe better geometry of dictionary and its relation to sparse approximation: VC dimension? Methods for constructing “good” redundant dictionaries (data dependent?) Watch the practitioners!

  13. Exponential General time O (2 d ) SPARSE SPARSE, geometry Polynomial Matrix time multiplication O ( d 2 ) FFT Linear time O ( d ) Logarithmic Chaining, HHS Streaming AAFFT time Pursuit wavelets, etc. O (log d )

  14. Computational Resources Time Space Randomness Communication

  15. Models: Sampling = measurements: m -sparse signal, length N = m log d length d

  16. Models: linear measurements = measurements: m -sparse signal, length N = m log d length d

  17. Models: Dictionary Orthonormal bases Fourier Wavelets Spikes Redundant dictionaries Piecewise constants Wavelet packets Chirps

  18. Results: Fourier Theorem: On signal with length , AAFFT d s builds -term Fourier representation in m r time using m poly(log d/ � ) m poly(log d/ � ) samples with error � s − r � 2 ≤ (1 + � ) � s − s m � 2 On each signal, succeed with high probability. G., Muthukrishnan, and Strauss 2005

  19. Why sublinear resources?

  20. 11223%456+4)7+6#8%46#59$4:;<=>=?@%+-'%!"#$:A<?B; 223G%456+4)7+6#8%46#59$4:CAA@%+-'%!"#$:A<?=E % C % C CAA CAA A<= A<= =A =A *+$,-$'./%0"' *+$,-$'./%0"' A<> A<> >A >A A<E A<E EA EA A<; A<; ;A ;A % A % A C ; D E F C ; D E F !"#$%&"'()& !"#$%&"'()& 11223%$++)+%"'%456+4)7+6# IH6#59$%)*%')"4/%"'5-!%4"7'69%)'%$6.J%&"'()& % CAA C > A<= E =A *+$,-$'./%0"' >A A<> ; A<E A EA ;A A<; ! ; % A ! E C ; D E F C ; D E !"#$%&"'()& !"#$ > H%CA Sparsogram

  21. Extensions, applications Generalize Fourier sampling algorithm to sublinear algorithm for linear chirps Multi-user detection for wireless comm. Radar detection and identification Calderbank, G., and Strauss 2006 Lepak, Strauss, and G.

  22. Results: Wavelets Theorem: On signal with length , streaming d s algorithm builds -term wavelet representation m r in time using poly( m log d/ � ) poly( m log d/ � ) linear measurements with error � s − r � 2 ≤ (1 + � ) � s − s m � 2 On each signal, succeed with high probability. G., Guha, Indyk, Kotidis, Muthukrishnan, and Strauss 2001

  23. Results: Chaining • Theorem: With probability at least , the random 1 − d − 3 measurement matrix has the following property. Φ Suppose that is a d -dimensional signal whose best m-term s approximation with respect to norm is . Given the s m � 1 O ( m log 2 d ) sketch of size and the number m, the v = Φ s Chaining Pursuit algorithm produces a signal with at s � most O(m) nonzero entries. This signal estimate satisfies � s − � s � 1 ≤ C log m � s − s m � 1 O ( m log 2 ( m ) log 2 ( d )) The time cost of the algorithm is G., Strauss, Tropp, and Vershynin 2006

  24. Algorithmic linear dimension reduction in � 1 Theorem: Let be a set of points in R d Y endowed with the norm. Assume that each � 1 point has at most non-zero coordinates. These m points can be linearly embedded in with � 1 O ( m log 2 d ) distortion , using only O (log 3 ( m ) log 2 ( d )) dimensions. Moreover, we can reconstruct a point from its low-dimensional sketch in time O ( m log 2 ( m ) log 2 ( d )) Vershynin 2006 G., Strauss, Tropp, and

  25. Results: HHS • Theorem: With probability at least , the random 1 − d − 3 measurement matrix has the following property. Suppose Φ that is a d -dimensional signal whose m largest entries are s given by . Given the sketch of size s m v = Φ s m polylog( d ) / � 2 and the number m, the HHS Pursuit algorithm produces a signal with m nonzero entries. This signal estimate s � satisfies � √ m � s − s m � 1 � s − � s � 2 ≤ � s − s m � 2 + m 2 polylog( d ) /� 4 The time cost of the algorithm is G., Strauss, Tropp, and Vershynin 2007

  26. Desiderata Uniformity: Sketch works for all signals simultaneously Optimal Size: measurements m polylog( d ) Optimal Speed: Update and output times are m polylog( d ) Must have high quality: answer to query has near-optimal error

  27. less information measure less compute less

  28. Related Work Reference Uniform Opt. Storage Sublin. Query GMS � � X CM � X � CRT, Don � � X Chaining � � � HHS � � � Remark: Numerous contributions in area are not strictly comparable Gilbert et al. 2001, 2005: Cormode-Muthukrishnan 2005; Candes-(Romberg)-Tao 2004, 2005; Donoho 2004, 2005....

  29. More formally....

  30. Signal Information Recovery signal space statistic space statistic map Ω U Ω U information map Φ (measurements) A information space recovery Φ Ω algorithm Golomb-Weinberger 1959

  31. More Formal Framework... What signal class are we interested in? What statistic are we trying to compute? How much nonadaptive information is necessary to do so? What type of information? Point samples? Inner products? Deterministic or random information? How much storage does the measurement operator require? How much computation time, space does the algorithm use? How much communication is necessary?

  32. Computational Harmonic Analysis? Algorithmic Harmonic Analysis = AHA!

  33. http://www.math.lsa.umich.edu/~annacg annacg@umich.edu

  34. Isolation = Approximate Group Testing

  35. Approximate group testing Want to find spikes at height , � noise � 1 = 1 m 1 /m Assign positions into groups by n = m log d d Φ of spikes isolated m ≥ c 1 m groups have ≤ c 2 m noise ≥ 1 / (2 m ) groups have single spike and low noise ≥ ( c 1 − c 2 ) m e ( − m log d ) except with probability Union bound over all spike configurations

Recommend


More recommend