what we know and what we do not know about practical
play

What we know and what we do not know about practical compressive - PowerPoint PPT Presentation

What we know and what we do not know about practical compressive sampling Deanna Needell Jan. 13, 2014 FGMIA 2014, Paris, France Outline Introduction Mathematical Formulation & Methods Practical CS Other notions of sparsity


  1. What we know and what we do not know about practical compressive sampling Deanna Needell Jan. 13, 2014 FGMIA 2014, Paris, France

  2. Outline ✧ Introduction ✧ Mathematical Formulation & Methods ✧ Practical CS ✧ Other notions of sparsity ✧ Heavy quantization ✧ Adaptive sampling

  3. The mathematical problem 1. Signal of interest f ∈ C d ( = C N × N ) 2. Measurement operator A : C d → C m ( m ≪ d ) 3. Measurements y = A f + ξ               f          ξ  y A    =   +        4. Problem: Reconstruct signal f from measurements y

  4. Sparsity Measurements y = A f + ξ .               f          y A  ξ  =   +          Assume f is sparse : def ✧ In the coordinate basis: � f � 0 = | supp( f ) | ≤ s ≪ d ✧ In orthonormal basis: f = Bx where � x � 0 ≤ s ≪ d In practice, we encounter compressible signals. ✦ f s is the best s -sparse approximation to f

  5. Many applications... ✧ Radar, Error Correction ✧ Computational Biology, Geophysical Data Analysis ✧ Data Mining, classification ✧ Neuroscience ✧ Imaging ✧ Sparse channel estimation, sparse initial state estimation ✧ Topology identification of interconnected systems ✧ ...

  6. Sparsity... Sparsity in coordinate basis: f=x

  7. Reconstructing the signal f from measurements y ✦ ℓ 1 -minimization [Candès-Romberg-Tao] Let A satisfy the Restricted Isometry Property and set: ˆ f = argmin � g � 1 such that � A f − y � 2 ≤ ε , g where � ξ � 2 ≤ ε . Then we can stably recover the signal f : f � 2 � ε + � x − x s � 1 � f − ˆ . � s This error bound is optimal.

  8. Restricted Isometry Property ✧ A satisfies the Restricted Isometry Property (RIP) when there is δ < c such that (1 − δ ) � f � 2 ≤ � A f � 2 ≤ (1 + δ ) � f � 2 whenever � f � 0 ≤ s . ✧ m × d Gaussian or Bernoulli measurement matrices satisfy the RIP with high probability when m � s log d . ✧ Random Fourier and others with fast multiply have similar property: m � s log 4 d .

  9. Sparsity... In orthonormal basis: f = Bx

  10. Natural Images Images are compressible in Wavelet bases . � N � � f = x j , k H j , k , x j , k = f , H j , k , � f � 2 = � x � 2 , j , k = 1 Figure 1: Haar basis functions Wavelet transform is orthonormal and multi-scale. Sparsity level of image is higher on detail coefficients.

  11. Sparsity in orthonormal basis B ✦ L1-minimization Method For orthonormal basis B , f = Bx with x sparse, one may solve the ℓ 1 -minimization program: � B − 1 ˜ ˆ � A ˜ f = argmin f � 1 subject to f − y � 2 ≤ ε . ˜ f ∈ C n Same results hold.

  12. Sparsity... In arbitrary dictionary: f = Dx

  13. The CS Process

  14. Example: Oversampled DFT 1 � n e − 2 π ikt / n ✧ n × n DFT: d k ( t ) = ✧ Sparse in the DFT → superpositions of sinusoids with frequencies in the lattice. ✧ Instead, use the oversampled DFT : ✧ Then D is an overcomplete frame with highly coherent columns → conventional CS does not apply .

  15. Example: Gabor frames ✧ Gabor frame: G k ( t ) = g ( t − k 2 a ) e 2 π ik 1 bt ✧ Radar, sonar, and imaging system applications use Gabor frames and wish to recover signals in this basis. ✧ Then D is an overcomplete frame with possibly highly coherent columns → conventional CS does not apply .

  16. Example: Curvelet frames ✧ A Curvelet frame has some properties of an ONB but is overcomplete. ✧ Curvelets approximate well the curved singularities in images and are thus used widely in image processing. ✧ Again, this means D is an overcomplete dictionary → conventional CS does not apply .

  17. Example: UWT ✧ The undecimated wavelet transform has a translation invariance property that is missing in the DWT. ✧ The UWT is overcomplete and this redundancy has been found to be helpful in image processing. ✧ Again, this means D is a redundant dictionary → conventional CS does not apply .

  18. ℓ 1 -Synthesis Method ✦ For arbitrary tight frame D , one may solve the ℓ 1 -synthesis program: � � ˆ f = D argmin � ˜ x � 1 subject to � A D ˜ x − y � 2 ≤ ε . x ∈ C n ˜ Some work on this method [Candès et.al., Rauhut et.al., Elad et.al.,...] ✦ To do: Understand the ℓ 1 -synthesis problem, necessary assumptions, recovery guarantees.

  19. ℓ 1 -Analysis Method ✦ For arbitrary tight frame D , one may solve the ℓ 1 -analysis program: � D ∗ ˜ ˆ � A ˜ subject to f − y � 2 ≤ ε . f = argmin f � 1 ˜ f ∈ C n

  20. Condition on A? ✦ D-RIP We say that the measurement matrix A obeys the restricted isometry property adapted to D (D-RIP) if there is δ < c such that (1 − δ ) � Dx � 2 2 ≤ � A Dx � 2 2 ≤ (1 + δ ) � Dx � 2 2 holds for all s -sparse x . ✦ Similarly to the RIP , many classes of m × d random matrices satisfy the D-RIP with m ≈ s log( d / s ) .

  21. CS with tight frame dictionaries ✦ Theorem [Candès-Eldar-N-Randall] Let D be an arbitrary tight frame and let A be a measurement matrix . Then the solution ˆ satisfying D-RIP f to ℓ 1 -analysis satisfies f − f � 2 � ε + � D ∗ f − ( D ∗ f ) s � 1 � ˆ . � s ✦ In other words, This result says that ℓ 1 -analysis is very accurate when D ∗ f has rapidly decaying coefficients and D is a tight frame.

  22. ℓ 1 -analysis: Experimental Setup n = 8192, m = 400, d = 491,520 A: m × n Gaussian, D: n × d Gabor

  23. ℓ 1 -analysis: Experimental Setup n = 8192, m = 400, d = 491,520 A: m × n Gaussian, D: n × d Gabor

  24. ℓ 1 -analysis: Experimental Results

  25. Other algorithms ✦ ℓ 1 -analysis is very accurate when D ∗ f has rapidly decaying coefficients and D is a tight frame. This is precisely because this method operates in “analysis” space. ✦ To do: analysis methods for non-tight frames, without decaying analysis coefficients (concatenations), other models ✦ What about operating in signal or coefficient space?

  26. Is it really a pipe? (Thanks to M. Davenport for this clever analogy.)

  27. CoSaMP C O S A MP (N-Tropp) input: Sampling operator A , measurements y , sparsity level s initialize: Set x 0 = 0 , i = 0 . repeat signal proxy: Set p = A ∗ ( y − Ax i ) , Ω = supp( p 2 s ) , T = Ω ∪ supp( x i ) . signal estimation: Using least-squares, set b | T = A † T y and b | T c = 0 . prune and update: Increment i and to obtain the next approximation, set x i = b s . x = x i output: s -sparse reconstructed vector �

  28. Signal Space CoSaMP S IGNAL S PACE C O S A MP (Davenport-N-Wakin) input: A , D , y , s , stopping criterion initialize: r = y , x 0 = 0 , ℓ = 0 , Γ = � repeat h = A ∗ r proxy: identify: Ω = S D ( h ,2 s ) merge: T = Ω ∪ Γ x = argmin z � y − Az � 2 z ∈ R ( D T ) update: � s.t. x , s ) Γ = S D ( � x ℓ + 1 = P Γ � x r = y − Ax ℓ + 1 ℓ = ℓ + 1 x = x ℓ output: �

  29. Signal Space CoSaMP ✦ Here we must contend with P Λ : C n → R ( D Λ ). Λ opt ( z , s ) : = argmin � z − P Λ z � 2 , Λ : | Λ |= s ✦ Estimate by S D ( z , s ) with | S D ( z , s ) | = s , that satisfies � � � � � � � � � P Λ opt ( z , s ) z − P S D ( z , s ) z � � P Λ opt ( z , s ) z � � z − P Λ opt ( z , s ) z � 2 ≤ min ǫ 1 2 , ǫ 2 2 for some constants ǫ 1 , ǫ 2 ≥ 0 .

  30. Approximate Projection ✦ Practical choices for S D ( z , s ) : ✧ Any sparse recovery algorithm! ✧ OMP ✧ CoSaMP ✧ ℓ 1 -minimization followed by hard thresholding

  31. Signal Space CoSaMP ✦ Theorem [Davenport-N-Wakin] Let D be an arbitrary tight frame, A be a measurement matrix satisfying D-RIP , and f a sparse signal with respect to D . Then the solution ˆ f from Signal Space CoSaMP satisfies � ˆ f − f � 2 � ε . (And similar results for approximate sparsity, depending on the approximation method used for Λ opt ( z , s ) .) ✦ To do: Design approximation methods that satisfy necessary recovery bounds (sparse approximation).

  32. Signal Space CoSaMP: Experimental Results Figure 2: Performance in recovering signals having a s = 8 sparse representation in a dictionary D with orthogonal, but not normalized, columns.

  33. Signal Space CoSaMP: Experimental Results (a) (b) Figure 3: Results with s = 8 sparse representation in a 4 × overcomplete DFT dictionary: (a) well-separated coefficients, (b) clustered coefficients.

  34. Signal Space CoSaMP: Recent improvements ✦ Recently improved results [Giryes-N and Hegde-Indyk-Schmidt] which relax the assumptions on the approximate projections. ✦ These results show that at least for RIP/incoherent dictionaries, standard algorithms like CoSaMP/OMP/IHT suffice for the approximate projections. To do: ✦ The interesting/challenging case is when the dictionary does not satisfy such a condition. Are there methods which provide these approximate projections? Or are they not even necessary?

  35. Natural images Sparse... 256 × 256 “Boats" image

  36. Natural images Sparse wavelet representation...

  37. Natural images Images are compressible in discrete gradient .

Recommend


More recommend