com pressive sensing
play

Com pressive Sensing and Applications Volkan Cevher - PowerPoint PPT Presentation

Com pressive Sensing and Applications Volkan Cevher volkan@rice.edu Rice University Acknowledgements Rice DSP Group (Slides) Richard Baraniuk Mark Davenport, Marco Duarte, Chinmay Hegde, Jason Laska, Shri


  1. Com pressive Sensing and Applications Volkan Cevher volkan@rice.edu Rice University

  2. Acknowledgements • Rice DSP Group (Slides) – Richard Baraniuk � Mark Davenport, � Marco Duarte, � Chinmay Hegde, � Jason Laska, � Shri Sarvotham, � Mona Sheikh � Stephen Schnelle… – Mike Wakin, Justin Romberg, Petros Boufounos, Dror Baron

  3. Outline • Introduction to Compressive Sensing (CS) – motivation – basic concepts • CS Theoretical Foundation – geometry of sparse and compressible signals – coded acquisition – restricted isometry property (RIP) – signal recovery • CS in Action • Summary

  4. Sensing

  5. Digital Revolution

  6. Pressure is on Digital Sensors • Success of digital data acquisition is placing increasing pressure on signal/ image processing hardware and software to support higher resolution / denser sam pling » ADCs, cameras, imaging systems, microarrays, … large num bers of sensors » image data bases, camera arrays, distributed wireless sensor networks, … increasing num bers of m odalities » acoustic, RF, visual, IR, UV, x-ray, gamma ray, …

  7. Pressure is on Digital Sensors • Success of digital data acquisition is placing increasing pressure on signal/ image processing hardware and software to support higher resolution / denser sam pling » ADCs, cameras, imaging systems, microarrays, … x large num bers of sensors » image data bases, camera arrays, distributed wireless sensor networks, … x increasing num bers of m odalities » acoustic, RF, visual, IR, UV deluge of data » how to acquire , store , fuse , process efficiently?

  8. Digital Data Acquisition • Foundation: Shannon/ Nyquist sampling theorem “if you sample densely enough (at the Nyquist rate), you can perfectly reconstruct the original analog data” time space

  9. Sensing by Sampling • Long-established paradigm for digital data acquisition – uniformly sam ple data at Nyquist rate (2x Fourier bandwidth) sample

  10. Sensing by Sampling • Long-established paradigm for digital data acquisition – uniformly sam ple data at Nyquist rate (2x Fourier bandwidth) too sample m uch data!

  11. Sensing by Sampling • Long-established paradigm for digital data acquisition – uniformly sam ple data at Nyquist rate (2x Fourier bandwidth) – com press data transmit/ store com press sample JPEG JPEG2 0 0 0 … receive decompress

  12. Sparsity / Compressibility large pixels wavelet coefficients (blue = 0) frequency wideband large signal Gabor (TF) samples coefficients time

  13. Sample / Compress • Long-established paradigm for digital data acquisition – uniformly sam ple data at Nyquist rate – com press data transmit/ store sample com press sparse / com pressible wavelet transform receive decompress

  14. What’s Wrong with this Picture? • W hy go to all the w ork to acquire N sam ples only to discard all but K pieces of data? transmit/ store sample com press sparse / com pressible wavelet transform receive decompress

  15. What’s Wrong with this Picture? linear processing nonlinear processing linear signal model nonlinear signal model (bandlimited subspace) (union of subspaces) transmit/ store sample com press sparse / com pressible wavelet transform receive decompress

  16. Compressive Sensing • Directly acquire “ com pressed ” data • Replace samples by more general “measurements” transmit/ store com pressive sensing receive reconstruct

  17. Com pressive Sensing Theory I Geom etrical Perspective

  18. Sampling • Signal is - sparse in basis/ dictionary – WLOG assume sparse in space domain sparse signal nonzero entries

  19. Sampling • Signal is - sparse in basis/ dictionary – WLOG assume sparse in space domain • Sam ples sparse measurements signal nonzero entries

  20. Compressive Sampling • When data is sparse/ compressible, can directly acquire a condensed representation with no/ little information loss through linear dim ensionality reduction sparse measurements signal nonzero entries

  21. How Can It Work? • Projection not full rank … … and so loses inform ation in general • Ex: Infinitely many ’s map to the same

  22. How Can It Work? • Projection not full rank… columns … and so loses information in general • But we are only interested in sparse vectors

  23. How Can It Work? • Projection not full rank… columns … and so loses information in general • But we are only interested in sparse vectors • is effectively M x K

  24. How Can It Work? • Projection not full rank… columns … and so loses information in general • But we are only interested in sparse vectors • Design so that each of its M x K submatrices are full rank

  25. How Can It Work? columns • Goal: Design so that its M x2 K submatrices are full rank – difference between two K -sparse vectors is 2 K sparse in general – preserve information in K -sparse signals – Restricted I som etry Property (RIP) of order 2 K

  26. Unfortunately… columns • Goal: Design so that its M x2 K submatrices are full rank (Restricted Isometry Property – RIP) • Unfortunately, a combinatorial, NP-com plete design problem

  27. Insight from the 80’s [ Kashin, Gluskin] • Draw at random – iid Gaussian – iid Bernoulli … columns • Then has the RIP with high probability as long as – M x2 K submatrices are full rank – stable embedding for sparse signals – extends to compressible signals in balls

  28. Compressive Data Acquisition • Measurements = random linear com binations of the entries of • WHP does not distort structure of sparse signals – no information loss sparse measurements signal nonzero entries

  29. CS Signal Recovery • Goal : Recover signal from measurements • Challenge : Random projection not full rank (ill-posed inverse problem) • Solution : Exploit the sparse/ compressible geom etry of acquired signal

  30. Concise Signal Structure • Sparse signal: only K out of N coordinates nonzero sorted index

  31. Concise Signal Structure • Sparse signal: only K out of N coordinates nonzero – model: union of K -dimensional subspaces aligned w/ coordinate axes sorted index

  32. Concise Signal Structure • Sparse signal: only K out of N coordinates nonzero – model: union of K -dimensional subspaces • Com pressible signal: sorted coordinates decay rapidly to zero power-law decay sorted index

  33. Concise Signal Structure • Sparse signal: only K out of N coordinates nonzero – model: union of K -dimensional subspaces • Com pressible signal: sorted coordinates decay rapidly to zero – model: ball: power-law decay sorted index

  34. CS Signal Recovery • Random projection not full rank • Recovery problem: given find • Null space • So search in null space for the “best” according to some criterion – ex: least squares

  35. CS Signal Recovery • Recovery: given find (sparse) (ill-posed inverse problem) • fast pseudoinverse

  36. CS Signal Recovery • Recovery: given find (sparse) (ill-posed inverse problem) • fast, w rong pseudoinverse

  37. Why Doesn’t Work for signals sparse in the space/ tim e dom ain least squares, minimum solution null space of is almost never sparse translated to (random angle)

  38. CS Signal Recovery • Reconstruction/ decoding: given find (ill-posed inverse problem) • fast, wrong • num ber of nonzero entries “find sparsest in translated nullspace”

  39. CS Signal Recovery • Reconstruction/ decoding: given find (ill-posed inverse problem) • fast, wrong correct: • only M = 2 K measurements required to reconstruct number of K -sparse signal nonzero entries

  40. CS Signal Recovery • Reconstruction/ decoding: given find (ill-posed inverse problem) • fast, wrong correct: • only M = 2 K measurements required to reconstruct number of K -sparse signal nonzero entries slow : NP-complete algorithm

  41. CS Signal Recovery • Recovery: given find (sparse) (ill-posed inverse problem) • fast, wrong • correct, slow correct, efficient • m ild oversam pling [ Candes, Romberg, Tao; Donoho] linear program number of measurements required

  42. Why Works for signals sparse in the space/ tim e dom ain minimum solution = sparsest solution (with high probability) if

  43. Universality • Random measurements can be used for signals sparse in any basis

  44. Universality • Random measurements can be used for signals sparse in any basis

  45. Universality • Random measurements can be used for signals sparse in any basis sparse coefficient vector nonzero entries

  46. Compressive Sensing • Directly acquire “ com pressed ” data • Replace N samples by M random projections transmit/ store random m easurem ents … receive linear pgm

  47. Com pressive Sensing Theory I I Stable Em bedding

  48. Johnson-Lindenstrauss Lemma • JL Lemma: random projection stably embeds a cloud of Q points whp provided Q points • Proved via concentration inequality • Same techniques link JLL to RIP [ Baraniuk, Davenport, DeVore, Wakin, Constructive Approximation , 2008]

Recommend


More recommend