adaptive sparse recovery
play

Adaptive Sparse Recovery Eric Price MIT 2012-04-26 Joint work - PowerPoint PPT Presentation

Adaptive Sparse Recovery Eric Price MIT 2012-04-26 Joint work with Piotr Indyk and David Woodruff, 2011-2012 Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 1 / 29 Outline Motivating Example 1 Eric Price (MIT) Adaptive Sparse


  1. General Compressive Sensing Want to observe n -dimensional vector x ◮ Which of n people have a genetic mutation. ◮ Image ◮ Traffic pattern of packets on a network. Normally takes n observations to find. But we know some structure on the input: ◮ Genetics: most people don’t have the mutation. ◮ Images: mostly smooth with some edges. ◮ Traffic: Zipf distribution. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 8 / 29

  2. General Compressive Sensing Want to observe n -dimensional vector x ◮ Which of n people have a genetic mutation. ◮ Image ◮ Traffic pattern of packets on a network. Normally takes n observations to find. But we know some structure on the input: ◮ Genetics: most people don’t have the mutation. ◮ Images: mostly smooth with some edges. ◮ Traffic: Zipf distribution. We use this structure to compress space (e.g. JPEG). Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 8 / 29

  3. General Compressive Sensing Want to observe n -dimensional vector x ◮ Which of n people have a genetic mutation. ◮ Image ◮ Traffic pattern of packets on a network. Normally takes n observations to find. But we know some structure on the input: ◮ Genetics: most people don’t have the mutation. ◮ Images: mostly smooth with some edges. ◮ Traffic: Zipf distribution. We use this structure to compress space (e.g. JPEG). Can we use structure to save on observations ? Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 8 / 29

  4. Cameras 5 megapixel camera takes 15 million byte-size observations. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  5. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  6. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Why do we need to bother with so many observations? [Donoho,Cand` es-Tao] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  7. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Why do we need to bother with so many observations? [Donoho,Cand` es-Tao] Cheap in visible light (silicon), very expensive in infrared. ◮ $30,000 for 256x256 IR sensor. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  8. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Why do we need to bother with so many observations? [Donoho,Cand` es-Tao] Cheap in visible light (silicon), very expensive in infrared. ◮ $30,000 for 256x256 IR sensor. Use structure to take only a few million observations. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  9. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Why do we need to bother with so many observations? [Donoho,Cand` es-Tao] Cheap in visible light (silicon), very expensive in infrared. ◮ $30,000 for 256x256 IR sensor. Use structure to take only a few million observations. ◮ What structure? Sparsity. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  10. Sparsity A vector is k -sparse if k components are non-zero. Images are almost sparse in the wavelet basis: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 10 / 29

  11. Sparsity A vector is k -sparse if k components are non-zero. Images are almost sparse in the wavelet basis: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 10 / 29

  12. Sparsity A vector is k -sparse if k components are non-zero. Images are almost sparse in the wavelet basis: Same kind of structure as in genetic testing! Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 10 / 29

  13. Linear sketching/Compressive sensing Suppose an n -dimensional vector x is k -sparse in known basis. Given Ax , a set of m << n linear products. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 11 / 29

  14. Linear sketching/Compressive sensing Suppose an n -dimensional vector x is k -sparse in known basis. Given Ax , a set of m << n linear products. Why linear? Many applications: ◮ Genetic testing: mixing blood samples. ◮ Streaming updates: A ( x + ∆ ) = Ax + A ∆ . ◮ Camera optics: filter in front of lens. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 11 / 29

  15. Linear sketching/Compressive sensing Suppose an n -dimensional vector x is k -sparse in known basis. Given Ax , a set of m << n linear products. Why linear? Many applications: ◮ Genetic testing: mixing blood samples. ◮ Streaming updates: A ( x + ∆ ) = Ax + A ∆ . ◮ Camera optics: filter in front of lens. Then it is possible to recover x from Ax . ◮ Quickly ◮ Robustly: get close to x if x is close to k -sparse. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 11 / 29

  16. Linear sketching/Compressive sensing Suppose an n -dimensional vector x is k -sparse in known basis. Given Ax , a set of m << n linear products. Why linear? Many applications: ◮ Genetic testing: mixing blood samples. ◮ Streaming updates: A ( x + ∆ ) = Ax + A ∆ . ◮ Camera optics: filter in front of lens. Then it is possible to recover x from Ax . ◮ Quickly ◮ Robustly: get close to x if x is close to k -sparse. Note: impossible without using sparsity ( A is underdetermined). Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 11 / 29

  17. Standard Sparse Recovery Framework Specify distribution on m × n matrices A (independent of x ). Given linear sketch Ax , recover ˆ x . Satisfying the recovery guarantee: � ˆ x − x � 2 � ( 1 + ǫ ) min � x − x k � 2 k -sparse x k with probability 2 / 3. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 12 / 29

  18. Standard Sparse Recovery Framework Specify distribution on m × n matrices A (independent of x ). Given linear sketch Ax , recover ˆ x . Satisfying the recovery guarantee: � ˆ x − x � 2 � ( 1 + ǫ ) min � x − x k � 2 k -sparse x k with probability 2 / 3. Solvable with O ( 1 ǫ k log n k ) measurements [Cand` es-Romberg-Tao ’06, Gilbert-Li-Porat-Strauss ’10] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 12 / 29

  19. Standard Sparse Recovery Framework Specify distribution on m × n matrices A (independent of x ). Given linear sketch Ax , recover ˆ x . Satisfying the recovery guarantee: � ˆ x − x � 2 � ( 1 + ǫ ) min � x − x k � 2 k -sparse x k with probability 2 / 3. Solvable with O ( 1 ǫ k log n k ) measurements [Cand` es-Romberg-Tao ’06, Gilbert-Li-Porat-Strauss ’10] Matching lower bound. [Do Ba-Indyk-P-Woodruff ’10, P-Woodruff ’11] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 12 / 29

  20. Adaptive Sparse Recovery Framework For i = 1 . . . r : ◮ Choose matrix A i based on previous observations (possibly randomized). ◮ Observe A i x . ◮ Number of measurements m is total number of rows in all A i . ◮ Number of rounds is r . Given linear sketch Ax , recover ˆ x . Satisfying the recovery guarantee: � ˆ x − x � 2 � ( 1 + ǫ ) min � x − x k � 2 k -sparse x k with probability 2 / 3. Solvable with O ( 1 ǫ k log n k ) measurements [Cand` es-Romberg-Tao ’06, Gilbert-Li-Porat-Strauss ’10] Matching lower bound. [Do Ba-Indyk-P-Woodruff ’10, P-Woodruff ’11] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 12 / 29

  21. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  22. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  23. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) This talk: O ( 1 ǫ k log log n k ) . [Indyk-P-Woodruff ’11] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  24. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) This talk: O ( 1 ǫ k log log n k ) . [Indyk-P-Woodruff ’11] ◮ Using r = O ( log log n log ∗ k ) rounds. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  25. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) This talk: O ( 1 ǫ k log log n k ) . [Indyk-P-Woodruff ’11] ◮ Using r = O ( log log n log ∗ k ) rounds. Even when r = 2, can get O ( k log n + 1 ǫ k log ( k /ǫ )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  26. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) This talk: O ( 1 ǫ k log log n k ) . [Indyk-P-Woodruff ’11] ◮ Using r = O ( log log n log ∗ k ) rounds. Even when r = 2, can get O ( k log n + 1 ǫ k log ( k /ǫ )) ◮ Separating dependence on n and ǫ . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  27. Applications of Adaptivity When does adaptivity make sense? Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  28. Applications of Adaptivity When does adaptivity make sense? Genetic testing: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  29. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  30. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  31. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  32. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  33. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Streaming algorithms: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  34. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Streaming algorithms: ◮ Adaptivity corresponds to multiple passes. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  35. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Streaming algorithms: ◮ Adaptivity corresponds to multiple passes. ◮ Router finding most common connections: No . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  36. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Streaming algorithms: ◮ Adaptivity corresponds to multiple passes. ◮ Router finding most common connections: No . ◮ Mapreduce finding most frequent URLs: Yes . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  37. Outline Motivating Example 1 Formal Introduction to Sparse Recovery/Compressive Sensing 2 Algorithm 3 Conclusion 4 Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 15 / 29

  38. Outline of Algorithm Theorem Adaptive 1 + ǫ -approximate k-sparse recovery is possible with O ( 1 ǫ k log log ( n / k )) measurements. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 16 / 29

  39. Outline of Algorithm Theorem Adaptive 1 + ǫ -approximate k-sparse recovery is possible with O ( 1 ǫ k log log ( n / k )) measurements. Lemma Adaptive O ( 1 ) -approximate 1 -sparse recovery is possible with O ( log log n ) measurements. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 16 / 29

  40. Outline of Algorithm Theorem Adaptive 1 + ǫ -approximate k-sparse recovery is possible with O ( 1 ǫ k log log ( n / k )) measurements. Lemma Adaptive O ( 1 ) -approximate 1 -sparse recovery is possible with O ( log log n ) measurements. Lemma implies theorem using standard tricks ([GLPS10]). Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 16 / 29

  41. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  42. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Non-adaptive lower bound: why is this hard? Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  43. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Non-adaptive lower bound: why is this hard? Hard case: x is random e i plus Gaussian noise w . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  44. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Non-adaptive lower bound: why is this hard? Hard case: x is random e i plus Gaussian noise w . Noise � w � 2 2 = Θ ( 1 ) so C -approximate recovery requires finding i . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  45. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Non-adaptive lower bound: why is this hard? Hard case: x is random e i plus Gaussian noise w . Noise � w � 2 2 = Θ ( 1 ) so C -approximate recovery requires finding i . Observations � v , x � = v i + � v , w � = v i + � v � 2 √ n z , for z ∼ N ( 0 , Θ ( 1 )) . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  46. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  47. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  48. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  49. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  50. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  51. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  52. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) where SNR denotes the “signal-to-noise ratio,” SNR = E [ signal 2 ] E [ v 2 i ] � 2 / n = 1 E [ noise 2 ] � v � 2 Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  53. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) where SNR denotes the “signal-to-noise ratio,” SNR = E [ signal 2 ] E [ v 2 i ] � 2 / n = 1 E [ noise 2 ] � v � 2 Finding i needs Ω ( log n ) non-adaptive measurements. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  54. 1-sparse recovery: changes in adaptive setting Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) . where SNR denotes the “signal-to-noise ratio,” � � E [ v 2 i ] SNR = Θ . � v � 2 2 / n Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 19 / 29

  55. 1-sparse recovery: changes in adaptive setting Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) . where SNR denotes the “signal-to-noise ratio,” � � E [ v 2 i ] SNR = Θ . � v � 2 2 / n If i is independent of v , this is O ( 1 ) . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 19 / 29

  56. 1-sparse recovery: changes in adaptive setting Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) . where SNR denotes the “signal-to-noise ratio,” � � E [ v 2 i ] SNR = Θ . � v � 2 2 / n If i is independent of v , this is O ( 1 ) . As we learn about i , we can increase E [ v 2 i ] for constant � v � 2 . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 19 / 29

  57. 1-sparse recovery: changes in adaptive setting Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) . where SNR denotes the “signal-to-noise ratio,” � � E [ v 2 i ] SNR = Θ . � v � 2 2 / n If i is independent of v , this is O ( 1 ) . As we learn about i , we can increase E [ v 2 i ] for constant � v � 2 . ◮ Equivalently, for constant E [ v 2 i ] we can decrease � v � 2 . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 19 / 29

  58. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits v SNR = 2 I ( i , � v , x � ) � log SNR = 1 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  59. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits 1 bit v SNR = 2 2 I ( i , � v , x � ) � log SNR = 2 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  60. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits 1 bit 2 bits v SNR = 2 4 I ( i , � v , x � ) � log SNR = 4 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  61. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits 1 bit 2 bits 4 bits v SNR = 2 8 I ( i , � v , x � ) � log SNR = 8 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  62. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits 1 bit 2 bits 4 bits 8 bits v SNR = 2 16 I ( i , � v , x � ) � log SNR = 16 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  63. Goal Shown intuition for specific distribution on x Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

  64. Goal Shown intuition for specific distribution on x Match previous convergence for arbitrary x = α e i + w ? Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

  65. Goal Shown intuition for specific distribution on x Match previous convergence for arbitrary x = α e i + w ? ◮ α may not be 1. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

  66. Goal Shown intuition for specific distribution on x Match previous convergence for arbitrary x = α e i + w ? ◮ α may not be 1. ◮ Work for a specific x with 3 / 4 probability. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

  67. Goal Shown intuition for specific distribution on x Match previous convergence for arbitrary x = α e i + w ? ◮ α may not be 1. ◮ Work for a specific x with 3 / 4 probability. ◮ Distribution over A , for fixed w . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

Recommend


More recommend