image masking schemes for local manifold learning methods
play

Image Masking Schemes for Local Manifold Learning Methods Marco F. - PowerPoint PPT Presentation

Image Masking Schemes for Local Manifold Learning Methods Marco F. Duarte Joint work with Hamid Dadkhahi IEEE ICASSP - April 23 2015 Manifold Learning Given training points in , learn the mapping to the underlying K -dimensional


  1. Image Masking Schemes for 
 Local Manifold Learning Methods Marco F. Duarte Joint work with Hamid Dadkhahi IEEE ICASSP - April 23 2015

  2. Manifold Learning • Given training points in , learn the mapping � to the underlying K -dimensional articulation manifold • Exploit local geometry 
 to capture parameter 
 differences by embedding 
 distances • ISOMAP, LLE, HLLE, … • Ex: � images of � rotating teapot � � � articulation space � = circle �

  3. Compressive Manifold Learning • Given training points in , learn the mapping � to the underlying K -dimensional articulation manifold • Isomap algorithm approximates geodesic distances using distances between neighboring points • Random measurements preserve these distances • Theorem : If , then the Isomap 
 residual variance in the projected [Hegde, Wakin, 
 Baraniuk 2008] domain is bounded by the additive � error factor translating 
 disk manifold 
 N = 4096 (Full Data) M = 50 M = 25 M = 100 ( K =2)

  4. Custom Projection Operators • Goal of Dimensionality Reduction: To preserve distances between points in the manifold, i.e., for • Collect pairwise differences into set of 
 secant vectors • Search for projection that preserves norms of secants: [Hegde, Sankaranarayanan, Baraniuk 2012]

  5. Custom Projection Operators • Usual approach: Principal Component Analysis (PCA) • Collect all secants into a matrix: • Perform eigenvalue decomposition on S : • Select top eigenvectors as projections • PCA minimizes the average squared distortion over secants, but can distort individual secants arbitrarily and therefore warp manifold structure [Hegde, Sankaranarayanan, Baraniuk 2012]

  6. Custom Projection Operators: NuMax • For target distortion , find matrix featuring the smallest number of rows that yields • This is equivalent to minimizing the rank 
 of the matrix such that • Use nuclear norm as proxy for rank to 
 obtain computationally efficient approach • Improves over random projections since matrix is 
 specifically tailored to manifold observed • May be difficult to link target distortion to matrix rank/ 
 number of rows [Hegde, Sankaranarayanan, Baraniuk 2012]

  7. Issues with Randomness and NuMax • Projections matrices have entries 
 with arbitrary values • Physics of sensing process, 
 hardware devices restrict types 
 of projections we can obtain • Example: Low-power imaging 
 for computational eyeglasses • Low-power imaging sensor 
 allows for individual selection of 
 pixels to record • Power consumption proportional 
 to number of pixels sampled • Random projections/NuMax involve 
 half/all pixels and do not enable 
 power savings • How to derive constrained 
 projection matrices that involve 
 [Mayberry, Hu, Marlin, 
 only few pixels? Salthouse, Ganesan 2014]

  8. Masking Strategies for Manifold Data • Select only a subset of the pixels of size M that minimizes distortion to manifold structure • Emulate strategies for projection design into mask design • Random Masking : 
 Pick M pixels uniformly at random across image M = 100 
 pixels

  9. Masking Strategies for Manifold Data • Select only a subset of the pixels of size M that minimizes distortion to manifold structure • Emulate strategies for projection design into mask design • Principal coordinate analysis : 
 Pick M coordinates that maximize variance among secants M = 100 
 pixels [Dadkhahi and Duarte 2014]

  10. Masking Strategies for Manifold Data • Select only a subset of the pixels of size M that minimizes distortion to manifold structure • Emulate strategies for projection design into mask design • Adaptation of NuMax : • Define secants from k -nearest neighbor graph: 
 • Pick masking matrix (row submatrix of I ) to minimize secant norm distortion after scaling: 
 • Combinatorial integer program M = 100 
 replaced by greedy approximation pixels [Dadkhahi and Duarte 2014]

  11. Isomap vs. Locally Linear Embedding • While Isomap employs distances between neighbors when designing the embedding, Locally Linear Embedding (LLE) employs additional local geometry , 
 representing each vector as a weighted 
 linear combination of its neighbors 
 • In particular, Isomap embedding is 
 sensitive to scaling of the point clouds, 
 while LLE isn’t

  12. Isomap vs. Locally Linear Embedding • While Isomap employs distances between neighbors when designing the embedding, Locally Linear Embedding (LLE) employs additional local geometry , 
 representing each vector as a weighted 
 linear combination of its neighbors 
 • In particular, Isomap embedding is 
 sensitive to scaling of the point clouds, 
 while LLE isn’t

  13. Isomap vs. Locally Linear Embedding • While Isomap employs distances between neighbors when designing the embedding, Locally Linear Embedding (LLE) employs additional local geometry , 
 representing each vector as a weighted 
 linear combination of its neighbors 
 • In particular, Isomap embedding is 
 sensitive to scaling of the point clouds, 
 while LLE isn’t • To preserve this additional local 
 information, we expand the set of secants 
 to include distances between neighbors of 
 each point

  14. Isomap vs. Locally Linear Embedding • While Isomap employs distances between neighbors when designing the embedding, Locally Linear Embedding (LLE) employs additional local geometry , 
 representing each vector as a weighted 
 linear combination of its neighbors 
 • In particular, Isomap embedding is 
 sensitive to scaling of the point clouds, 
 while LLE isn’t • To preserve this additional local 
 information, we expand the set of secants 
 to include distances between neighbors of 
 each point

  15. 
 
 
 Manifold-Aware Pixel Selection for LLE • Expand the set of secants considered: 
 • Compute squared norms of secants 
 in obtained from the original 
 images and from masked images; 
 collect into “norm” vectors 
 • Choose mask that maximizes sum of 
 cosine similarities between original and 
 masked “norm” vectors: 
 • Replace combinatorial optimization by 
 greedy forward selection algorithm • Cosine similarity is invariant to (local) scaling of point cloud

  16. Manifold-Aware Pixel Selection for LLE M = 100 pixels M = 100 pixels

  17. Performance Analysis: 
 LLE Embedding Error Full Data Full Data 10 PCA PCA SPCA SPCA Embedding Error 8 PCoA PCoA MAPS − LLE MAPS − LLE 6 MAPS − Isomap MAPS − Isomap Random Random 4 2 50 100 150 200 250 300 Embedding Dim./Masking Size m M

  18. Performance Analysis: 
 LLE Embedding Error Full Data PCA 2500 SPCA Embedding Error PCoA 2000 MAPS − LLE MAPS − Isomap 1500 Random 1000 500 50 100 150 200 250 300 Embedding Dim./Masking Size m M

  19. Performance Analysis: 
 LLE Embedding Error Full Data 50 SPCA PCA PCoA SPCA Embedding Error 40 MAPS − LLE PCoA MAPS − Isomap MAPS − LLE Random 30 20 10 50 100 150 200 250 300 Embedding Dim./Masking Size m M

  20. Performance Analysis: 2-D LLE M = 100 pixels

  21. Computational Eyeglasses: Eye Gaze Tracking Average Gaze Estimation Error 45 Full Data PCA SPCA 40 PCoA MAPS − LLE 35 MAPS − Isomap Random 30 Error measured 
 25 in “target” pixels 50 100 150 200 250 300 Embedding Dim./Masking Size m M

  22. Conclusions • Compressive sensing (CS) for manifold-modeled images via random or customized projections (NuMax) • New sensors enable CS by masking images, 
 i.e., restricting the type of projections • Our MAPS algorithms find image masks that best preserve geometric structure used during manifold learning for image datasets • Greedy algorithms provide good preservation of learned manifold embeddings, suitable for parameter estimation • While Isomap relies on distances between neighbors, LLE also leverages local geometric structure; different algorithms are optimal for these cases • Concept of subsampling as feature selection - supervised and unsupervised learning? http://www.ecs.umass.edu/~mduarte mduarte@ecs.umass.edu

Recommend


More recommend