differentiable mutual information and matrix exponential
play

Differentiable Mutual Information and Matrix Exponential for - PowerPoint PPT Presentation

Differentiable Mutual Information and Matrix Exponential for Multi-Resolution Image Registration Abhishek Nan, Matthew Tennant, Uriel Rubin, Nilanjan Ray Medical Imaging with Deep Learning, 2020 https://github.com/abnan/DRMIME Registration


  1. Differentiable Mutual Information and Matrix Exponential for Multi-Resolution Image Registration Abhishek Nan, Matthew Tennant, Uriel Rubin, Nilanjan Ray Medical Imaging with Deep Learning, 2020 https://github.com/abnan/DRMIME

  2. Registration

  3. With MSE Source: https://anhir.grand-challenge.org/

  4. Problems? Multi-modal images ● MSE won’t work ● Source: https://www.mathworks.com/discovery/image-registration.html

  5. Solution? Mutual Information ●

  6. Issues? Mutual Information for images is computed using joint histograms. ● Histograms are not differentiable. ● No gradient descent? ●

  7. Differentiable mutual information The function T is realized by a neural ● network with parameter θ . V( θ ) is differentiable and can be used as a ● objective function in place of MI.

  8. MINE for images

  9. Currently

  10. Matrix exponential Matrix exponential of a square matrix A is given by the following: ● Geometric transformation matrices can be obtained by exponential of a linear combination of basis ● matrices.

  11. Matrix Exponential (Examples)

  12. Why matrix exponentials?

  13. So far...

  14. More problems? Medical/Microscopy images often are extremely high resolution. So gradient descent can be ● extremely slow. Optimization for neural networks is non-convex. ●

  15. Solution? Gaussian Pyramids ● Source: https://en.wikipedia.org/wiki/Pyramid_(image_processing)

  16. Can we do better? ● What if we did simultaneous optimization for all levels? ○ Each level of optimization is for different images. So MI between them changes as well. Solution? ● A single MINE can be trained for all of these! ○ How? ● Mini-batches can be constructed by sampling from all levels ○

  17. What about the loss? Since we are doing simultaneous optimization, with modern deep learning frameworks, it’s very ● easy to combine the loss from each level and perform joint optimization. For eg, for just 1 level: ● Loss = MI(F, G v (M)) ○ For 4 levels: ● Loss = (¼) * [MI(F 1 , G v (M 1 )) + MI(F 2 , G v (M 2 )) + MI(F 3 , G v (M 3 )) + MI(F 4 , G v (M 4 ))] ○

  18. Evaluation Public datasets ● Available ground truth ●

  19. Results

  20. Source: https://projects.ics.forth.gr/cvrl/fire/

  21. Thank you!

Recommend


More recommend