3d vision
play

3D Vision Viktor Larsson Spring 2019 Schedule Feb 18 - PowerPoint PPT Presentation

3D Vision Viktor Larsson Spring 2019 Schedule Feb 18 Introduction Feb 25 Geometry, Camera Model, Calibration Mar 4 Features, Tracking / Matching Mar 11 Project Proposals by Students Mar 18 Structure from Motion (SfM) + papers Mar 25


  1. 3D Vision Viktor Larsson Spring 2019

  2. Schedule Feb 18 Introduction Feb 25 Geometry, Camera Model, Calibration Mar 4 Features, Tracking / Matching Mar 11 Project Proposals by Students Mar 18 Structure from Motion (SfM) + papers Mar 25 Dense Correspondence (stereo / optical flow) + papers Apr 1 Bundle Adjustment & SLAM + papers Apr 8 Student Midterm Presentations Apr 15 Multi-View Stereo & Volumetric Modeling + papers Easter break Apr 22 Apr 29 3D Modeling with Depth Sensors + papers May 6 3D Scene Understanding + papers May 13 4D Video & Dynamic Scenes + papers May 20 papers May 27 Student Project Demo Day = Final Presentations

  3. 3D Vision – Class 3 Features & Correspondences feature extraction, image descriptors, feature matching, feature tracking Chapters 4, 8 in Szeliski’s Book [Shi & Tomasi, Good Features to Track, CVPR 1994]

  4. Overview • Local Features • Invariant Feature Detectors • Invariant Descriptors & Matching • Feature Tracking

  5. Importance of Features Features are key component of many 3D Vision algorithms

  6. Importance of Features Schönberger & Frahm, Structure-From-Motion Revisited, CVPR 2016

  7. Feature Detectors & Descriptors • Detector : Find salient structures • Corners, blob-like structures, ... • Keypoints should be repeatable • Descriptor : Compact representation of image region around keypoint • Describes patch around keypoints • Establish matches between images by comparing descriptors

  8. Feature Detectors & Descriptors (Lowe, Distinctive Image Features From Scale-Invariant Keypoints , IJCV’04)

  9. Feature Matching vs. Tracking Matching Tracking • Extract features independently • Extract features in first image • Match by comparing descriptors • Find same feature in next view

  10. Wide Baseline Matching • Requirement to cope with larger variations between images  • Translation, rotation, scaling geometric • Perspective foreshortening transformations  • Non-diffuse reflections photometric transformations • Illumination

  11. Good Detectors & Descriptors? • What are the properties of good detectors and descriptors? • Invariances against transformations • How to design such detectors and descriptors? • This lecture: • Feature detectors & their invariances • Feature descriptors, invariances, & matching • Feature tracking

  12. Overview • Local Features Intro • Invariant Feature Detectors • Invariant Descriptors & Matching • Feature Tracking

  13. Good Feature Detectors? • Desirable properties? • Precise (sub-pixel perfect) localization • Repeatable detections under • Rotation • Translation • Illumination • Perspective distortions • … • Detect distinctive / salient structures

  14. Feature Point Extraction • Find “distinct” keypoints (local image patches) • As different as possible from neighbors homogeneous edge corner

  15. Comparing Image Regions • Compare intensities pixel-by-pixel I ´ (x,y) I(x,y) • Dissimilarity measure: Sum of Squared Differences / Distances ( SSD )     2 SSD  I ( x , y )  I ( x , y )  x y ฀

  16. Finding Stable Features • Measure uniqueness of candidate • Approximate SSD for small displacement Δ    2 SSD  w ( x i ) I ( x i   )  I ( x i ) i     2 w ( x i ) I ( x i )   I  I     I ( x i )      x  y     i   ฀ 2 I x   I I x I x I y   w ( x i )  T      T M   x 2   I x I y I y i ฀ • possible weights ฀ ฀

  17. Finding Stable Features homogeneous edge corner Suitable feature positions should maximize i.e. maximize smallest eigenvalue of M

  18. Harris Corner Detector • Use small local window: • Directly computing eigenvalues λ 1 , λ 2 of M is computationally expensive • Alternative measure for “ cornerness ”: = 𝜇 1 ⋅ 𝜇 2 − 𝑙 𝜇 1 + 𝜇 2 2 • Homogeneous: 𝜇 1 , 𝜇 2 small ⇒ 𝑆 small 2 < 0 • Edge: 𝜇 1 ≫ 𝜇 2 ≈ 0 ⇒ 𝑆 = 𝜇 1 ⋅ 0 − 𝑙𝜇 1 • Corner: 𝜇 1 , 𝜇 2 large ⇒ 𝑆 large

  19. Harris Corner Detector • Alternative measure for “ cornerness ” • Select local maxima as keypoints • Subpixel accuracy through second order surface fitting (parabola in 1D)

  20. Harris Corner Detector • Keypoint detection: Select strongest features over whole image or over each tile (e.g. 1000 per image or 2 per tile) • Invariances against geometric transformations • Shift / translation?

  21. Geometric Invariances Rotation Harris: Yes Scale Harris: No Affine (approximately invariant w.r.t. perspective/viewpoint) Harris: No

  22. 2D Transformations of a Patch Harris corners VIP Harris corners MSER SIFT

  23. Scale-Invariant Feature Transform (SIFT) • Detector + descriptor (later) • Recover features with position, orientation and scale (Lowe, Distinctive Image Features From Scale-Invariant Keypoints , IJCV’04)

  24. Position • Look for strong responses of Difference-of- Gaussian filter ( DoG ) 3 2  k • Approximates Laplacian of Gaussian ( LoG ) • Detects blob-like structures • Only consider local extrema

  25. Scale • Look for strong DoG responses over scale space    ฀  1/2 image ( σ =2) ฀ ฀ ฀    ฀  ฀ orig. image   4 ฀ 2 Slide credits: Bastian Leibe, Krystian Mikolajczyk ฀ ฀

  26. Scale • Only consider local maxima/minima in both position and scale • Fit quadratic around extrema for sub- pixel & sub-scale accuracy

  27. Minimum Contrast and “ Cornerness ” all features

  28. Minimum Contrast and “ Cornerness ” after suppressing edge-like features

  29. Minimum Contrast and “ Cornerness ” after suppressing edge-like features + small contrast features

  30. Invariants So Far • Translation? Yes • Scale? Yes • Rotation? Yes

  31. Orientation Assignment • Compute gradient for each pixel in patch at selected scale • Bin gradients in histogram & smooth histogram • Select canonical orientation at peak(s) • Keypoint = 4D coordinate 2  0 (x, y, scale, orientation)

  32. Invariants So Far • Translation • Scale • Rotation • Brightness changes: • Additive changes? • Multiplicative changes?

  33. 2D Transformations of a Patch Harris corners VIP Harris corners MSER SIFT

  34. Affine Invariant Features Perspective effects can locally be approximated by affine transformation

  35. Extreme Wide Baseline Matching • Detect stable keypoints using the Maximally Stable Extremal Regions ( MSER ) detector • Detections are regions , not points! (Matas et al., Robust Wide Baseline Stereo from Maximally Stable Extremal Regions, BMVC’02)

  36. Maximally Stable Extremal Regions Extremal regions: • Much brighter than surrounding • Use intensity threshold

  37. Maximally Stable Extremal Regions Extremal regions: • OR: Much darker than surrounding • Use intensity threshold

  38. Maximally Stable Extremal Regions • Regions: Connected components at a threshold • Region size = #pixels • Maximally stable: Region constant near some threshold

  39. A Sample Feature

  40. A Sample Feature T is maximally stable wrt. surrounding

  41. From Regions To Ellipses • Compute „ center of gravity “ • Compute Scatter (PCA / Ellipsoid)

  42. From Regions To Ellipses • Ellipse abstracts from pixels! • Geometric representation: position/size/shape

  43. Achieving Invariance • Normalize to „ default “ position, size, shape • For example: Circle of radius 16 pixels

  44. • Normalize ellipse to circle (affine transformation) • 2D rotation still unresolved

  45. • Same approach as for SIFT: Compute histogram of local gradients • Find dominant orientation in histogram • Rotate local patch into dominant orientation

  46. Summary: MSER Features • Detect sets of pixels brighter/darker than surrounding pixels • Fit elliptical shape to pixel set • Warp image so that ellipse becomes circle • Rotate to dominant gradient direction (other constructions possible as well)

  47. MSER Features - Invariants • Constant brightness changes (additive and multiplicative) • Rotation, translation, scale • Affine transformations  Affine normalization of feature leads to similar patches in different views !

  48. 2D Transformations of a Patch Harris corners VIP In practice hardly observable for small patches ! Harris corners MSER SIFT

  49. Viewpoint Invariant Patches (VIP) • Use known planar geometry to remove perspective distortion • Or: Use vanishing points to rectify patch (Wu et al., 3D Model Matching with Viewpoint Invariant Patches (VIPs), CVPR’08)

  50. Learning Feature Detectors • In the age of deep learning, can we learn good detectors from data? • How can we model repeatable feature detection? • Learn ranking function H(x|w): R 2 → [ -1, 1] with parameters w • Interesting points close to -1 or 1 (Savinov et al., Quad-networks: unsupervised learning to rank for interest point detection, CVPR’17)

Recommend


More recommend