proximal point algorithm in hadamard spaces
play

Proximal point algorithm in Hadamard spaces Miroslav Bacak T el - PowerPoint PPT Presentation

Proximal point algorithm in Hadamard spaces Miroslav Bacak T el ecom ParisTech Optimisation G eom etrique sur les Vari et es - Paris, 21 novembre 2014 Contents of the talk 1 Basic facts on Hadamard spaces 2 Proximal point


  1. Proximal point algorithm in Hadamard spaces Miroslav Bacak T´ el´ ecom ParisTech Optimisation G´ eom´ etrique sur les Vari´ et´ es - Paris, 21 novembre 2014

  2. Contents of the talk 1 Basic facts on Hadamard spaces 2 Proximal point algorithm 3 Applications to computational phylogenetics

  3. Proximal point algorithm in Hadamard spaces Why? Well. . . it is used in: • Phylogenetics: computing medians and means of phylogenetic trees. • diffusion tensor imaging: the space P ( n, R ) of symmetric positive definite matrices n × n with real entries is a Hadamard space if it is equipped with the Riemannian metric A − 1 XA − 1 Y � X, Y � A := Tr � � , X, Y ∈ T A ( P ( n, R )) , for every A ∈ P ( n, R ) . • Computational biology: shape analyses of tree-like structures:

  4. Tree-like structures in organisms Figure: Bronchial tubes in lungs Figure: Human circulatory system Figure: Transport system in plants

  5. Definition of Hadamard space Let ( H , d ) be a complete metric space where: 1 any two points x 0 and x 1 are connected by a geodesic x : [0 , 1] → H : t �→ x t , 2 and, d ( y, x t ) 2 ≤ (1 − t ) d ( y, x 0 ) 2 + td ( y, x 1 ) 2 − t (1 − t ) d ( x 0 , x 1 ) 2 , for every y ∈ H . Then ( H , d ) is called a Hadamard space. For today: assume that local compactness.

  6. Geodesic space x 2 x 1

  7. Geodesic space x 2 x 1

  8. Geodesic space (1 − λ ) x 1 + λx 2 x 2 x 1

  9. Definition of nonpositive curvature A geodesic triangle in a geodesic space: x 3 x 2 x 1

  10. Terminology remark CAT( κ ) spaces, for κ ∈ R , were introduced in 1987 by Michail Gromov C = Cartan A = Alexandrov T = Toponogov We are particularly interested in CAT(0) spaces.

  11. Examples of Hadamard spaces 1 Hilbert spaces, the Hilbert ball 2 complete simply connected Riemannian manifolds with Sec ≤ 0 3 R -trees: a metric space T is an R -tree if • for x, y ∈ T there is a unique geodesic [ x, y ] • if [ x, y ] ∩ [ y, z ] = { y } , then [ x, z ] = [ x, y ] ∪ [ y, z ] 4 Euclidean buildings 5 the BHV tree space (space of phylogenetic trees) 6 L 2 ( M, H ) , where ( M, µ ) is a probability space: � 1 �� d ( u ( x ) , v ( x )) 2 d µ ( x ) 2 u, v ∈ L 2 ( M, H ) d 2 ( u, v ) := , M

  12. Convexity in Hadamard spaces Let ( H , d ) be a Hadamard space. These spaces allow for a natural definition of convexity: Definition A set C ⊂ H is convex if, given x, y ∈ C, we have [ x, y ] ⊂ C. Definition A function f : H → ( −∞ , ∞ ] is convex if f ◦ γ is a convex function for each geodesic γ : [0 , 1] → H .

  13. Convexity in Hadamard spaces Let ( H , d ) be a Hadamard space. These spaces allow for a natural definition of convexity: Definition A set C ⊂ H is convex if, given x, y ∈ C, we have [ x, y ] ⊂ C. Definition A function f : H → ( −∞ , ∞ ] is convex if f ◦ γ is a convex function for each geodesic γ : [0 , 1] → H .

  14. Convexity in Hadamard spaces Let ( H , d ) be a Hadamard space. These spaces allow for a natural definition of convexity: Definition A set C ⊂ H is convex if, given x, y ∈ C, we have [ x, y ] ⊂ C. Definition A function f : H → ( −∞ , ∞ ] is convex if f ◦ γ is a convex function for each geodesic γ : [0 , 1] → H .

  15. Examples of convex functions 1 The indicator function of a convex closed set C ⊂ H : ι C ( x ) := 0 , if x ∈ C, and ι C ( x ) := ∞ , if x / ∈ C. 2 The distance function to a closed convex subset C ⊂ H : d C ( x ) := inf c ∈ C d ( x, c ) , x ∈ H . 3 The displacement function of an isometry T : H → H : δ T ( x ) := d ( x, Tx ) , x ∈ H .

  16. Examples of convex functions 1 The indicator function of a convex closed set C ⊂ H : ι C ( x ) := 0 , if x ∈ C, and ι C ( x ) := ∞ , if x / ∈ C. 2 The distance function to a closed convex subset C ⊂ H : d C ( x ) := inf c ∈ C d ( x, c ) , x ∈ H . 3 The displacement function of an isometry T : H → H : δ T ( x ) := d ( x, Tx ) , x ∈ H .

  17. Examples of convex functions 1 The indicator function of a convex closed set C ⊂ H : ι C ( x ) := 0 , if x ∈ C, and ι C ( x ) := ∞ , if x / ∈ C. 2 The distance function to a closed convex subset C ⊂ H : d C ( x ) := inf c ∈ C d ( x, c ) , x ∈ H . 3 The displacement function of an isometry T : H → H : δ T ( x ) := d ( x, Tx ) , x ∈ H .

  18. Examples of convex functions 4 Let c : [0 , ∞ ) → H be a geodesic ray. The function b c : H → R defined by b c ( x ) := lim t →∞ [ d ( x, c ( t )) − t ] , x ∈ H , is called the Busemann function associated to the ray c. 5 The energy of a mapping u : M → H given by �� d ( u ( x ) , u ( y )) 2 p ( x, d y )d µ ( x ) , E ( u ) := M × M where ( M, µ ) is a measure space with a Markov kernel p ( x, d y ) . E is convex continuous on L 2 ( M, H ) .

  19. Examples of convex functions 4 Let c : [0 , ∞ ) → H be a geodesic ray. The function b c : H → R defined by b c ( x ) := lim t →∞ [ d ( x, c ( t )) − t ] , x ∈ H , is called the Busemann function associated to the ray c. 5 The energy of a mapping u : M → H given by �� d ( u ( x ) , u ( y )) 2 p ( x, d y )d µ ( x ) , E ( u ) := M × M where ( M, µ ) is a measure space with a Markov kernel p ( x, d y ) . E is convex continuous on L 2 ( M, H ) .

  20. Examples of convex functions 6 Given a 1 , . . . , a N ∈ H and w 1 , . . . , w N > 0 , set N w n d ( x, a n ) p , � f ( x ) := x ∈ H , n =1 where p ∈ [1 , ∞ ) . • If p = 1 , we get Fermat-Weber problem for optimal facility location. A minimizer of f is called a median. • If p = 2 , then a minimizer of f is the barycenter of N � µ := w n δ a n , n =1 or the mean of a 1 , . . . , a N .

  21. Examples of convex functions 6 Given a 1 , . . . , a N ∈ H and w 1 , . . . , w N > 0 , set N w n d ( x, a n ) p , � f ( x ) := x ∈ H , n =1 where p ∈ [1 , ∞ ) . • If p = 1 , we get Fermat-Weber problem for optimal facility location. A minimizer of f is called a median. • If p = 2 , then a minimizer of f is the barycenter of N � µ := w n δ a n , n =1 or the mean of a 1 , . . . , a N .

  22. Strongly convex functions A function f : H → ( −∞ , ∞ ] is strongly convex with parameter β > 0 if f ((1 − t ) x + ty ) ≤ (1 − t ) f ( x ) + tf ( y ) − βt (1 − t ) d ( x, y ) 2 , for any x, y ∈ H and t ∈ [0 , 1] . Each strongly has a unique minimizer. Example Given y ∈ H , the function f := d ( y, · ) 2 is strongly convex. Indeed, d ( y, x t ) 2 ≤ (1 − t ) d ( y, x 0 ) 2 + td ( y, x 1 ) 2 − t (1 − t ) d ( x 0 , x 1 ) 2 , for each geodesic x : [0 , 1] → H .

  23. 1 Basic facts on Hadamard spaces 2 Proximal point algorithm 3 Applications to computational phylogenetics

  24. Proximal point algorithm Let f : H → ( −∞ , ∞ ] be convex lsc. Optimization problem: min x ∈H f ( x ) . Recall: no (sub)differential, no shooting (singularities). Implicit methods are appropriate. The PPA generates a sequence � � f ( y ) + 1 d ( y, x i − 1 ) 2 x i := J λ i ( x i − 1 ) := arg min , 2 λ i y ∈H where x 0 ∈ H is a given starting point and λ i > 0 , for each i ∈ N .

  25. Proximal point algorithm Let f : H → ( −∞ , ∞ ] be convex lsc. Optimization problem: min x ∈H f ( x ) . Recall: no (sub)differential, no shooting (singularities). Implicit methods are appropriate. The PPA generates a sequence � � f ( y ) + 1 d ( y, x i − 1 ) 2 x i := J λ i ( x i − 1 ) := arg min , 2 λ i y ∈H where x 0 ∈ H is a given starting point and λ i > 0 , for each i ∈ N .

  26. Convergence of proximal point algorithm Theorem (M.B., 2011) Let f : H → ( −∞ , ∞ ] be a convex lsc function attaining its minimum. Given x 0 ∈ H and ( λ i ) such that � ∞ 1 λ i = ∞ , the PPA sequence ( x i ) converges to a minimizer of f. (Resolvents are firmly nonexpansive - cheap version for λ i = λ. ) Disadvantage: The resolvents � f ( y ) + 1 � d ( y, x i − 1 ) 2 x i := J λ i ( x i − 1 ) := arg min , 2 λ i y ∈H are often difficult to compute.

  27. Convergence of proximal point algorithm Theorem (M.B., 2011) Let f : H → ( −∞ , ∞ ] be a convex lsc function attaining its minimum. Given x 0 ∈ H and ( λ i ) such that � ∞ 1 λ i = ∞ , the PPA sequence ( x i ) converges to a minimizer of f. (Resolvents are firmly nonexpansive - cheap version for λ i = λ. ) Disadvantage: The resolvents � f ( y ) + 1 � d ( y, x i − 1 ) 2 x i := J λ i ( x i − 1 ) := arg min , 2 λ i y ∈H are often difficult to compute.

  28. Splitting proximal point algorithm Let f 1 , . . . , f N be convex lsc and consider N � f ( x ) := f n ( x ) , x ∈ H . n =1 Example (Median and mean) f n := d ( · , a n ) 2 . f n := d ( · , a n ) , Key idea: apply resolvents J n λ ’s of f n ’s in a cyclic or random order.

  29. Splitting proximal point algorithm Let f 1 , . . . , f N be convex lsc and consider N � f ( x ) := f n ( x ) , x ∈ H . n =1 Example (Median and mean) f n := d ( · , a n ) 2 . f n := d ( · , a n ) , Key idea: apply resolvents J n λ ’s of f n ’s in a cyclic or random order.

  30. Splitting proximal point algorithm Let f 1 , . . . , f N be convex lsc and consider N � f ( x ) := f n ( x ) , x ∈ H . n =1 Example (Median and mean) f n := d ( · , a n ) 2 . f n := d ( · , a n ) , Key idea: apply resolvents J n λ ’s of f n ’s in a cyclic or random order.

Recommend


More recommend