A Secant Method for Nonsmooth Optimization Asef Nazari CSIRO Melbourne CARMA Workshop on Optimization, Nonlinear Analysis, Randomness and Risk Newcastle, Australia 12 July, 2014 Asef Nazari A Secant Method for Nonsmooth Optimization
Outline Background Secant Method 1 Background The Problem Optimization Algorithms and components Subgradient and Subdifferential 2 Secant Method Definitions Optimality Condition and Descent Direction The Secant Algorithm Numerical Results Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential The problem Unconstrained Optimization Problem x ∈ R n f ( x ) min f : R n → R n Locally Lipshitz ¨ O Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential The problem Unconstrained Optimization Problem x ∈ R n f ( x ) min f : R n → R n Locally Lipshitz ¨ O Why just unconstrained? Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Transforming Constrained into Unconstrained Constrained Optimization Problem min x ∈ Y f ( x ) where Y ⊂ R n . Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Transforming Constrained into Unconstrained Constrained Optimization Problem min x ∈ Y f ( x ) where Y ⊂ R n . Distance function dist ( x , Y ) = min y ∈ Y � y − x � Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Transforming Constrained into Unconstrained Constrained Optimization Problem min x ∈ Y f ( x ) where Y ⊂ R n . Distance function dist ( x , Y ) = min y ∈ Y � y − x � Theory of penalty function (under some conditions) x ∈ R n f ( x ) + σ dist ( x , y ) min f ( x ) + σ dist ( x , y ) is a nonsmooth function Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Sources of Nonsmooth Problems Minimax Problem x ∈ R n f ( x ) min f ( x ) = max 1 ≤ i ≤ m f i ( x ) Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Sources of Nonsmooth Problems Minimax Problem x ∈ R n f ( x ) min f ( x ) = max 1 ≤ i ≤ m f i ( x ) System of Nonlinear Equations f i ( x ) = 0 , i = 1 , . . . , m , Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Sources of Nonsmooth Problems Minimax Problem x ∈ R n f ( x ) min f ( x ) = max 1 ≤ i ≤ m f i ( x ) System of Nonlinear Equations f i ( x ) = 0 , i = 1 , . . . , m , we often do x ∈ R n � ¯ min f ( x ) � where ¯ f ( x ) = ( f 1 ( x ) , . . . , f m ( x )) Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Structure of Optimization Algorithms Step 1 Initial Step x 0 ∈ R n Step 2 Termination Criteria Step 3 Finding descent direction d k at x k Step 4 Finding step size f ( x k + α k d k ) < f ( x k ) Step 5 Loop x k +1 = x k + α k d k and go to step 2. Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Classification of Algorithms Based on d k and α k Directions d k = −∇ f ( x k ) Steepest Descent Method d k = − H − 1 ( x k ) ∇ f ( x k ) Newton Method d k = − B − 1 ∇ f ( x k ) Quasi-Newton Method Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Classification of Algorithms Based on d k and α k Directions d k = −∇ f ( x k ) Steepest Descent Method d k = − H − 1 ( x k ) ∇ f ( x k ) Newton Method d k = − B − 1 ∇ f ( x k ) Quasi-Newton Method Step sizes h ( α ) = f ( x k + α d k ) exactly solve h ′ ( α ) = 0 exact line search 1 loosly solve it, inexact line serach 2 Trust region methods Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential When the objective function is smooth Descent direction f ′ ( x k , d ) < 0 f ′ ( x k , d ) = �∇ f ( x k ) , d � ≤ 0 descent direction −∇ f ( x k ) steepest descent direction �∇ f ( x k ) � ≤ ε good stopping criteria (First Order Necessary Conditions) Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subgradient Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subgradient Basic Inequality for convex differentiable function: f ( x ) ≥ f ( y ) + ∇ f ( y ) T . ( x − y ) ∀ x ∈ Dom ( f ) Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subgradient g ∈ R n is a subgradient of a convex function f at y if f ( x ) ≥ f ( y ) + g T . ( x − y ) ∀ x ∈ Dom ( f ) Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subgradient g ∈ R n is a subgradient of a convex function f at y if f ( x ) ≥ f ( y ) + g T . ( x − y ) ∀ x ∈ Dom ( f ) Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subgradient g ∈ R n is a subgradient of a convex function f at y if f ( x ) ≥ f ( y ) + g T . ( x − y ) ∀ x ∈ Dom ( f ) if x < 0, g = − 1 Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subgradient g ∈ R n is a subgradient of a convex function f at y if f ( x ) ≥ f ( y ) + g T . ( x − y ) ∀ x ∈ Dom ( f ) if x < 0, g = − 1 if x > 0, g = 1 Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subgradient g ∈ R n is a subgradient of a convex function f at y if f ( x ) ≥ f ( y ) + g T . ( x − y ) ∀ x ∈ Dom ( f ) if x < 0, g = − 1 if x > 0, g = 1 if x = 0, g ∈ [ − 1 , 1] Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subdifferential Set of all subgradients of f at x , ∂ f ( x ), is called subdifferential. Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subdifferential Set of all subgradients of f at x , ∂ f ( x ), is called subdifferential. for f ( x ) = | x | , the subdifferential at x = 0 is ∂ f (0) = [ − 1 , 1] For x > 0 ∂ f ( x ) = { 1 } For x < 0 ∂ f ( x ) = {− 1 } Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subdifferential Set of all subgradients of f at x , ∂ f ( x ), is called subdifferential. for f ( x ) = | x | , the subdifferential at x = 0 is ∂ f (0) = [ − 1 , 1] For x > 0 ∂ f ( x ) = { 1 } For x < 0 ∂ f ( x ) = {− 1 } f is differentiable at x ⇐ ⇒ ∂ f ( x ) = {∇ f ( x ) } Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Subdifferential Set of all subgradients of f at x , ∂ f ( x ), is called subdifferential. for f ( x ) = | x | , the subdifferential at x = 0 is ∂ f (0) = [ − 1 , 1] For x > 0 ∂ f ( x ) = { 1 } For x < 0 ∂ f ( x ) = {− 1 } f is differentiable at x ⇐ ⇒ ∂ f ( x ) = {∇ f ( x ) } for Lipschitz functions (Clarke) ∂ f ( x 0 ) = conv { lim ∇ f ( x i ) : x i − → x 0 ∇ f ( x i ) exists } Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Optimality Condition For a smooth convex function f ( x ∗ ) = ⇒ 0 = ∇ f ( x ∗ ) x ∈ Dom ( f ) f ( x ) ⇐ inf Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Optimality Condition For a smooth convex function f ( x ∗ ) = ⇒ 0 = ∇ f ( x ∗ ) x ∈ Dom ( f ) f ( x ) ⇐ inf For a nonsmooth convex function f ( x ∗ ) = ⇒ 0 ∈ ∂ f ( x ∗ ) x ∈ Dom ( f ) f ( x ) ⇐ inf Asef Nazari Secant Method
Outline The Problem Background Optimization Algorithms and components Secant Method Subgradient and Subdifferential Directional derivative In Smooth Case f ( x k + λ d ) − f ( x k ) f ′ ( x k , d ) = lim = �∇ f ( x k ) , d � λ λ → 0 + Asef Nazari Secant Method
Recommend
More recommend