Home Laboratory Computer Lab Theses

2.5 Unconstrained problems

Definition 2.14: Unconstrained Programs

An unconstrained program is a problem of the form to minimize (or maximize) f ( x ) without any constraints on the variables x :

min { f ( x ) : x n x }

Remark 2.15

Note that, being the feasible domain of x unbounded, theorem 2.1 does not apply. Thus, one does not know with certainty whether a minimum actually exists for that problem a a For unconstrained optimization problems, the existence of a minimum can actually be guaranteed if the objective function is such that lim x + f ( x ) = + (O-coercive function). . Moreover, even if the objective function is convex, one such minimum may not exist (think of f : x exp x !). Hence, we shall proceed with the theoretically unattractive task of seeking minima and maxima of functions which need not have them!

Given a point x n x , necessary conditions help determine whether or not a point is a local or a global minimum of a function f . For this purpose, we are mostly interested in obtaining conditions that can be checked algebraically.

Definition 2.15: Descent Direction

Suppose that f : n x is continuous at x ¯ . A vector d n x is said to be a descent direction, or an improving direction, for f at x ¯ if

δ > 0 : f ( x ¯ + λ d ) < f ( x ¯ ) λ ( 0 , δ )

Moreover, the cone of descent directions at x ¯ , denoted by ( x ¯ ) , is given by

( x ¯ ) : = { d : δ > 0  such that  f ( x ¯ + λ d ) < f ( x ¯ ) λ ( 0 , δ ) }

The definition 2.15 provides a geometrical characterization for a descent direction. Yet, an algebraic characterization for a descent direction would be more useful from a practical point of view. In response to this, let us assume that f is differentiable and define the following set at x ¯ :

0 ( x ¯ ) : = { d : f ( x ¯ ) d < 0 }

This is illustrated in figure ?? where the half-space 0 ( x ¯ ) and the gradient f ( x ¯ ) are translated from the origin to x ¯ for convenience.

Lemma 2.1 proves that every element d 0 ( x ¯ ) is a descent direction at x ¯ .

Lemma 2.1: Algebraic Characterization of a Descent Direction

Suppose that f : n x is differentiable at x ¯ . If there exists a vector d such that f ( x ¯ ) d < 0 , then d is a descent direction for f at x ¯ . That is,

0 ( x ¯ ) ( x ¯ )

Proof. f being differentiable at x ¯ ,

f ( x ¯ + λ d ) = f ( x ¯ ) + λ f ( x ¯ ) d + o ( λ )

where lim λ 0 o ( λ ) λ = 0 . Rearranging the terms and dividing by λ 0 , we get

f ( x ¯ + λ d ) f ( x ¯ ) λ = f ( x ¯ ) d + o ( λ ) λ

Since f ( x ¯ ) d < 0 and lim λ 0 o ( λ ) λ = 0 , there exists a δ > 0 such that f ( x ¯ ) d + o ( λ ) λ < 0 for all λ ( 0 , δ ) . □

We are now ready to derive a number of necessary conditions for a point to be a local minimum of an unconstrained optimization problem.

Theorem 2.3: First-Order Necessary Condition for a Local Minimum

Suppose that f : n x is differentiable at x . If x is a local minimum, then f ( x ) = 0 .

Proof. The proof proceeds by contraposition. Suppose that f ( x ) 0 . Then, letting d = f ( x ) , we get f ( x ) d = f ( x ) 2 < 0 . By lemma 2.1,

δ > 0 : f ( x + λ d ) < f ( x ) λ ( 0 , δ )

hence contradicting the assumption that x is a local minimum for f . □

Remark 2.16: Obtaining Candidate Solution Points

The above condition is called a first-order necessary condition because it uses the first-order derivatives of f . This condition indicates that the candidate solutions to an unconstrained optimization problem can be found by solving a system of n x algebraic (nonlinear) equations. Points x ¯ such that f ( x ¯ ) = 0 are known as stationary points. Yet, a stationary point need not be a local minimum; it could very well be a local maximum or even a saddle point.

Example 2.8. Consider the problem

min x x 6 + x 4 + x 3 x 2

The gradient vector of the objective function is given by

f ( x ) = 6 x 5 + 4 x 3 + 3 x 2 2 x

which has three distinct roots x 1 , x 2 and x 3 . Out of these values, x 2 gives the smallest cost value. Yet, we cannot declare x 2 to be the global minimum because we do not know whether a (global) minimum exists for this problem. Indeed, as shown in figure 2.4 , none of the stationary points is a global minimum because f decreases to as | x | .

pict

Figure 2.4:: Illustration of the objective function and its derivative in Example 2.13

More restrictive necessary conditions can also be derived in terms of the Hessian matrix 2 f whose elements are the second-order derivatives of f . One such second-order condition is given below.

Theorem 2.4: Second-Order Necessary Conditions for a Local Minimum

Suppose that f : n x is twice differentiable at x . If x is a local minimum, then f ( x ) = 0 and 2 f ( x ) is positive semidefinite.

Proof. Consider an arbitrary direction d . Then, from the differentiability of f at x , we have

f ( x + λ d ) = f ( x ) + λ f ( x ) d + λ 2 2 d 2 f ( x ) d + o ( λ 2 )

where lim λ 0 o ( λ 2 ) λ 2 = 0 . Since x is a local minimum, from Theorem 2.3, f ( x ) = 0 . Rearranging the terms and dividing by λ 2 , we get

f ( x + λ d ) f ( x ) λ 2 = 1 2 d 2 f ( x ) d + o ( λ 2 ) λ 2

Since x is a local minimum, f ( x + λ d ) f ( x ) for λ sufficiently small. By taking the limit as λ 0 , it follows that d 2 f ( x ) d 0 . Since d is arbitrary, 2 f ( x ) is therefore positive semidefinite. □

Example 2.9. Consider the problem

min x 2 1 2 x [ 3 2 2 1 . 5 ] x = 1 2 x A x

The gradient vector of the objective function is given by

f ( x ) = A x

so that the only stationary point in 2 is x ¯ = ( 0 , 0 ) . Now, consider the Hessian matrix of the objective function at x ¯ :

2 f ( x ¯ ) = A = ( 3 2 2 1 . 5 ) x 2

It is easily checked that 2 f ( x ¯ ) is indefinite, eig ( A ) = { 3 . 8 , 2 . 3 } . Therefore, by Theorem 2.4 , the stationary point x ¯ is not a (local) minimum (nor is it a local maximum). Such stationary points are called saddle points (see Figure 2.5 ).

pict pict

Figure 2.5:: Plot and contour plot of saddle point case

The conditions presented in theorems 2.3 and 2.4 are necessary conditions. That is, they must hold true at every local optimal solution. Yet, a point satisfying these conditions need not be a local minimum. Theorem 2.5 gives sufficient conditions for a stationary point to be a global minimum point, provided the objective function is convex on n x .

Theorem 2.5: First-Order Sufficient Conditions for a Global Minimum for Convex Functions

Suppose that f : n x is differentiable at x and convex on n x . If f ( x ) = 0 , then x is a global minimum of f on n x .

Proof. f being convex on n x and differentiable at x , by Theorem 2.6, we have

f ( x ) f ( x ) + f ( x ) [ x x ] x n x

But since x is a stationary point,

f ( x ) f ( x ) x n x

Theorem 2.6: First-Order Condition of Convexity

Let C be a convex set in n x with a nonempty interior, and let f : C be a function. Suppose f is continuous on C and differentiable on int ( C ) . Then f is convex on int ( C ) if and only if

f ( y ) f ( x ) + f ( x ) [ y x ]

holds for any two points x , y C .

Proof. ... □

Theorem 2.7: Second-Order Condition of Convexity

Let C be a convex set in n x with a nonempty interior, and let f : C be a function. Suppose f is continuous on C and twice differentiable on int ( C ) . Then f is convex on int ( C ) if and only if

2 f ( x ) 0 x C

Proof. ... □

Remark 2.17

The first-order condition of strict convexity is:

f ( y ) > f ( x ) + f ( x ) [ y x ] x , y C

While the second-order condition is:

2 f ( x ) 0 x C

The convexity condition required by theorem 2.5 is actually very restrictive, in the sense that many practical problems are nonconvex. In theorem 2.8, we give sufficient conditions for characterizing a local minimum point, provided the objective function is strictly convex in a neighborhood of that point.

Theorem 2.8: Second-Order Sufficient Conditions for a Strict Local Minimum

Suppose that f : n x is twice differentiable at x . If f ( x ) = 0 and 2 f ( x ) is positive definite, then x is a local minimum of f .

Proof. f being twice differentiable at x , we have

f ( x + d ) = f ( x ) + f ( x ) d + 1 2 d 2 f ( x ) d + o ( d 2 )

for each d n x , where lim d 0 o ( d 2 ) d 2 = 0 . Let λ L denote the smallest eigenvalue of 2 f ( x ) . Then, 2 f ( x ) being positive definite we have λ L > 0 and d 2 f ( x ) d λ L d 2 . Moreover, from f ( x ) = 0 we get

f ( x + d ) f ( x ) [ λ L 2 + o ( d 2 ) d 2 ] d 2

Since lim d 0 o ( d 2 ) d 2 = 0 ,

η > 0  such that  | o ( d 2 ) d 2 | < λ L 2 d 𝔹 η ( 0 )

and finally,

f ( x + d ) f ( x ) λ L 2 d 2 > 0 d 𝔹 η ( 0 ) { 0 }

i.e., x is a strict local minimum of f . □

Example 2.10. Consider the problem

min x 2 1 2 [ x 1 1 x 2 2 ] [ 5 2 2 2 ] [ x 1 1 x 2 2 ] = 1 2 [ x 1 1 x 2 2 ] A [ x 1 1 x 2 2 ]

The gradient vector and Hessian matrix at x ¯ = ( 1 , 2 ) are given by

f ( x ¯ ) = A [ x 1 1 x 2 2 ] = 0
2 f ( x ¯ ) = A 0

Note that A 0 since eig ( A ) = { 6 , 1 } and also that 2 f 0 is the second-order characterization of strict convexity. Hence, by Theorem 2.8 , x ¯ is a local minimum of f ( x ¯ is also a global minimum of f on 2 since f is convex). The objective function is pictured in Figure 2.6 below.

We close this subsection by re-emphasizing the fact that every local minimum of an unconstrained optimization problem min { f ( x ) : x n x } is a global minimum if f is a convex function on n x . Yet, convexity of f is not a necessary condition for each local minimum to be a global minimum.

pict pict

Figure 2.6:: Plot and contour plot of convex quadratic function