Home Laboratory Computer Lab Theses

2.8 General NLP problems

In this section, we shall consider general, nonlinear programming problems with both equality and inequality constraints,

 minimize  f ( x )  subject to  g i ( x ) 0 , i = 1 , , n g h i ( x ) = 0 , i = 1 , , n h

Before stating necessary and sufficient conditions for such problems, we shall start by revisiting the KKT conditions for inequality constrained problems, based on the method of Lagrange multipliers described in 2.7.

2.8.1 KKT Conditions for Inequality Constrained NLP Problems Revisited

Consider the problem to minimize a function f ( x ) for x S : = { x n x : g i ( x ) = 0 , i = 1 , . . . , n g } and suppose that x is a local minimum point. Clearly, x is also a local minimum of the inequality constrained problem where the inactive constraints g i ( x ) 0 , i 𝒜 ( x ) , have been discarded. Thus, in effect, inactive constraints at x don’t matter; they can be ignored in the statement of optimality conditions.

On the other hand, active inequality constraints can be treated to a large extent as equality constraints at a local minimum point. In particular, it should be clear that x is also a local minimum to the equality constrained problem

 minimize  f ( x )  subject to  g i ( x ) = 0 , i 𝒜 ( x )

That is, it follows from Theorem 2.13 that, if x is a regular point, there exists a unique Lagrange multiplier vector ν n g such that

f ( x ) + i 𝒜 ( x ) ν i g i ( x ) = 0

Assigning zero Lagrange multipliers to the inactive constraints, we obtain

f ( x ) + g ( x ) ν = 0 ν i = 0 , i 𝒜 ( x )

This latter condition can be rewritten by means of the following equations:

ν i g i ( x ) = 0 i = 1 , , n g

The argument showing that ν 0 is a little more elaborate. By contradiction, assume that ν l < 0 for some l 𝒜 ( x ) . Let A ( n g + 1 ) × n x ) be the matrix whose rows are f ( x ) and g i ( x ) , i = 1 , . . . , n g . Since x is a regular point, the Lagrange multiplier vector ν is unique. Therefore, the condition

A y = 0

can only be satisfied by y : = η ( 1 ν ) with η . Because ν l < 0 , we know by Theorem 2.1 that there exists a direction d ¯ n x such that A d ¯ < 0 . In other words,

d ¯ 0 ( x ) 𝒟 0 ( x )

which contradicts the hypothesis that x is a local minimizer of f on S (see Remark 2.18).

Overall, these results thus constitute the KKT optimality conditions as stated in Theorem 2.10. But although the foregoing development is straightforward, it is somewhat limited by the regularity-type assumption at the optimal solution. Obtaining more general constraint qualifications (see Remark 2.21) requires that the KKT conditions be derived using an alternative approach, e.g., the approach described earlier in Section 2.6. Still, the conversion to equality constrained problem proves useful in many situations, e.g., for deriving second-order sufficiency conditions for inequality constrained NLP problems.

2.8.2 Optimality Conditions for General NLP Problems

We are now ready to generalize the necessary and sufficient conditions given in Theorems 2.10, 2.13, 2.15 and 2.16 to general NLP problems.

Theorem 2.17: First- and Second-Order Necessary Conditions

Let f : n x , g i : n x , i = 1 , , n g and h i : n x , be twice continuously differentiable functions on n x . Consider the problem of minimizing f ( x ) subject to the constraints h ( x ) = 0 and g ( x ) = 0 . If x is a local minimum of the optimization problem and is a regular point of the constraints, then there exist unique vectors ν and λ such that

f ( x ) + h ( x ) λ + g ( x ) ν = 0 g ( x ) 0 h ( x ) = 0 g ( x ) ν = 0

and

y ( 2 f ( x ) + i = 1 n g ν i 2 g i ( x ) + i = 1 n h λ i 2 h i ( x ) ) y 0

for all y such that g i ( x ) y = 0 , i = 1 𝒜 ( x ) and h ( x ) y = 0

Theorem 2.18: Second-Order Sufficient Conditions

Let f : n x , g i : n x , i = 1 , , n g and h i : n x , i = 1 , , n h , be twice continuously differentiable functions on n x . Consider the problem of minimizing f ( x ) subject to the constraints h ( x ) = 0 and g ( x ) = 0 . If there exists x , ν and λ satisfying the KKT conditions in Theorem 2.17, and

y x x 2 ( x , ν , λ ) y > 0

for all y 0 such that

g i ( x ) y = 0 i 𝒜 ( x )  with  ν i > 0 g i ( x ) y 0 i 𝒜 ( x )  with  ν i = 0 h ( x ) y = 0

where = f ( x ) + h ( x ) λ + g ( x ) ν , then x is a strict local minimum.

Likewise, the KKT sufficient conditions given in Theorem 2.11 for convex, inequality constrained problems can be generalized to general convex problems as follows:

Theorem 2.19: KKT Sufficient Conditions for Convex Programs

Let f : n x , g i : n x , i = 1 , , n g be convex and differentiable functions. Let also h i : n x , i = 1 , , n h be affine functions. Consider the problem to minimize f ( x ) subject to x S : = { x n x : g ( x ) 0 , h ( x ) = 0 } . If ( x , ν , λ ) satisfies the KKT conditions of Theorem 2.17 then x is a global minimizer for f on S .

Theorem 2.20: Farkas’ Theorem

Let A m × n and c n . Then, exactly one of the following two statements holds:

System 1. x n such that A x 0 and c x > 0 ,
System 2. y n such that A y = c and y 0 .

Proof. See, e.g., [6, Theorem 2.4.5] for a proof. □

Farkas’ Theorem is used extensively in the derivation of optimality conditions of (linear and) nonlinear programming problems. A geometrical interpretation of Farkas’ Theorem is shown in Fig. 1.A.1.. If a1, . . . , am denote the rows of A, then system 2 has a solution if c lies in the convex cone generated by a1, . . . , am; On the other hand, system 1 has a solution if the closed convex cone x : Ax 0 and the open half-space x : cTx ¿ 0 have a nonempty intersection.

Corollary 2.1: Gordan’s Theorem

Let A m × n . Then, exactly one of the following two statements holds:

System 1. x n such that A x < 0 ,
System 2. y m y 0 such that A y = 0 and y 0 .

Proof. System 1 can be equivalently written as A x + ρ e where ρ > 0 is a scalar and e is a vector of m ones. Rewriting this in the form of System 1 in Farkas’ Theorem 2.20, we get ( A e ) p and ( 0 , . . . , 0 , 1 ) p > 0 where p : = ( x ρ ) . The associated System 2 by Theorem 2.20 states that ( A e ) ( 0 , . . . , 0 , 1 ) and y 0 for some y m , i.e., A y = 0 , e y = 1 and y 0 , which is equivalent to the System 2 of the corollary. □

Lemma 2.4

Let P and Q be two symmetric matrices, such that Q 0 and P 0 on the null space of Q (i.e., y P y > 0 y 0 such that Q y = 0 ). Then,

c ¯ > 0  such that  P + c Q 0 c > c ¯

Proof. Assume the contrary. Then,

k > 0 , x k , x k = 1  such that  x k P x k + k x k Q x k 0

Consider a subsequence { x k } 𝒦 converging to some x ¯ with x ¯ = 1 . Dividing the last equation by k , and taking the limit as k 𝒦 , we obtain

x ¯ Q x ¯ 0

On the other hand, Q being positive semidefinite, we must have

x ¯ Q x ¯ 0

Hence x ¯ Q x ¯ = 0 . That is, using the hypothesis, x ¯ P x ¯ > 0 . This contradicts the fact that

x ¯ P x ¯ + limsup k 0 k 𝒦 k x k Q x k 0