In this section, we shall consider general, nonlinear programming problems with both equality and inequality constraints,
Before stating necessary and sufficient conditions for such problems, we shall start by revisiting the KKT conditions for inequality constrained problems, based on the method of Lagrange multipliers described in 2.7.
Consider the problem to minimize a function and suppose that is a local minimum point. Clearly, is also a local minimum of the inequality constrained problem where the inactive constraints , have been discarded. Thus, in effect, inactive constraints at don’t matter; they can be ignored in the statement of optimality conditions.
On the other hand, active inequality constraints can be treated to a large extent as equality constraints at a local minimum point. In particular, it should be clear that is also a local minimum to the equality constrained problem
That is, it follows from Theorem 2.13 that, if is a regular point, there exists a unique Lagrange multiplier vector such that
Assigning zero Lagrange multipliers to the inactive constraints, we obtain
This latter condition can be rewritten by means of the following equations:
The argument showing that is a little more elaborate. By contradiction, assume that for some . Let be the matrix whose rows are and . Since is a regular point, the Lagrange multiplier vector is unique. Therefore, the condition
can only be satisfied by with . Because , we know by Theorem 2.1 that there exists a direction such that . In other words,
which contradicts the hypothesis that is a local minimizer of on (see Remark 2.18).
Overall, these results thus constitute the KKT optimality conditions as stated in Theorem 2.10. But although the foregoing development is straightforward, it is somewhat limited by the regularity-type assumption at the optimal solution. Obtaining more general constraint qualifications (see Remark 2.21) requires that the KKT conditions be derived using an alternative approach, e.g., the approach described earlier in Section 2.6. Still, the conversion to equality constrained problem proves useful in many situations, e.g., for deriving second-order sufficiency conditions for inequality constrained NLP problems.
We are now ready to generalize the necessary and sufficient conditions given in Theorems 2.10, 2.13, 2.15 and 2.16 to general NLP problems.
Theorem 2.17: First- and Second-Order Necessary Conditions
Let , and , be twice continuously differentiable functions on . Consider the problem of minimizing subject to the constraints and . If is a local minimum of the optimization problem and is a regular point of the constraints, then there exist unique vectors and such that
and
for all such that and
Theorem 2.18: Second-Order Sufficient Conditions
Let , and , be twice continuously differentiable functions on . Consider the problem of minimizing subject to the constraints and . If there exists , and satisfying the KKT conditions in Theorem 2.17, and
for all such that
where , then is a strict local minimum.
Likewise, the KKT sufficient conditions given in Theorem 2.11 for convex, inequality constrained problems can be generalized to general convex problems as follows:
Theorem 2.19: KKT Sufficient Conditions for Convex Programs
Let , be convex and differentiable functions. Let also be affine functions. Consider the problem to minimize subject to . If satisfies the KKT conditions of Theorem 2.17 then is a global minimizer for on .
Theorem 2.20: Farkas’ Theorem
Let and . Then, exactly one of the following two statements holds:
System 1.
,
System 2.
.
Proof. See, e.g., [6, Theorem 2.4.5] for a proof. □
Farkas’ Theorem is used extensively in the derivation of optimality conditions of (linear and) nonlinear programming problems. A geometrical interpretation of Farkas’ Theorem is shown in Fig. 1.A.1.. If a1, . . . , am denote the rows of A, then system 2 has a solution if c lies in the convex cone generated by a1, . . . , am; On the other hand, system 1 has a solution if the closed convex cone x : Ax 0 and the open half-space x : cTx ¿ 0 have a nonempty intersection.
Corollary 2.1: Gordan’s Theorem
Let . Then, exactly one of the following two statements holds:
System 1.
,
System 2.
.
Proof. System 1 can be equivalently written as where is a scalar and is a vector of ones. Rewriting this in the form of System 1 in Farkas’ Theorem 2.20, we get and where . The associated System 2 by Theorem 2.20 states that and for some , i.e., , and , which is equivalent to the System 2 of the corollary. □
Lemma 2.4
Let and be two symmetric matrices, such that and on the null space of (i.e., ). Then,
Proof. Assume the contrary. Then,
Consider a subsequence converging to some with . Dividing the last equation by , and taking the limit as , we obtain
On the other hand, being positive semidefinite, we must have
Hence . That is, using the hypothesis, . This contradicts the fact that