98 Classical Optimization Techniques
2.5.1 Kuhn–Tucker Conditions
As shown above, the conditions to be satisfied at a constrained minimum point,X∗, of
the problem stated in Eq. (2.58) can be expressed as
∂f
∂xi+
∑
j∈J 1λj∂gj
∂xi= 0 , i= 1 , 2 ,... , n (2.73)λj> 0 , j∈J 1 (2.74)These are calledKuhn–Tucker conditionsafter the mathematicians who derived them
as the necessary conditions to be satisfied at a relative minimum off(X) [2.8]. These
conditions are, in general, not sufficient to ensure a relative minimum. However, there is
a class of problems, calledconvex programming problems,†for which the Kuhn–Tucker
conditions are necessary and sufficient for a global minimum.
If the set of active constraints is not known, the Kuhn–Tucker conditions can be
stated as follows:
∂f
∂xi+
∑mj= 1λj∂gj
∂xi= 0 , i= 1 , 2 ,... , nλjgj= 0 ,‡ j = 1 , 2 ,... , m
gj≤ 0 , j= 1 , 2 ,... , mλj≥ 0 , j= 1 , 2 ,... , m(2.75)
Note that if the problem is one of maximization or if the constraints are of the type
gj≥ , the 0 λjhave to be nonpositive in Eqs. (2.75). On the other hand, if theproblem is
one of maximization with constraints in the formgj≥ , the 0 λjhave to be nonnegative
in Eqs. (2.75).2.5.2 Constraint Qualification
When the optimization problem is stated asMinimizef (X)subject to
gj(X)≤ 0 , j= 1 , 2 ,... , mhk(X)= 0 k= 1 , 2 ,... , p(2.76)
the Kuhn–Tucker conditions become∇f+∑mj= 1λj∇gj−∑pk= 1βk∇hk= 0λjgj= 0 , j= 1 , 2 ,... , m†See Sections 2.6 and 7.14 for a detailed discussion of convex programming problems.
‡This condition is the same as Eq. (2.64).