1.1.1.2 Properties of Convex Functions
1 Let be convex functions defined on a convex subset . Their summation(1.5) is convex, and if at least of one is a strictly convex function, then their summation is strictly convex.
2 Let a be a positive number and be a (strictly) convex function defined in a convex subset . Then the product is (strictly) convex.
3 Let be a (strictly) convex function defined in , and be an increasing convex function defined on the range of in . Then, the composite function defined in is a (strictly) convex function.
4 Let be convex functions defined on a convex subset . If these functions are bounded from above, their pointwise supremum(1.6) is a convex function on .
5 Let be concave functions defined on a convex subset . If these functions are bounded from below, their pointwise infimum(1.7) is a concave function on .
1.1.2 Optimality Conditions
We introduce the following definitions for the solution of general nonlinear optimization problems:
Definition 1.6 (Local Minimum)
is called a local minimum if there exists ball of radius around , , such that
(1.8)
Definition 1.7 (Global Minimum)
is called a global minimum if
(1.9)
A constrained nonlinear optimization problem, which aims to minimize a real valued function
subject to the inequality constraints and equality constraints is denoted as(1.10)
Problem (1.10) is a nonlinear optimization problem, if and only if, at least one of
is a nonlinear function. We assume that the aforementioned functions are continuous and differentiable.Definition 1.8 (Active Constraints)
An inequality constraint
is called active at a point if . Conversely, is called inactive if .Remark 1.1
If one step of the dual simplex algorithm consists of changing one element of the active set, i.e. let
, then the dual pivot involving the constraint yields .The first‐order constraint qualifications that will be presented in the following text are necessary prerequisites to identify whether a feasible point
is a local optimum of the function .Linear independence constraint qualification: The gradients for all and for all are linearly independent.
Slater constraint qualification: The constraints for all are pseudo‐convex1 at , while the constraints for all are quasi‐convex or quasi‐concave.2 In addition, the gradients are linearly independent and there exists such that and .
1.1.2.1 Karush–Kuhn–Tucker Necessary Optimality Conditions
Let
and be differentiable at a feasible solution , and let have continuous partial derivatives at . In addition, let be the number of active inequality constraints at . Then if one of the aforementioned constraint qualifications hold, there exist Lagrange multipliers such that(1.11)
These conditions are the Karush–Kuhn–Tucker (KKT) Necessary Conditions and they are the basis for the solution of nonlinear optimization problems.
1.1.2.2 Karun–Kush–Tucker First‐Order Sufficient Optimality Conditions
Consider the sets
and . Then, if the following conditions hold:is pseudo‐convex at with respect to all other feasible points x.
for all are quasi‐convex at with respect to all other feasible points x.
for all are quasi‐convex at with respect to all other feasible points x.
for all are quasi‐concave at with respect to all other feasible points