Karush-Kuhn-Tucker conditions
Karush-Kuhn-Tucker conditions (KKT conditions) are the first-order necessary conditions for a solution in nonlinear programming to be optimal, provided that a regularity condition — a constraint qualification — is satisfied. Named after Harold W. Kuhn and Albert W. Tucker, who formalized them in 1951, they generalize the method of Lagrange multipliers — which handles equality constraints — to inequality constraints. In the modern formulation, they apply to optimization problems where an objective function is minimized subject to both equality and inequality constraints.
For a problem with objective \(f(x)\), equality constraints \(h_i(x) = 0\), and inequality constraints \(g_j(x) \leq 0\), the KKT conditions require four things: stationarity (the gradient of the Lagrangian vanishes), primal feasibility (all constraints are satisfied), dual feasibility (Lagrange multipliers for inequalities are non-negative), and complementary slackness (a multiplier is non-zero only if its corresponding constraint is active). These conditions, when combined with an appropriate constraint qualification, completely characterize local optima.
From Lagrange to KKT
The Lagrange multiplier method, developed in the 18th century, solves equality-constrained optimization by introducing multipliers that convert constraints into penalties embedded in the objective. The KKT framework extends this to inequality constraints by recognizing that an inequality constraint can be either active (binding the optimum, treated as an equality) or inactive (irrelevant to the optimum, multiplier zero). This is the content of complementary slackness: \(\lambda_j \cdot g_j(x) = 0\) for each inequality constraint. Either the constraint bites (\(g_j = 0\)) and the multiplier prices its tightness (\(\lambda_j > 0\)), or the constraint is loose (\(g_j < 0\)) and the multiplier vanishes (\(\lambda_j = 0\)).
The extension is not merely technical. It transforms the geometry of constrained optimization. The Lagrangian dual problem, constructed from the KKT multipliers, provides lower bounds on the optimal value and, under strong duality, exact solutions. In convex optimization, the KKT conditions become both necessary and sufficient for global optimality — a remarkable sharpening that separates convex from general nonlinear programming. This is why interior point methods and other convex solvers treat the KKT system as the target: solving the KKT equations is equivalent to solving the original problem.
The Shadow Price Interpretation
Each KKT multiplier has an economic and systems-theoretic meaning: it is the marginal rate of change of the optimal objective value with respect to a perturbation in the corresponding constraint. A multiplier of 3.7 means that relaxing the constraint by one unit would improve the objective by 3.7 units. This is not merely a derivative. It is a price — the price of constraint, the value of permission, the cost of boundary.
In mechanism design and social choice, these prices guide resource allocation and equilibrium clearing. In control theory, KKT multipliers reveal which constraints actually limit system performance. The multipliers are the hidden wiring of constrained systems: they tell you not just what the optimum is, but why it is where it is, and what would happen if the rules changed. The duality between primal variables (what to do) and dual variables (what the constraints are worth) is the structural backbone of Lagrangian duality.
Regularity and Its Violations
The KKT conditions are necessary only when a constraint qualification holds — a regularity condition ensuring that the local geometry of the constraints is not pathological. The most common qualification in convex settings is Slater's condition, which requires the existence of a strictly feasible point. When constraints are linear, no additional qualification is needed. But when constraints are nonlinear and degenerate — when gradients of active constraints become linearly dependent at the optimum — the KKT conditions may fail to hold even at a genuine optimum.
This is not a niche pathology. It is the boundary between tractable and treacherous optimization. When constraint qualifications fail, one must retreat to the Fritz John conditions, which always hold but provide weaker structural information. The gap between KKT and Fritz John is precisely the gap between problems we can fully characterize and problems that resist clean characterization. Every constraint qualification is a bet that the geometry of the feasible set is well-behaved near the optimum — a bet that is usually safe in engineered systems and frequently unsafe in natural ones.
The KKT conditions are often taught as a technical checklist — stationarity, feasibility, complementary slackness — but this misses their deeper structure. They are the equilibrium conditions of a constrained system, the point at which the gradient of ambition (the objective) is exactly balanced by the gradient of constraint (the multipliers). Every optimizer is a negotiator between what it wants and what is permitted. The KKT conditions are the treaty that ends that negotiation. And like all treaties, they only hold when both parties have enough power to enforce their terms — that is what constraint qualifications really mean.