Penalty methods
Contents
Penalty methods#
In contrast to barrier methods, penalty methods solve a sequence of unconstrained optimization problems whose solution is usually infeasible to the original constrained problem. As this penalty is increased, the iterates are forced towards the feasible region.
Consider the equality-constrained problem
where \(h(x)\) is an \(p\)-dimensional vector, whose \(j\)-th component \(h_{j}(x)\) is twice continuously differentiable.
The best-known penalty is the quadratic-loss function
The weight of the penalty is controlled by a positive penalty parameter \(\rho\). The penalty method consists of solving a sequence of unconstrained minimization problems of the form
for an increasing sequence \(\{ \rho^{k} \}\) of positive values tending to infinity.
Penalty methods share many of the properties of barrier methods. Under appropriate conditions, the sequence of penalty function minimizers defines a continuous trajectory. In the latter case, it is possible to get estimates of the Lagrange multipliers at the solution.
Consider the quadratic-loss penalty function
Its minimizer \(x(\rho)\) satisfies
Defining \(\lambda_{j}(\rho) := -\rho \cdot h_{j}(x(\rho))\) one obtains
If \(x(\rho)\) converges to a solution \(x^{*}\) that is a regular point of the constraints, then \(\lambda(\rho)\) converges to the Lagrange multiplier \(\lambda^{*}\) associated with \(x^{*}\).
Penalty functions suffer from the same problems of ill-conditioning as barrier functions. As the penalty parameter increases, the condition number of the Hessian matrix of \(\pi(x(\rho), \rho)\) increases, tending to \(\infty\) as \(\rho \to \infty\). Therefore the unconstrained minimization problems can become increasingly difficult to solve.
PM: Example 1#
Consider the problem
Then the penalty method solves a sequence of unconstrained minimization problems
for increasing values \(\rho^{k}\). The necessary conditions for optimality are
For \(\rho > 1/4\) this yields the solution
which is a local as well as a global minimizer. (The unconstrained problem has no minimum if \(\rho \leq 1/4\).) Note that \(x(\rho)\) is infeasible to the original constrained problem, since
At any solution \(x(\rho)\) we can define a Lagrange multiplier estimate
As \(\rho\) tends to \(\infty\) one obtains
and indeed \(x^{*} = (2, 1)^{T}\) is the minimizer for the constrained problem. Further,
and indeed \(\lambda^{*} = -1\) is the Lagrange multiplier at \(x^{*}\).
The ill-conditioning of the penalty function is demonstrated by the Hessian matrix at \(x(\rho)\):
It can be shown that its condition number is approximately \(25 \rho / 4\). When \(\rho\) is large, the Hessian matrix is ill-conditioned.
It is also possible to apply penalty methods to problems with inequality constraints. Consider for example the problem
For example, the quadratic-loss penalty in this case is
This function has continuous first derivatives
but its second derivatives can be discontinuous at points where some constraint \(g_{i}\) is satisfied exactly. Hence, a careful implementation is necessary. One cannot safely use Newton’s method to minimize the function. For this reason, straightforward penalty methods have not been widely used for solving general inequality-constrained problems.
For a convergence discussion of penalty methods, see [GNS08] (chapter 16.2.3).