In economics, most optimization problems involve scarcity. We want to maximize something (utility, profit, output), but we face constraints (budget, technology, time, policy).
This chapter develops the tools to solve such problems.
Constrained Optimization and Second-Order Conditions
Using the substitution method, solve each constrained optimization problem. Determine the optimal values and use the second-order condition to verify whether you have a maximum or a minimum.
(i)f(x,y)=x2+2xy+4y2 subject to x+y=8
(ii)f(x,y)=3x2+y2−2xy subject to x+y=1
(iii)f(x,y)=10x+40y subject to x1/2y1/2=2
(iv)f(x,y)=lnx+2lny subject to y=16−4x
(v)f(x,y)=2x1/2y1/2 subject to x+y=1
(vi)f(x,y)=2x2−4xy+9y2−8y+36 subject to x+y=12
(vii)f(x,y,z)=2x2+5xy−y2−3xz subject to x+y+z=14
Answers
(i)(x,y)=(8,0), minimum (ii)(x,y)=(31,32), minimum (iii)(x,y)=(4,1), minimum (iv)(x,y)=(34,332), maximum (v)(x,y)=(21,21), maximum (vi) [Solution omitted in original, usually solved as (x,y)=(8,4) minimum] (vii)(x,y,z)=(1,4,9), a saddle point
An alternative to the substitution method is the Lagrange multiplier method.
Given an optimization problem say, to maximize an objective function
f(x1,x2)
subject to a constraint:
g(x1,x2)=c.
Begin by defining the Lagrangian function:
L(x1,x2,λ)=f(x1,x2)−λ(g(x1,x2)−c).
where λ is known as the Lagraingain multiplier.
Then the First-Order Conditions (F.O.C): simply get the first partials of the Lagrangian function
∂x1∂L(x1,x2,λ)=0,∂x2∂L(x1,x2,λ)=0 and ∂λ∂L(x1,x2,λ)=0
And solve for the critical values of x1∗ and x2∗.
Note that the last condition simply reproduces the constraint.
Let’s continue with the example above: We wish to maximize equation (1) subject to (2). To find the first-order conditions (FOCs) for this specific Lagrangian, we take the partial derivative of L with respect to each of the three arguments (x1, x2, and λ) and set them equal to zero.
which is positive, hence ∣Hˉ∣ is negative definite and we have met the sufficient condition for a maximum.
That is, evaluate the determinant at the stationary point we found earlier (x1=4, x2=4), the resulting matrix is:
∣Hˉ∣=∣Hˉ2∣=∣∣0414−1/16010−1/64∣∣=5/16
which is positive, hence ∣Hˉ∣ is negative definite and we have met the sufficient condition for a maximum. In other words, for a bivariate case with one constraint, a Lagrangian function is negative (positive) definite at a stationary point if the determinant of its bordered Hessian is positive (negative) when evaluated at that point. In this case we have sufficient condition for the stationary point identified by the Lagrangian method to be a maximum (minimum).
More generally, for the case of n arguments and m constraints, we have a sufficient condition for a maximum if the determinant of the bordered Hessian (∣Hˉn+m∣) has the same sign as (−1)n, and the largest n−m leading principal minors alternate in sign. In this case, the quadratic form is negative definite subject to the constraint.
Alternatively, if the largest n−m leading principal minors of the bordered Hessian (including the determinant of the bordered Hessian itself) all have the same sign as (−1)m, then the quadratic form is positive definite subject to the constraint, and the stationary point represents a minimum.
As a rule, we essentially need to evaluate the determinants of the largest n−m leading principal minors of the bordered Hessian.
Also it is important to note is that the Lagrangian method generates a variable not obtained with the
substitution method i.e., the Lagrangian multiplier λ, which represents the effect of a small
change in the constraint on the optimal value of the objective function.
Although we have seen just a simple case, optimization of a bivariate function subject to one constraint, the
Lagrangian method is powerful enough to deal with multivariate functions and multiple constraints
Try these ^^
Use the sufficient condition to determine whether the stationary points identified in (i), (ii), and (iv) of previous questions above represent a maximum, minimum, or saddle point.
Answers
(i)
Hˉ=⎣⎡022211216⎦⎤
which has the determinant -24. Since m=1, the sign of a minimum must be (−1)1=−1. Thus, the stationary point is a minimum.
(ii)
Hˉ=⎣⎡01/2311/2100308010016⎦⎤
The determinant of the leading principal minor ∣Hˉ3∣ is -11 and the full determinant ∣Hˉ4∣ is -184. Since m=1, all minors must have the sign (−1)1=−1 for a minimum. Hence, we have a minimum.
(iv)
Hˉ=⎣⎡0444641641−λ4641−λ641⎦⎤
With λ=1/32, the determinant is 1. Since n=2 and m=1, the sign for a maximum must be (−1)n=(−1)2=+1. Thus, we have a maximum.
The bordered Hessian Hˉ for n=3 variables and m=1 constraint is a 4×4 matrix. The border consists of the first derivatives of the constraint g(x,y,z)=x+y+z−14, which are gx=1,gy=1,gz=1. Evaluating the second-order partial derivatives, we obtain:
In general, we can write the Lagrangian function for a constrained optimization problem with an objective function with n variables subject to m equality constraints as follows:
As an exercise, verify that we have a maximum by evaluating the bordered Hessian. Note that in this case m−n=3−1, so we need to evaluate the determinants of the largest two principal minors, ∣Hˉ∣=∣Hˉ3∣ and ∣Hˉ2∣.
Perhaps we should explain this before moving on? We need to eliminate the Lagrange multipliers (λ1 and λ2) using the first three equations.
Step 1: Isolate λ1:
From the second FOC (∂V∂L):
3V1=2λ1⟹λ1=3V2
Step 2: Isolate λ2:
Subtract the third FOC from the first FOC to cancel out λ2:
(3S1−4λ1−λ2)−(3J1−12λ1−λ2)=0
3S1−3J1−4λ1+12λ1=0
Step 3: Combine and Solve for V:
Substitute our λ1 value (3V2) into that result:
3S1−3J1=41(3V2)−121(3V2)
3SJJ−S=12V2−36V2
3SJJ−S=36V6−36V2=36V4=9V1
Now, rearrange to solve for V:
9V(J−S)=3SJ
V=9(J−S)3SJ=3(J−S)SJ
We proceed by simplify the first constraint by substituting V=3(J−S)SJ into the first constraint:
4S+21(3(J−S)SJ)+12J=6
4S+6(J−S)SJ+12J=6
To clear the denominators, multiply the entire equation by 12(J−S):
3S(J−S)+2SJ+J(J−S)=72(J−S)
3SJ−3S2+2SJ+J2−SJ=72J−72S
4SJ−3S2+J2=72J−72S
Next, from the second constraint, we know J=24−S. We substitute this into our simplified equation:
4S(24−S)−3S2+(24−S)2=72(24−S)−72S
96S−4S2−3S2+(576−48S+S2)=1728−72S−72S
And grouping the S terms:
−6S2+48S+576=1728−144S
−6S2+192S−1152=0
Divide by -6 to make it a standard quadratic:
S2−32S+192=0
Using the quadratic formula or factoring (S−24)(S−8)=0 gives S=24. This would imply J=0, which is impossible given the lnJ term in our objective function.
S=8: This is our valid solution.
The final Values S=8, J=24−8=16 and V=3(J−S)SJ=3(16−8)8×16=24128=316 or 531
Hence, the optimal point for this system is:
S∗=8,V∗=531,J∗=16.
Which gives an utility of 2.17, which is lower than 2.21 (without the second constraint).
Now the second order conditions, or bordered Hessian.
which matches the sign of (−1)n where in this case n=3, i.e., both are negative.
Hence the stationary point S∗=8, V∗=531, and J∗=16 represent a maximum.
But don’t we have to evaluete the determinant other principal minors of the bordered Hessian as well? No, because in this case, we have to evaluate only n−m=3−2=1 largest Hessians.
In general, for the case of n arguments and m constraints, we have a sufficient condition for a maximum if the largest n−m leading principal minors of the bordered Hessian alternate in sign, with the sign of the full determinant (∣Hˉ∣) being the same as (−1)n.
Alternatively, we have a sufficient condition for a minimum if the largest n−m leading principal minors of the bordered Hessian all have the same sign, which must be the same as the sign of (−1)m. If the minors do not follow either of these patterns, the second-order condition for a local extremum is not satisfied.
Try these ^^
Lagrange Method with Multiple Constraints
Use the Lagrange method to determine the optimal values for the choice variables in the following problem involving multiple constraints.
Problem:
Find the stationary point for the function:
z=x2+2xy+w2
subject to the constraints:
2x+y+3w=24andx+w=8
Answer
(w,x,y)=(2,2,6)
Worked Solution:
Set up the Lagrangian:
We introduce two Lagrange multipliers, λ1 and λ2, for the two constraints:
L=x2+2xy+w2−λ1(2x+y+3w−24)−λ2(x+w−8)
First-Order Conditions (F.O.C.s):
∂x∂L=2x+2y−2λ1−λ2=0
∂y∂L=2x−λ1=0⟹λ1=2x
∂w∂L=2w−3λ1−λ2=0
∂λ1∂L=2x+y+3w−24=0
∂λ2∂L=x+w−8=0
Solving the System:
From the second F.O.C., we have λ1=2x.
Substitute λ1 into the third F.O.C.: 2w−3(2x)−λ2=0⟹λ2=2w−6x.
Substitute both λ1 and λ2 into the first F.O.C.:
2x+2y−2(2x)−(2w−6x)=0
2x+2y−4x−2w+6x=0⟹4x+2y−2w=0⟹2x+y−w=0
Now use the constraints: From x+w=8, we have w=8−x.
Substitute w into our derived equation: 2x+y−(8−x)=0⟹3x+y=8⟹y=8−3x.
Substitute w=8−x and y=8−3x into the first constraint:
So far we have considered constraints which are strictly equalities. We now consider the case when
constraints take the form of inequalities rather than equalities. This involves a slight modification to
the Lagrangian method, which we present below.
The solutions to this problem satisfies the following conditions:
Stationarity
∂xk∂L=0, for k=1,…,n
Primal feasibility
gi(x)≤ci,, for i=1,…,m
Dual feasibility
λi≥0, for i=1,…,m
Complementary slackness
λi(ci−gi(x))=0,, for i=1,…,m.
The above are known as the Kuhn-Tucker conditions for an optimization problem with inequality constraints.
The last two Kuhn-Ticker conditions are known as complementary slackness conditions (meaning that both
optimal points and first-partials evaluated at the optimal points cannot simultaneously both be zero).
We proceed by considering solutions that arise in four cases corresponding to the four possible ways in which the complementary slackness conditions may be met. These cases are
In this case the condition ∂L/∂x1=0 shows that x1=1, and the condition ∂L/∂x2=0 shows that x2=0. The pair that x1=1 and that x2=0 satisfies the constraint
2x12+x22≤89
as well as the constraint x2 which s nonnegative. Hence:
Here the F.O.C ∂L/∂x2=0 requires that λ1=1, which satisfies the condition that λ1≥0. This in turn requires that x1=1/2, and satisfy the condition ∂L/∂x1=0. The complementary slack condition requires that
89−2(1/2)2−x22=0
which is satisfied for x2=1 or x2=−1. But the constraint that x2 is nonnegative allows us to consider only x2=1. Hence:
Here the condition ∂L/∂x1=0 gives x1=1. The condition λ2x2=0 requires that x2=0. Therefore this solution is identical to the one found when λ1=0 and λ2=0. Same as Case 1:
Lastly, we consider the case where λ1=0 and λ2=0. The condition λ2x2=0 requires that x2=0. But if x2=0, the condition ∂L/∂x2=0 requires that λ2=0. Therefore there is no feasible solution in which λ1=0 and λ2=0.
The optimal solution is x=1,y=0, which implies that λ2=0 (via complementary slackness). At this point, the objective function reaches its maximum value of h(1,0)=2.
(ii) The Lagrangian function is:
L=2a2+b2−λ1(2a+b−9)−λ2(a2+b2−16)
The solution that provides the maximum value must satisfy: