140 Donald Wittman APPENDIX A1 Mathematical Background y x
48
140 Donald Wittman APPENDIX A1 Mathematical Background y x . Budget Constraint - a line Greater Utility Constrained Calculus: Example: Maximizing utility subject to a budget constraint. The triangle bounded by the x and y axes and the budget constraint is the feasible set. Choose the highest iso-utility curve (indifference curve) that is feasible. Note that the constraint holds with equality. Feasible set Higher Utility y x Linear Programming: The constraints and iso-curves are all linear. The constraints do not have to hold with equality. That is, some of the constraints are not-binding (e.g., the line going north-east). The solution will be at one of the points of intersection. y x Higher Utility Feasible Set Non-Linear Programming: The solution may be anywhere, including the interior of the feasible set if the point of highest utility is interior to the feasible set. This would be the case in the drawing to the left if the feasible set were expanded to the right so that it covered the innermost circle.
140 Donald Wittman APPENDIX A1 Mathematical Background y x
07.math.appendixy
x
x and y axes and the budget constraint
is the feasible set. Choose the highest
iso-utility curve (indifference curve)
holds with equality.
linear. The constraints do not have to
hold with equality. That is, some of the
constraints are not-binding (e.g., the
line going north-east). The solution will
be at one of the points of intersection.
y
x
if the point of highest utility is interior
to the feasible set. This would be the
case in the drawing to the left if the
feasible set were expanded to the right
so that it covered the innermost circle.
141
Closed Convex Sets Almost all optimization problems make extensive
use of convex sets. This statement holds true
for maximization via calculus and linear programming as well as
non-linear optimization
problems. Today I will show that closed convex sets and hyperplanes
are the bases behind the
optimization techniques that we will cover. These concepts are
meant to give insight into the
more mechanical methods that we employ and to show that we are
fundamentally using the
identical method in all that we have covered in previous courses.
The proofs regarding convex
sets are thus not meant for memorization but to provide basic
understanding.
We first start off with a number of definitions: Convex Set:
Geometric definition--a set is convex if and only if a (straight)
line connecting any two points in
the set is also in the set.
Convex sets:
Non Convex sets:
Convex Set: Algebraic definition--Let x,y be two vectors in N space
in the set S. E.g.,
x = (x1, x2) y = (y1, y2) . S is convex if and only if Px + [1−P] y
! S for all P such that 0 ≤ P ≤ 1.
142
2
x = x , x 1 2
Note that a convex set is not the same thing as a convex
function.
Boundary point: Intuitively boundary points are points, which are
on the edge of a set. More
formally, a point is a boundary point if every neighborhood (ball)
around the point contains
points not in the set and in the set. They may be members of the
set; e.g., in the set 0 ≤ x ≤ 1, the
numbers 0 and 1 are boundary points and members of the set. In the
set 0 < x < 1, the numbers 0
and 1 are boundary points and not members of the set.
Closed set: a closed set is a set that includes all its boundary
points. The following is a closed
set: 0 ≤ x ≤ 1. In economics we almost always consider closed
convex sets. The reason can be
illustrated by considering the opposite. Maximize the amount of
gold (G) you get if G < 1.
There is no maximum!
Preference set: The set of all points, which are indifferent or
preferred to a given point.
Economics majors are acquainted with indifference curves. A
preference set is the set of all
points on the indifference curve plus the all of the points
strictly preferred to the indifference
curve. In the typical drawing of an indifference curve the
preference set would include the
indifference curve plus all of the points upwards and to the right
of the indifference curve.
We can view the ordinary two-dimensional graph of indifference
curves as a two-dimensional
topographical map showing height in terms of contour lines. Here
height is utility and the
contour lines are indifference curves. Consider the following
diagram. There is a mountain
143
(representing utility) sticking out from the paper, the higher the
mountain the higher the utility.
Assume that it is an ice cream cone cut in half lengthwise, filled
with ice cream and put upside
down on the paper. The cone part is the indifference curve formed
by a horizontal slice parallel
to the paper; the cone plus the ice cream represents all points
such that utility is either equal to or
greater. We assume the set of such points is convex.
y
x
Indifference curves
Theorem 1. Sufficient conditions for a local optimum to be a global
optimum. If the feasible set F is a closed and convex set and if
the objective function is a continuous
function on that set and the set of points (such that the function
is indifferent or preferred) create a convex set, then the local
optimum is a global optimum.
This is a very important result. When these conditions hold, we
know that that myopic
optimization will end up at a global optimum. If either the
feasible set or the preference set is not
convex, then the local optimum need not be a global optimum.
Feasible set is not convex. 2nd peak is a local optimal but not a
global optimum
(See following page)
144
145
Preference set is not convex. Right-most tangent is not a global
optimal.
146
We next turn our attention toward hyperplanes. Hyperplane: a line
in N-space dividing the space in two.
In two-space a hyperplane is a line. E.g., y = a + bx or a = y - bx
or in more general
notation, A = c1x1 + c2x2. In three-space we have a plane: y = A +
BX + CZ, or in more general
notation, …a = c1x1 + c2x2 + c3x3. In four or more space we have a
hyperplane. Note that y ≥ a
+ Bx or a ≥ c1x1 + c2x2 creates a convex halfspace.
HYPERPANES CREATE CONVEX HALFSPACES
x 1
Supporting hyperplane has one or more points in common with a
(closed) convex set but no
interior points. The set lies to one side of the hyperplane.
SUPPORTING HYPERPLANES
The following diagrams illustrate other cases of supporting
hyperplanes.
Theorem 2. Given z a boundary point of a closed convex set, there
is at least one
supporting hyperplane at z. Theorem 3: If 2 convex sets intersect
but without interior points, there is a supporting
hyperplane that separates them.
Neoclassical optimization:
hyperplane separating the two convex sets
Neoclassical optimization (usually only consider edge) Max U (x, y)
+ (g(x) - 5) ¬
In the following figure, the shaded area is the feasible set and
the curved line is the indifference
curve, which characterizes the preference set. The straight line is
the hyperplane. We usually
give it a less technical term – the budget set or price line. Given
these prices, the point of
intersection maximizes utility.
149
There are other results that are of use. The following two theorems
state that the intersection of
closed convex sets is closed and convex.
Theorem 4. The intersection of two convex sets is convex.
Proof: If x ! (S ∩ T) and y ! (S ∩ T), then Z = Px + [1 - P] y ! S
because S convex Z = Px + [1 - P] y ! T because T convex Therefore
Z ! (S ∩ T)
Theorem 5. The intersection of two closed sets is closed.
150
Donald Wittman APPENDIX A2 Concave and Quasiconcave Functions A.
CONCAVE FUNCTIONS Concave functions
x
f(x)
Three definitions of concavity: (1) Concave Function: The line
connecting any two points of the function lies on the function
or
below.
(2) Concave Function: Algebraic definition: Pf x( ) + 1 ! P[ ] f y(
) " f Px + 1 ! P[ ]y( ). Note that x
and y may be vectors in n-space
(3) Concave Function: Tangency definition: The tangent is always
outside or on the function. The sum of two concave functions is
also concave.
151
In order to determine whether a function is concave, we generally
would prefer a method that did
not rely on drawing a picture of the function. Hence, we have the
second derivative test.
If ! ! f x( ) = d2 f dx 2
" 0 , for all x, then the function is concave. This is a necessary
and sufficient
condition for concavity.
Example A: f(x) = -x4 is concave since f''(x) = -12x2! 0 for all
x.
Example B: f(x) = - x3 is not concave since f''(x) = - 6x and - 6x
> 0 for x < 0.
Example C: x ≥ 0 and f(x) = − x3 is concave since f''(x) = − 6x ≤ 0
for all x ≥ 0.
Notice that the test for concavity has a great similarity to the
test for a maximum. To test for a
maximum, you find out if the second derivative is less than zero
when the first derivative equals
zero. To speak imprecisely, the test for a maximum discovers
whether the function is locally
concave. In contrast, the test for concavity requires the second
derivative to be non-positive
everywhere.
! f x( ) =12 " 6x = 0 x = 2
! ! f x( ) = "6 < 0 Therefore at x = 2, f(x) is a maximum and
not a minimum.
tangent
tangent above function (locally) where f '(x) = 0. Going down
hill.
The analogous test for concavity: ! ! f x( ) = "6 < 0 .
Therefore f is concave.
152
An Important Property of Concavity: If f(x) is concave, then the
set of x such that f(x) ≥ k is convex for all k.
f(x)
153
B. MULTI-DIMENSIONAL CONCAVE FUNCTIONS Next, we would like to
expand our intuition by visualizing concave functions and convex
sets in
two dimensions. If f x1, x2( ) = z , then f x1, x2( ) is concave if
any vertical plane looks like the
single dimension picture (see the slice in the right hand side of
the following picture).
f(x ,x )1 2
slice
What happens if we take a horizontal slice? We then we have an
iso-function. If f is utility, iso-
utility is an indifference curve. Looking at x1, x2 :
x 2
x 1
from the figure to the above and left.
Notice that the set of x1, x2 such that
f x1, x2( ) ! k is a solid convex space.
If f is concave, then the set x1, x2 such that f x1, x2( ) ! k is
convex for all k. The reverse relationship need not be true (a
convex set does not imply a concave function). Hessian of 2nd order
derivatives test—sufficient conditions. Let f i be the partial of f
with respect to the partial of i. f11 f12 f21 f 22
= H for all x1, x2 H1 < 0 H2 > 0
! f concave where H1 = f1 1 and H2 = H
Note well this is only a sufficient condition for concavity. The
necessary and sufficient
conditions have loose inequalities and consider permutations (this
will be covered later).
154
Convex Function: The line connecting any
two points on a convex function does not lie
below the function.
f(x)
x
Convex Function, tangency definition: the function never lies below
any tangent to the function. Convex function (a third definition):
A function f is convex if and only if -f is concave. Hessian test:
! ! f x( ) " 0 for all x is a necessary and sufficient condition
for f being convex. Note that the set of all x such that
f x( ) ! K is convex.
f(x) k
Students often get mixed up between concave and convex functions
and convex sets. Think of a
concave function as creating a cave. This will help distinguish a
concave function from a convex
function. While there is a convex set there is no such thing as a
concave set. The set of x such
that f(x) ≥ k is convex when f(x) is a concave function; the set of
x such that f(x) ≤ k is convex
when f(x) is a convex function.
155
D. QUASICONCAVITY Intuitive concept for one dimension: the function
has only one peak. Quasiconcavity: Algebraic definition:
f px + 1! P[ ]y( ) " f x( ) and/or f (y)
f(x)
f x( ) ! k
f(x)
Any horizontal cut creates a convex set of indifferent or higher
points. Another lecture covers the hessian test for
quasiconcavity.
156
Note well that the sum of two quasiconcave functions need not be
quasi-concave. E.g., try x3
and -x2. For x ≥ 0 both are quasiconcave.
x3 -x2
Adding them together:
x 3 ! x 2 = 0 at x= 0 0 − 0 = 0 at x=1/2 1/8 −1/4 = −1/8 at x =1 1
− 1 = 0
A quasiconcave function has to have one peak no matter what way you
slice it. These two boxes below look like there is only one peak,
but if you go from A to B you will drop
down to C.
157
Donald Wittman APPENDIX A3 Semidefinite determinants Students are
acquainted with second order conditions. In the single variable
case, if the first
order conditions are equal to zero (f'(x) = 0), then f'' < 0 is
sufficient for a local maximum.
These results can be generalized to the n-variable case by looking
at the Hessian matrix of
second derivatives when the firsts order conditions are satisfied.
Now this approach to
maximization is very limited. It only tests for local maxima; there
may be a much higher peak
elsewhere. Furthermore, it provides only a sufficient condition.
For example, -x4 is a maximum
at x = 0; yet its second derivative is −12x2 = 0 at x = 0.
Therefore the sufficiency test is not
applicable.
In programming we are interested in global optima. We do not look
at just a point (where f' = 0),
but the function as a whole. We test whether f is a concave
function. It is if f'' ≤ 0 for all relevant
values of x (say for x ≥ 0). If there is more than one variable, we
then deal with the Hessian of
second derivatives (this will be explained later). Thus global
optimization in non-linear
programming uses related ideas to the local optimization analysis
(which in crude but incorrect
terms finds local concavity) typically found in courses in
intermediate economics, but there are
three important differences: (1) Conditions on the second
derivative of f apply to all values of x
not just where the first order conditions are satisfied. (2)
Inequalities need not be strict (zero
values are allowed). (3) Permutations of the Hessian of second
order partials are considered.
We will start with strict concavity and strict convexity, which are
easier to handle. We know
from the single variable case that if f'' < 0, then f(x) is a
strictly concave function. The two-
variable test is a bit more complicated as one has to make sure
that the function is concave not
only in one direction but in any direction. To start with, suppose
that f is a function of three
variables, f(x1, x2, x3). Letting fi stand for the partial of f
with respect to xi, we have the following
matrix of second derivatives:
!
"
# # #
$
%
& & & .
Note that the matrix is symmetric as fij = fji by Young’s theorem.
We will now look at the
determinants (denoted by ||) formed from this matrix (denoted by
[]).
If the following holds, then f is strictly concave:
f11 < 0, f11 f12
f21 f22 f23
f31 f32 f33
f11 f21 f31
f12 f22 f32
= f11 f22 f33 + f12 f23 f31 + f13 f21 f32 ! f31 f22 f13 ! f32 f23
f11 ! f33 f21 f12
If the determinants derived from the matrix of second derivatives
of f are negative definite for all relevant values of x, then f is
strictly concave. This holds for any number of variables,
not just 3.
Why not only if? f(x) = –x4 is strictly concave, but its second
derivative = 0 at x = 0.
The matrix is known as being positive definite if the determinants
of the successive principal
minors are all positive for all relevant values of x.
159
If the determinants derived from the matrix of second derivatives
of f are positive definite
for all relevant values of x, then f is strictly convex.
Things get a bit more complicated when the above inequalities are
no longer strict.
Let
H k be a k-by-k determinant formed by any permutation from the set
of N variables along the
diagonal of H, an N-by-N matrix. Note that the bar is to remind us
that we are permuting.
H =
!
"
# # # #
$
%
& & & &
H 2 = f11 f12
f12 f11
= f11 f22 - f21 f12
As can be seen, the second expression is redundant. This is for two
reasons: 1) the second
determinant is equivalent to the first because we have interchanged
a row and then a column,
which does not alter the value of the determinant. 2) In the second
place, we will only be
considering second partials so that the matrices are symmetric.
Therefore this alteration will
have no effect.
!
"
# # #
$
%
& & &
H 2 = f11 f12
f23 f22
Note that the second determinant is redundant of the first, the
fourth redundant of the third and
the sixth redundant of the fifth.
Finally,
H 3 = |H|.
Note that the diagonals in the sub-determinants are always fii from
the original diagonal and that
the off diagonal elements are matched to the diagonal. The same
holds true for a bordered
hessian, which we will get to later (the borders remain borders,
diagonals remain diagonals and
other elements match).
are inappropriate.
The matrix is negative semidefinite if for all relevant values of
x, all of the permutations of the
determinants of the successive principal minors alternate in sign
starting with a minus. That is, if
for all the permutations,
H 1 ! 0;H 2 " 0;H 3 ! 0, then H is a negative semidefinite
matrix.
A matrix of second derivatives of f is negative semidefinite if and
only if f is concave. If for all the permutations
H 1 ! 0;H 2 ! 0;H 3 ! 0, then H is a positive semidefinite
matrix.
161
A matrix of second derivatives of f is positive semidefinite if and
only if f is convex. Just for fun (if you call this fun), I have
written this out for the 4-by-4 case, where aij = fij.
If H = a11 a12
a21 a22 a13 a14
H 2 = a11 a12
H3 = a11 a12 a13 a21 a22 a23 a31 a32 a33
and a11 a12 a14 a21 a22 a24 a41 a42 a44
and a11 a13 a14 a31 a33 a34 a41 a43 a44
and a22 a23 a24 a32 a33 a34 a42 a43 a44
H 4 =|H | A matrix is positive semidefinite if for all values the
derived determinants H1 ! 0, H2 ! 0, H3 ! 0 A matrix is negative
semidefinite if for all values the derived determinants H1 ! 0, H2
" 0, H3 ! 0
162
We next consider the tests for quasiconcavity and quasiconvexity.
Again, we will use the
concepts of negative (or positive) definite series of determinants
and negative (or positive)
semidefinite series of determinants, but this time the determinants
will be bordered with the first
derivatives and therefore the signs get reversed.
It is easy to get confused about signs. Here is the way that I
remember. For concavity, we have a
negative semidefinite expansion 1!1" 0;2 ! 2 # 0;3! 3" 0; etc. In
testing for quasiconcavity, we
reverse signs because we have a border: hence, 1!1" 0;2 ! 2 # 0;3!
3" 0; etc. As we will see,
with a border, the following will always be true1!1= 0;2 ! 2 " 0.
So, we will only have to pay
attention starting with the 3! 3determinant. For convexity, we have
a positive semidefinite
expansion 1!1" 0;2 ! 2 " 0;3! 3" 0; etc. In testing for
quasiconvexity, we again reverse signs
because we have a border: 1!1" 0;2 ! 2 " 0;3! 3" 0; etc. Again it
is true that with a border, the
following will always be true1!1= 0;2 ! 2 " 0. So, we will only
have to pay attention starting
with the 3! 3determinant.
We start with the two-variable case. Notice right away that the
border is composed of first
derivatives. Notice also that if we consider successive
determinants, the first determinant will
always equal 0 and the second will always be negative or 0. So we
start with the third. If the 3x3
determinant is strictly positive (and the 4x4 determinant is
strictly negative when we have 3
variables, and so forth), we have a negative definite symmetric
matrix.
0 f1 f2 f1 f11 f12 f2 f21 f22
!
"
# # #
$
%
& & &
If the bordered hessian of f is negative definite (starting with a
positive for the 1x1), then f is strictly quasiconcave.
If the bordered hessian of −f is negative definite (starting with a
positive for the 1x1), then f
is strictly quasiconvex.
163
Note that employing the minus sign in front of f, means the value
of the determinant is opposite
of the value of the determinant for f when there are an odd number
of columns and the same
value of the determinant as for f when there are an even number of
columns. Hence if the
expansion is always negative, then f is strictly quasiconvex.
When the value of one or more determinants is 0, we have to go back
to permutations and semi-
definiteness, with loose inequalities.
The bordered hessian of f is negative semidefinite (starting with a
positive for the 1x1), if and only if f is quasiconcave.
The bordered hessian of −f is negative semidefinite (starting with
a positive for the 1x1), if
and only if f is quasiconvex.
Example 1. f(x,y) = x + y2 where x, y ≥ 0. fx = 1; fy = 2y
fxx = 0; fyy = 2; fxy = 0
H = 0 0 0 2
H1 = 0, 2 ! 0; H2 = 0 Hence, we have a positive semidefinite
determinant and f is convex. This example illustrates
why we need to do permutations. If we did not do the second
permutation, we could have said
that f is negative semidefinite, but this would be incorrect.
164
Example 2. f(x,y) = x2 + y2 where x, y ≥ 0. fx = 2x; fy = 2y
fxx = 2; fyy = 2; fxy = 0
H = 2 0 0 2
H1 = 2 > 0; H2 = 4 > 0 Therefore we have a positive definite
matrix and a strictly convex function. We know that a convex
function is quasiconvex, as well. But it is useful to demonstrate
this by
looking at the bordered hessian (this is also a good way to check
about the appropriate signing of
a bordered hessian if you forget). So let us look at the bordered
hessian. We start with the 3x3
border as the 2x2 determinants are never positive when we have a
bordered hessian and the 1x1
is always 0.
H = 0 f1 f2 f1 f11 f12 f2 f21 f22
= 0 2x 2y 2x 2 0 2y 0 2
0 2x 2y
2x 2 0
= !8y2 ! 8x2 " 0 . So it is quasiconvex.
I have actually given you a slightly different algorithm. I said
that if you multiply f by minus 1,
and the –f function is quasiconcave, then f is convex. So let us
multiply by –1.
H = 0 !2x !2y
0 !2x !2y
!2x !2 0
= 8y2 + 8x2 ! 0. So –f is quasiconcave and therefore f is
quasiconvex.
165
Example 3. f(x,y) = x +2y ! 3xy where x, y ≥ 1 fx = 1 ! 3y; fy = 2
! 3x
fxx = 0; fyy = 0; fxy = !3
H = 0 !3 !3 0
H1 = 0,H1 = 0; H2 = !9 < 0
Therefore f is neither concave nor convex. Note that we have two
values for H1 as we have to
deal with each permutations, and that we have to deal with
permutations because we have zeros
and semidefiniteness. This time we need to go to the bordered
hessian because we could not
establish concavity or convexity.
H = 0 1 ! 3y 2 ! 3x
1 ! 3y 0 !3 2 ! 3x !3 0
0 1 ! 3y 2 ! 3x
1 ! 3y 0 !3
= !6(1 ! 3y)(2 ! 3x) < 0 because x, y ≥ 1. So, quasiconconvex.
Note well that if
x, y ≥ 0 , f is neither quasiconcave nor quasiconvex because y <
1/3 and x > 2/3 would generate a
positive number while y > 1/3 and x > 2/3 would generate a
negative number. Concavity and
convexity are global conditions not local.
Example 4: f(x,y) = x3y x ≥ 0, y ≥ 0 fx = 3x2y fxx = 6xy fy = x3
fxy = 3x2 fyy = 0
6xy 3x2
3x2 0
166
Therefore f is neither concave or convex. Let us see if the
function is quasiconcave or quasiconvex by looking at the bordered
hessian.
0 3x2y x3
3x2y 6xy 3x2
x3 3x2 0
0
3x2y
x3
3x2y
6xy
3x2 = 9x7y + 9x7y - 6x7y = 12x7y ≥ 0 for x, y ≥ 0 Therefore
negative semidefinite and therefore quasiconcave. Note that when we
have a bordered hessian, we do not pay attention to
H 1 as it is always 0 (the
border remains the same) and
H 2 as all 2x2 are negative (here too, remember that the
border
remains with the 0 in the upper left-hand corner). Note that in the
case we were saved from doing
permutations as we only had the 3x3 matrix to worry about.
Note that a function can be both concave and quasiconvex or both
convex and quasiconcave. A
straight line is both concave and convex (as well as quasiconcave
and quasiconvex).
Example 5
Suppose that f = −x2 − y2 − z2 and that x, y, and z are
non-negative.
We know that −x2 is concave and that the same holds true for the
other two variables. We also
know that the sum of concave functions is concave and that a
concave function is quasiconcave.
Nevertheless, let us test directly whether f is quasiconcave as
this exercise will help to cement
our understanding.
167
H =
=
0 !2x !2y !2z !2x !2 0 0 !2y 0 !2 0 !2z 0 0 !2
H2 = 0 fx fx fxx
and 0 fy fy fyy
and 0 fz fz fzz
It is always the case that these are ! 0.
H 3 =
= 0 !2x !2y
, 0 !2x !2z
, 0 !2z !2y !2z !2 0 !2y 0 !2
These equal 8x2 +8y2 ! 0; 8x2 +8z2 ! 0; and 8z2 +8y2 ! 0,
respectively. Notice once again that if I switch 1 row and 1
column, the value of the determinant remains the
same. In the last case, the z and y rows and z and y columns were
switched.
H 4 = H =
=
0 !2x !2y !2z !2x !2 0 0 !2y 0 !2 0 !2z 0 0 !2
Expanding by co-factors (by going down the last column) and
repeating the first two rows for
easy multiplication, we get,
H = 2z !2x !2 0 !2y 0 !2 !2z 0 0
!2x !2y !2z
!2 0 0
0 !2x !2y
!2x !2 0
168
So, the bordered hessian is negative semidefinite and therefore the
function is quasiconcave.
Note that because of the border, we have positive, negative,
positive, negative starting with the
first 1 by 1 in contrast to the test for concavity where negative
semidefinite means having
negative, positive, negative, positive without any border.
Example 6
Suppose that f = x2 + y2 + z2 and that x, y, and z are
non-negative.
We know that x2 is convex and that the same holds true for the
other two variables. We also
know that the sum of convex functions is convex and that a convex
function is quasiconvex.
Nevertheless let us test directly whether f is quasiconvex as this
exercise will further help to
cement our understanding.
H =
=
0 2x 2y 2z 2x 2 0 0 2y 0 2 0 2z 0 0 2
H2 = 0 fx fx fxx
and 0 fy fy fyy
and 0 fz fz fzz
It is always the case that these are ! 0.
H 3 =
0 fx fy fx fxx fxy fy fyx fyy
= 0 2x 2y 2x 2 0 2y 0 2
, 0 2x 2z 2x 2 0 2z 0 2
, 0 2z 2y 2z 2 0 2y 0 2
These equal −8x2 −8y2 ! 0,−8x2 −8z2 ! 0, and −8z2 −8y2 ! 0,
respectively.
169
H 4 = H =
=
0 2x 2y 2z 2x 2 0 0 2y 0 2 0 2z 0 0 2
Expanding by co-factors (by going down the last column) and
repeating the first two rows for
easy multiplication, we get,
H = !2z 2x 2 0 2y 0 2 2z 0 0
2x 2y 2z
2 0 0
+ 2 0 2x 2y 2x 2 0 2y 0 2
0 2x 2y
2x 2 0
= !16z2 !16y2 !16x2 " 0
Notice that that each expansion is now negative. We do not have a
commonly used term for an
all-negative expansion. We use positive semidefinite for an
all-positive expansion, and negative
semidefinite for an alternating negative and positive expansion
(when there is no border) and an
alternating positive and negative expansion (when there is a
border). That is why the easiest way
to remember all the rules is to test for convexity and
quasiconvexity by multiplying the function
by minus one and testing for concavity or quasiconcavity.
HOMEWORK: Find out if concave, convex, quasiconcave, quasiconvex or
none of the above for x, y ≥ 0. Your
homework will be completed faster if you remember that a sum of
concave functions is concave
(but remember that a sum of quasiconcave functions need not be
quasiconcave).
1) (x + y)2 + x 2) x3 + y1/2 3) x1/3 ! 3xy
170
4) x1/2 + y1/3 + 3(xy)1/2 5) 2 log(x+1) + (1/3)log(y+1), where log
means natural log.
171
Nonlinear programming is a generalization of linear programming,
unconstrained maximization
and constrained maximization.
Maximize ! = f x1, x2 ,x3 . . .xN( )
subject to
. . . . gM x1, x2, . . .xN( ) ! RM xi " 0
Max ! = x1 100 " x1 2[ ] " x2 300 " x2
3[ ] + x1x2 = f x1,x2( )
g2 x1, x2( ) = x1 2 + 3x2 ! 5
x1, x2 ! 0
Remember that g j x1, x2 . . .( ) is the jth function of the
variables x1, x2 . . . we could have
g x1 . . .( ), h x . . .( ), k x . . .( ) but this would be
notationally difficult.
If f and g j are linear functions then we have a linear program. If
we have no constraints but f is
nonlinear then we have optimization using calculus. If we have the
constraints holding with strict
equality, we have optimization with equality constraints.
172
2) Graphical Representation of Nonlinear programming.
Nonlinear programming does not have the graphical restrictions
found in standard economic
undergraduate texts. This can be illustrated via several
examples.
Minimize C = x1 ! 4( )2 + x2 ! 4( )2 subject to
2x1 + 3x2 ! 6
!3x1 ! 2x2 " !12
x1, x2 ! 0
1
2
3
4
5
6
0
2!!!!!!!!!!!!!!2
The feasible set is shaded in.
A) Optimal solution as seen here need not be located at an extreme
point.
B) Only one constraint exactly fulfilled even though two
constraints (including non-
negativity there are four constraints) and two unknowns.
173
Minimize C = x1 ! 4( )2 + x2 ! 4( )2 subject to
x1 + x2 ! 5
6 5 4 3 2 1
1 2 3 4 5 6
Again the feasible set is shaded in. In this example, the optimal
solution does not lie on the
boundary and thus none of the constraints are exactly fulfilled. So
can't narrow down to
boundary points. Here the constrained and unconstrained optimums
are identical.
174
Our third example displays a problem—non convexity – that we will
try to avoid. In this
case, we have a non-convex constraint space (we may also have a
non-convex preference set
which would also cause problems) We have illustrated this before,
but now we do so in a
different way..
2x1 + 3x2 ! 12
x1, x2 ! 0
6 5 4 3 2 1
1 2 3 4 5 6
The straight lines are the indifference curves. The further to the
right, the higher the
indifference curve. We have a local maximum at 2, 3 but not a
global maximum.
175
3) Kuhn Tucker Necessary Conditions
In calculus, the FOC or necessary conditions for an interior
maximum were f i x1, x2( ) = 0
for all i. We now consider the necessary conditions for a maximum
in nonlinear programming.
These are known as the Kuhn-Tucker necessary conditions. We start
off with a very simple
problem—a maximization problem with non-negativity constraints.
This will enable us to see
the logic behind the Kuhn-Tucker conditions.
Max f x( )
0 (C)
where f x( ) is a one variable differentiable function. Let ! f
"
df dx
. If a local maximum occurs in an interior point (Figure A), the
first order condition
is ! f x( ) = 0 . This is the same necessary condition as the free
extremum. Even in the second
example (Figure B), where we have a local extremum at a boundary
point, ! f x( ) = 0 . As a third
possibility, we may have a local maximum at a boundary even though
! f 0( ) < 0 .
Thus the necessary conditions for a maximum are:
A. ! f x( ) = 0 and x > 0 B. ! f x( ) = 0 and x = 0 C. ! f x( )
< 0 and x = 0
For (C) we are at the lower boundary and we cannot decrease x more
and make f (x) larger.
These conditions can be stated more elegantly as follows:
! f x( ) ≤ 0 x ≥ 0 and x ! f x( ) = 0 non-negativity says at least
one constraint must equal zero
176
For a minimum the necessary conditions are as follows: ! f x( ) ≥ 0
x ≥ 0 and x ! f x( ) = 0
More generally, when ! = f x1, x2 . . . xN( ) = f x( ) , x is a
vector, and xi ! 0 , then the K-T
necessary conditions for a maximum are:
fi x( ) = !f !xi
Finally, we consider the most general case with inequality
constraints:
Max ! = f x( ) = f x1, x2 , . . . xN( )
subject to
g2 x( ) ! R2
gM x( ) ! RM
This can be rewritten as a Lagrange multiplier
"Max" W x1, x2 , . . . xN ,!1,!2 , . . .!M ( ) = f x1, x2, . . .
xN( ) + !1 R1 " g1( ) + !2 R2 " g2( ) . . . !M RM " gM( )
Max is in quotes. Since we are finding a saddle-point not a
maximum.
Necessary condition—K-Tucker necessary conditions
!W !xi
• xi = 0.
!W !"i
• "i = 0. .
177
Note that the Lagrangian method has converted the constraints into
first order conditions. Also
note that the KT complementary slack requirements !W
i!" # i" = 0 means that we have only
added zero's to the objective function.
A natural question is to ask how the non non-negativity constraints
work. A simple example in
the homework should provide insight.
178
4. Objective Function Requirements A. f x1, x2, . . . xn( ) is
either (i) concave or
(ii) quasiconcave and there exists no point x* = x1 * , . . . ,
xn
*( ) such that f i x *( ) = 0 for all i.
The rationale behind these assumptions: Say you have a concave
objective function like the
following diagram:
f(x)
x
Then everything is OK. When you max you end up at a global max. The
same holds true if
function looked like this.
But, if the objective function looks like the following,
then one could find what looks like a "max" at the inflection point
using the 2nd derivative test.
So, to avoid this problem we must rule out complete local
satiation.
179
4) Arrow Enthoven Constraint Qualifications
In order to avoid certain pathologies the following assumptions
concerning the constraints must
be true. These are known as the Arrow Enthoven Constraint
Qualifications:
a) Every constraint functiong j x( ) in the maximization problem is
twice differentiable and
quasiconvex in (an open set containing) the non negative
orthant.
b) There exists a point x0 in the non negative orthant such that
all the constraints are non-
binding (although satisfied). That is, the feasible set has an
interior point.
c) One of the following is true:
i. g j x( ) is convex, or
ii. There exists no point x* in the feasible set where gj i x*( ) =
0 for all i and j.
Discussion of points (a), (b) and (c):
(a) We know from earlier lectures that, for a maximization problem,
quasiconvex constraints
create convex feasible sets (that is, if f(x) is quasiconvex, the
set of all x such that f(x) ! k is
convex). From the opposite perspective, we have seen diagrams
illustrating how non-convex
feasible sets may yield non-optimal solutions even when the
indifference curve is tangent to the
feasible set. So requirement (a) is ensuring convex feasible sets.
In a homework exercise, we will
engage in a mathematical example showing that the KT conditions are
not sufficient for a
maximum when quasiconvexity of the constraints is violated.
Fortunately, we already know how
to test for quasiconvexity.
(b) In a homework exercise we will see via a simple example the
problems that arise when (b) is
violated.
180
(c) This is getting rid of constraint inflection points where the
derivative is 0 (and is thus parallel
to getting rid of objective function inflection points). If i
holds, that is, the function is convex,
then we do not have to worry about inflection points. If ii holds,
then we do not have to worry
either even though we only have quasi-convexity because the
condition rules out the possibility
of inflection points.
Example: Testing to see whether c is satisfied.
There is one constraint, LOG(x1 +1)! LOG(x2 +1) " 7 . Then g1 = 1/
(x1 +1) and g2 =
−1/(x1 +1); g11 = !1/ (x1 +1) 2;g12 = 0;g21 = 0;g22 = 1/ (x2
+1)
2. Looking at the hessian of second
derivatives, we can immediately see that the hessian !1/
(x1+1)
2 0
is not positive
definite. So, the constraint is not convex. Hence, c.i is not
satisfied.
We next test for quasiconvexity.
0 1/ (x1+1) !1/ (x2+1)
1 / (x1+1) !1/ (x1+1) 2 0
!1/ (x2+1) 0 1 / (x2+1) 2
0 1 / (x1+1) !1/ (x2+1)
1 / (x1+1)
2 !1/ (x2+1) 2(x1+1)
2 = 0
So the constraint is quasiconvex (as well as being
quasiconcave).
(c.ii) is satisfied as there are no set of values where both g1 and
g2 = 0. (Indeed there are no non-
negative values of x1 and x2 where either g1 or g2 is equal to 0.)
Hence the AECQ conditions are
satisfied.
181
Max ! = f x( ) x = x1, x2 ,x3, . . . xn
ST
xi ! 0 i =1, . . . , N
If the following is true then x* gives a global maximum.
(1) The objective function, f, is differentiable and quasiconcave
in an open set containing the
non-negative orthant.
(2) Each constraint, g j , is differentiable and quasiconvex in an
open set containing the non-
negative orthant.
(3) x* satisfies the Kuhn-Tucker necessary conditions.
(4) The set g j satisfies the AECQ conditions (this is a redundant
because an element of (3) is
that the AECQ or another set of constraint qualification
hold):
a) g j is either convex
or quasiconvex but there exists no point x* such that gj i x*( ) =
0 for all i, j,
b) There is an interior point to the constraint set.
(5) The second derivative of f exists and
a) f x( ) is concave
b) or quasiconcave and there exists no point x* such that f i x *(
) = 0 for all i (rules out
the local satiation).
182
A word about necessary and sufficient conditions: Sufficient means
that if the conditions
hold you will have a maximum; necessary means that if the
conditions do not hold, you will not
have a maximum (equivalently, only if the conditions hold is a
maximum possible).
HOMEWORK
A. In the following examples, first establish (1) whether the
constraint qualifications are met and
(2) whether the conditions on the objective function are met. Then
find the KT necessary
conditions and solve for all values. If (1) or (2) are not
satisfied show that the KT necessary
conditions can be met, yet the objective is not achieved or that it
is impossible for the K-T
conditions to be met. In all cases, x ≥ 0. Also solve
graphically.
Note that a monotonic function of one variable is quasiconcave and
quasiconvex. In some of
these problems, certain conditions are violated. Therefore the K-T
conditions are no longer
necessary and sufficient. Therefore it is possible for the KT
conditions to be satisfied and yet the
point not be a maximum.
1. Max f x( ) = x subject to x ≤ 10 (Note that λ is the value of
relaxing the constraint and
that it is at least as large as the marginal value of x.)
2. Max f x( ) = x2 subject to x3 ≤ 8
3. Max
f x( ) = x .2 subject to x3 ≤ 8 and x ≤ 10
4. Max f x( ) = x + y subject to x ! 2( )2 + y " 4 and x ! 2( )2 !
y " !4 .
In this case, the AECQ conditions cannot be satisfied.
183
HOMEWORK 2
NOTE: other homework assignments may be substituted for this and
the following
homework assignments.
1. Check if objective function satisfies either:
A. Concavity or
3. Write out Lagriangian
5. Provide an economic interpretation of Kuhn-Tucker conditions.
E.g. marginal
something equals marginal something else.
6. If possible determine which KT equations hold with
equality
7. Perform requested comparative statics based on equalities in
6.
A. A Firm maximizes Profit = [A - BQ(K, L)]Q(K, L) -wL - Ki
Let Q(K, L) be the production function. Q has a hessian of second
order derivitives that is
negative definite. A > 2BQ. A, B > 0. QL, QK > 0. QKL =
0.
Answer questions 1-5.
Also show that Profits are homogeneous of degree 0 in A, B i and w.
Show that the Profit
function is convex.
Find the effect of an increase in A on K; the effect of an increase
of w on K, the effect of an
increase in A on Profits; and the effect of an increase in i on
profits.
184
B. Peak load Pricing C.E.G.
Competitive Firm P1, P2 > 0 Prices in the two periods
given.
Q1 — supply of good in period 1 Q1 > 0 Q2 — supply of good in
period 2 Q2 > 0 Assume that Q2 > Q1
variable cost = C = AQ12 + BQ1Q2 + DQ22A, B, D > 0
4AD > B2
i > 0 interest rate which is the cost of capital K ≥ 0
Max ∏ = P1Q1 + P2Q2 – AQ12 – BQ1Q2 – DQ22 – Ki Present value of
price
Subject to
.01 Q1 ≤ K supply of good uses 1% of capacity
.01 Q2 ≤ K Find the effect of a change in P1 on Q2 Note that Q1 and
Q2 are not functions of K and L, but primitives (variables). The
firm is choosing Q1, Q2 and K subject to constraints.
C. A Shipping Problem
A monopolist ships from point A to B and back. We assume that the
demand curve is
higher for going out than for coming back.
X0 = Number of paid trips out XB = Number of paid trips back
X0 ≥ XB Treat this as a constraint (that is, a ship cannot return
if it does not go out).
We assume that X0 > XB and that AB > 2BXB
P0 = A0 – BX0 A0, AB , B ≥ 0 Pb= AB – BXB
185
HOMEWORK 3
(1) In order for WAIBPC to be a profit function, what restrictions
are there on A and B and C.
W = Wage, I = Interest Rate, P = Price of Good.
(2) Suppose Q(K, L) = 1 3K
1 3L and the firm is competitive in input and ouput markets.
A) Derive Profit maximizing relations.
B) What are second order conditions?
C) Find the slope of the isoquant.
D) Show that ∏ is homogeneous of degree 1.
(3) A monopolist has the following demand curve:
P = A – BQ A, B > 0
CQ2 is the cost curve C > 0
A) Find FOC, SOC.
B) Find the effect of an increase in C on Q.
C) Find the effect of an increase in C on ∏.
D) Write out the profit function, ∏*, explicitly. Remember that a
profit function is a
function of the exogenous variables only.
The cost of a round trip is C(Xo)2 where C is a constant.
Find the effect of an increase of AB on X0.
186
Starting with a Cobb Douglas production function Q = 1 3K
1 3L , find L*, K* and ∏*. Show that
∏* satisfies the properties of a profit function then derive the
cost function.
This is a lot of mathematical manipulation, but you need to do it
to see how everything fits
together.
187
HOMEWORK 5
1. Show that Q B,W + Bi i+2B3 (wi) 1/2[ ] is a cost function.
2. Starting with a Cobb-Douglas technology (with decreasing returns
to scale).
a. Derive the first order conditions for profit maximization.
(3)
b. Show that the second order conditions are satisfied. (4)
c. Determine the elasticity of substitution. (5)
3. Suppose that Q(K,L) is concave with a negative definite Hessian
of second order partials.
a. Find the first order conditions for the minimizing cost subject
to an output constraint. (3)
b. Find the effect of a change in W on cost. (3)
4. Again assume Q(K,L) has a negative defnite Hessian of second
partials.
a. Find the first order conditions of a perfectly competitive
profit maximizing firm (3) and
show the relationship between marginal products and factor prices
when L,K>0. (1)
b. Show that profits are homogeneous of degree O in input and
output prices. (3)
c. Show that if there are constant returns to scale, there are zero
profits. (4)
d. Find the effect of a change in i on profits. (3)
5. Suppose that Q(K,L) = K+L.
a. Derive the isoquant and draw it. (3)
b. Derive the minimal cost function (4) and derive, explain, and
draw the isocost curve. (4)