Mathematical methods for economic theory

Martin J. Osborne

4.2 Optimization: definitions

The optimization problems we study take the form
maxxX f(x) subject to x ∈ S
where X is a set of n-vectors, f is a function defined on X, x is an n-vector (which we can also write as (x1, ..., xn)), and S is a subset of X. We call f the objective function, x the choice variable, and S the constraint set.

To be very clear about what we mean by the “max” operator, here is a precise definition.

Definition
Let f be a function of many variables defined on a set X and let S be a subset of X. The point x* ∈ S solves the problem
maxxX f(x) subject to x ∈ S
if
f(x) ≤ f(x*) for all x ∈ S.
In this case we say that x* is a maximizer of f(x) subject to the constraint x ∈ S, and that f(x*) is the maximum (or maximum value) of f(x) subject to the constraint x ∈ S.
A minimizer is defined analogously. An optimization problem may have many maximizers and minimizers, but has at most one maximum and at most one minimum.

Consider the function f of one variable shown in the following figure, with S equal to the blue line segment. Both x* and x** are maximizers of f(x) subject to the constraint x ∈ S, and x" is a minimizer. (The graph of f is horizontal between x3 and x4 and between x5 and x6.) The number M is the maximum of f(x) subject to x ∈ S and the number m is the minimum.

x1 x' x2 x3 x'' x4 x5 x* x** x6 M m f(x)

What about the point x'? This point is decidedly not a maximizer, because f(x*) > f(x'), for example. But it is a maximum among the points close to it. We call such a point a “local maximizer”. Recall that the distance between two n-vectors x and y is the square root of ∑n
i=1
(xi − yi)2 (the length of the straight line joining x and y).

Definition
Let f be a function of many variables defined on a set X and let S be a subset of X. The point x* ∈ S is a local maximizer of f(x) subject to the constraint x ∈ S if there is a number ε > 0 such that
f(x) ≤ f(x*) for all x ∈ S for which the distance between x and x* is at most ε.
A local minimizer is defined analogously. Sometimes we refer to a maximizer as a global maximizer to emphasize that it is not only a local maximizer. Every global maximizer is, in particular, a local maximizer (ε can take any value), and every global minimizer is a local minimizer.

For the function f in the figure above we have f(x) ≤ f(x') for all x between x1 and x2 (for example), so that x' is a local maximizer of f(x); because x' − x1 = x2 − x', we can take ε = x' − x1 in the definition.

But note that the point x" in this figure is also a local maximizer of f(x), even though it is a global minimizer. The function is constant between x3 and x4. The point x4 is (slightly) closer to x" than is the point x3, so we can take the ε in the definition of a local maximizer to be x4 − x". For every point x within the distance ε of x", we have f(x) = f(x"), so that in particular f(x) ≤ f(x"). The point is that the definition of a local maximizer does not require the value of the function at nearby points to be less than f(x*), only less than or equal to f(x*).

In economic theory we are almost always interested in global maximizers, not merely local maximizers. For example, the standard theory is that a consumer chooses the bundle she most prefers among all those that are available; she is not satisfied by a bundle that is merely better than the other bundles that are nearby. Similarly, a firm is assumed to choose the input-output vector that maximizes its profit among all those that are feasible; it is not satisfied by an input-output vector that merely yields a higher profit than does similar vectors.

Transforming the objective function

Let g be an increasing function of a single variable. (That is, if z' > z then g(z') > g(z).) Then the set of solutions to the problem
maxx f(x) subject to x ∈ S
is identical to the set of solutions to the problem
maxx g(f(x)) subject to x ∈ S.
Why? If x* is a solution to the first problem then by definition
f(x) ≤ f(x*) for all x ∈ S.
But if f(x) ≤ f(x*) then g(f(x)) ≤ g(f(x*)), so that
g(f(x)) ≤ g(f(x*)) for all x ∈ S.
Hence x* is a solution of the second problem.

This fact is sometimes useful: for example, if the objective function is f(x1x2) = xα
1
xβ
2
and x1 and x2 are positive for all points in the constraint set, it may be easier to work with the logarithm of this function, namely αln x1 + βln x2. Since log z is an increasing function, the set of solutions of the problem in which xα
1
xβ
2
is the objective function is the same as the set of solutions of the problem is which αln x1 + βln x2 is the objective function.

Minimization problems

Throughout the tutorial, I concentrate on maximization problems. What about minimization problems? Any discussion of maximization problems encompasses minimization problems because of a simple fact: any minimization problem can be turned into a maximization problem by multiplying the objective function by minus one. That is, the problem
minx f(x) subject to x ∈ S
is equivalent to—in particular, has the same set of solutions as—the problem
maxxf(x) subject to x ∈ S.
Thus we can solve any minimization problem by multiplying the objective function by minus one applying the results for maximization problems.

Although I emphasize maximization problems, I usually state results separately both for maximization problems and for minimization problems.