In several areas of mathematics, isomorphism appears as a very general concept. The word derives from the Greek "iso", meaning “equal”, and "morphosis", meaning “to form” or “to shape.” Informally, an isomorphism is a map that preserves sets and relations among elements. When this map or this correspondence is established with no choices involved, it is called canonical isomorphism.
The construction of a dual may be iterated. In other words, we form a dual space \( \displaystyle \left( V^{\ast} \right)^{\ast} \) of the dual space V*; for simplicity of notation, we shall denote it by
V**.
We prefer to speak of a function f (not mentioning its variable), keeping f(x) for the image of x under this map. This approach is very convenient when dealing with linear functionals. For example, Dirac's delta function
\[
\delta_a \,:\ f \,\mapsto f(a)
\]
is a linear functional on a set of test functions.
The second dual V** is canonically (i.e., in a natural way) isomorphic to V. This means that any vector v ∈ V canonically defines a linear functional δv on V* (i.e. an element of the second dual V**) by the rule:
Note, that the null space of T is ker(T) = {0} due to uniqueness of linear functional. Finally, letting the element v vary in 𝔽-vector space V, we obtain a linear map
Theorem 7:
The linear map δ : V ⇾ V**, defined by v ↦ δv (φ ↦ φ(v), is injective function. When V is finite dimensional, it is a canonical isomorphism.
The meaning of δv = 0 is φ(v) = δv(φ) = 0 for all linear functionals φ. It implies that v = 0 due to Lemma 1. Hence, kerδ = {0}, so the linear map \eqref{EqDual.4} is injective. In the finite-dimensional case, we have
Example 11:
The "dual" of a vector is the linear transformation that it encodes onto the real or complex number. Another way of expressing the same concept is: the "dual" of a linear transformation from space to one dimension (ℂ or ℝ) is a certain vector in that space. A non-mathematician might, as a result of this, ascribe somewhat magical properties to the dot product because that is just what the dot product does: locate an answer to a vector question on the real number line or complex plane. Mathematicians and physicists have long studied the notion of Duality and have given various interpretations of it. One of these is that when two different mathematical abstractions share [some] properties, they are endowed with a correspondence that is very useful in practice. More complex transformations, such as Fourier, permit one to make theoretical progress on otherwise intractable problems in the transform space, achieving a result otherwise not possible and then transform that result back to the original for a practical answer that was out of reach in the pre-transformed expression.
The figure below is a simple example of duality and how one may move easily from one to the other and back. Notice the commonality of six and eight in each figure. Each object has a set of 6 somethings and a set of eight somethings, but they are not the same things. For the cube it is six faces and eight vertices. Count the equivalent for the octahedron. Note the each have the same number of edges. There is a mathematical relation between the cube and the octahedron which permits one to "convert" one shape to the other; reversing that process returns to the original shape. Naturally, there is a scale factor involved, explaining why the outside cube is much larger than the one inside. The mathematical relationship is described above where the center of the faces of one are linked, thereby creating the other.
Since dimV** = dimV* = dimV, the condition ker(T) = {0} implies that
T is an invertible transformation (isomorphism).
The isomorphism T is very natural because it was
defined without using a basis, so it does not depend on
the choice of basis. So, informally we say that V** is canonically isomorphic
to V; the rigorous statement is that the map T described above (which we consider to be a natural and canonical) is an isomorphism from V to V**.
When we defined the dual space V* from V, we did so by picking a special basis (the dual basis), therefore, the isomorphism from V to V* is not canonical because it depends on the basis.
Since the dual space is intuitively the space of “rulers” (or measurement-instruments) of our vector space, and as so there is no canonical relationship between rulers and vectors because measurements need scaling, and there is no canonical way to provide one scaling for space.
On the other hand, the isomorphism \eqref{EqDual.4} between the initial vector space V and its double dual V** is canonical since mapping \eqref{EqDual.4} does not depend on local basis.
If we were to calibrate the measure-instruments, a canonical way to accomplish it is to calibrate them by acting on what they are supposed to measure.
Example 12:
A typical linear programming problem that was first formulated by the Russian scientist Leonid Kantorovich in 1939, asks to minimize the cost function c • x subject to constraints:
where A is an m × n matrix with real coefficients and b ∈ ℝm×1 ≌ ℝm. As usual, the cost function c • x = c1x1 + ⋯ + cnxn is a dot product of two vectors. Any vector satisfying the constraint conditions is called feasible.
Theorem A:
If the feasible region is unbounded and the coefficients of the objective function are positive, then the minimum value of the objective function exists. Moreover, the minimum value of the objective function occurs at one (or more) of the corner points at the boundary of the feasible region.
The dual problem starts from the same A, b, and c, and reverses everything. In the primary problem, c is in the cost function and b is in the constraint, In the dual, b and c are
switched, The dual unknown y ∈ 𝔽1, m is a row vector with m components, and the feasible set
has y A ≤ c instead of A x ≥ b.
In short, the dual of a minimum problem is a maximum problem. Now y ≥ 0:
The dual of this problem is the original minimum problem. A set of all
solutions y ∈ ℝ1×m ≌ ℝm
satisfying (12.2) is called a polyhedron.
There is a symmetry
between the primal and dual problems. The simplex method applies equally well to a
maximization—anyway, both problems get solved at once.
Theorem B:
When both optimization problems have feasible vectors, they have
optimal solutions x* and y*, respectively. The minimum cost c • x* equals the maximum income y* • b.
If optimal vectors do not exist, there are two possibilities: Either both feasible sets are
empty, or one is empty and the other problem is unbounded (the maximum is +∞ or the
minimum is −∞).
Application to Economics and Finance
The central question of economics is how to allocate scarce resources amongst alternative ends. This turns into a series of sub-issues:
How much of what to apply and when?
How does the application of any particular factor/resource change the output in both absolute and relative terms?
What is the best combination of resources that produces the most efficient result?
The analytical technology required to answer these questions involves a process known as "Optimization" which, in practical terms, becomes finding the maximum or a minium of a function. This can require intertwined functions. For instance, the optimization of net income (revenue minus cost) involves the simultaneous maximization of revenue and minimization of cost.
Life involves trade-offs. Constraints force trade-offs into the picture. You can't go to the movie and the ballgame on the same night. One must make choices. A rational life means making choices which optimize your time/happiness/profits/career/retirement savings. The ideas of maximization and minimization become more interesting in the presence of constraints. Mathematica has very highly evolved algorithms for solving these kinds of problems.
The Primal Problem
It all begins with stating the goal as an "objective function." This is the mathematical expression of what you wish to achieve. So here we have some x and y (inputs, products, choices) that must be combined in a certain way to get us where we want to be.
In order to illustrate the concepts, we consider a simple primal minimization problem:
minimize a cost function \( \displaystyle \qquad f(x_1 , x_2 ) = 120\,x_1 + 100\, x_2 \qquad \iff \qquad f({\bf x}) = {\bf c} \bullet {\bf x} , \quad \) where c = (120, 100) and x = (x₁, x₂),
First, we find the solution of the formulated primal minimization problem
geometrically, and then solve it analytically. So we plot the constraint functions
cntrPlt1 =
ContourPlot[{2 x + 2 y == 8, 5 x + 3 y == 15, 3*x + 8*y == 17}, {x, 0,
6}, {y, 0, 5}, ContourStyle -> Thickness[0.01]];
LLabeled[%, "Figure 12.1: Constraint Lines"]abeled[%, "Figure 12.1: Constraint Lines"]
Then we cut off parts of these lines that do not define the feasible region.
This allows us to plot the feasible region that displays the area bounded by all of the constraints.
Also shown are four red points that represent so-called "corner solutions" where optimal values may be found.
Note below that we have added two more constraints. Both x₁ and x₂ must be non-negative. This example is considered "stylized" in that is simplifies a context for teaching purposes.
The Simplex Method (Danzig, 1946) finds the optimal solution, which must be at one of the corners of the feasible region. Three corners of the feasible region are shown below as red points. Sections of the lines that go through two points are known as "active constraints" as they are boundaries of the feasible region. Portions of the lines which are not boundaries of the feasible region are considered "inactive constraints" as values outside the feasible region cannot result in an optimal solution.
Mathematica has several ways to find the minimum of this function, here are two:
The objective function f(x₁, x₂), when equated to a constant, defines a line. Upon using several constants, we obtain several parallel lines that may be derived using several arbitrary values, then plotted.
The dashed line for the optimal solution to the objective function passes through the optimal point in purple.
Now we formulate the dual problem: maximize the objective function
f(y₁, y₂, y₃) = 8 y₁ + 15 y₂ + 17 y₃ (note this is a function of three variables) subject to only two constraints:
In our example, the Primal represents the buyer who wishes to minimize his cost. He needs two inputs, x1 and x2, each with constraints imposed by others (perhaps these are different suppliers with different minimum order sizes): buy two of each and the total cost is not less than 8 (constraint g1); buy 5 x1 and 3 x2 and the total cost is not less than 15 (constraint g2); or buy 3 x1 and 8 x2 and the total cost is not less than 17 (constraint g3). Our buyer needs at least one of each (neither x1 or x2 may be zero). The resulting matrices represents information about this system. That information, together with the minimizing techniques above, tell the buyer that portions 3/2 of x1 and 5/2 of x2 is the least overall cost. When x1 costs $120 and x2 cost $100, the optimal combination at those portions equals $430:
120*3/2 + 100*5/2
430
The Dual is from the prospective of the seller of x1 and x2. His goal is to maximize his revenue. His products are complements of one another. They may be acquired in any combination of two. Thus, his sales of product y are any pair selected from the set {y1, y2, y3}, permitting one of y to be zero. Shipping is accomplished through others who, in turn, impose constraints (gg1 and gg2). Using the same techniques, this time applying maximization, we determine that his optimal production combination is to produce 35 y1, 10 y2 and zero y3. At prices $8 for y1 and $15 for y2 we have the same answer, $430, for the Dual we had for the Primal: