Suppose that V is finite-dimensional vector space and let β = { e1, e2, … , en }
be a basis of V. For each i = 1, 2, … , n, define a linear functional ej : V ↦ 𝔽 by setting
It is customary to denote dual bases in V* by the same letters as corresponding bases in V but using lower indices ("downstairs") for V and upper indices ("upstairs") for V*. Then vectors v = viei in V can be represented
by objects vi with an upper index, and dual vectors φ = φiei in V* by objects φi with a lower index. It is convenient to use the Einstein summation convention and drop the summation symbol ∑. In this language, the canonical isomorphism between V and V* is realized by
lowering and raising indices. Upper index objects vi represent coordinate vectors relative to a basis ei of V and are also called covariant vectors.
Lower index objects φi are the coordinates of dual vectors relative to the dual basis ei and are also called contravariant vectors.
Theorem 2:
The set β* is a basis of V*. Hence, V* is finite-dimensional and dimV* = dimV.
First we check that the functionals { e1, e2, … , en } are linearly independent. Suppose that there exist 𝔽-scalars k1, k2, … , kn so that
Note that the 0 on the right denotes the zero functional; i.e., the functional that sends
everything in V to 0 ∈ 𝔽. The equality above is an equality of maps, which should hold
for any v ∈ V we evaluate either side on. In particular, evaluating both sides on every element from the basis β, we have
Again, this means that both sides should give the same result when evaluating on any v ∈ V. By linearity, it suffices to check that this is true on the basis β. Indeed, for each i, we get
again by the definition of the ei and the bi. Thus, φ and b1e1 + ⋯ + bnen agree on the basis, so
we conclude that they are equal as elements of V*. Hence {e1, e2, … , en} spans V* and therefore
forms a basis of V*.
Example 5:
In vector space ℝ³, let us consider a (not orthogonal) basis
Now suppose that we realize that I made a typo enterning matrices e₁, e₂, e₃. Suppose that in entries of vector e₃ I made one mistake and base vectors must be as follows:
but not between individual vectors in V
and its dual V*. Moreover, changing only one element of a basis β in V changes several elements in the dual basis β* (see Example 5).
Corollary 1:
Let φ ∈ V* be a nonzero functional on V, and let v ∈ V be any element such that φ(v) = <φ|v> ≠ 0.Then V is a direct sum of two subspaces:
Let v ∈ V be a vector that is not annihilated by linear form φ(v) = <φ|v> ≠ 0. Then we can choose a basis β of V such that β = { e1 = v, e2, … , en }, so that the first vector in this set is v. Let β* = { e1, e2, … , en } be a dual basis in V*. By definition of dual basis, e1(v) = 1, and for all other basis elements, we have
e1(ej) = 0, j = 2, 3, … , n. We can prove that kerφ⊂span{e2, … , en} and span{e2, … , en}⊂kerφ, so kerφ=span{e2, … , en}.
Therefore, the kernel (or null space) of e1 is n − 1 dimensional.
Hence, the functional φ has the following representation in the dual basis:
\[
\phi = \phi ({\bf v})\,{\bf e}^1 = c\, {\bf e}^1 , \qquad c\ne 0.
\]
This functional maps the rest basis elements { e2, … , en } to zero, so V = span(v) ⊕ kerφ because every vector u ∈ V has a unique representation
\[
{\bf u} = c^1 \cdot {\bf e}_1 + c^2 \cdot {\bf e}_2 + \cdots + c^n \cdot{\bf e}_n ,
\]
and
\[
{\bf v} = {\bf e}_1 = {\bf e}_1 + 0\cdot {\bf e}_2 + \cdots + 0\cdot{\bf e}_n .
\]
Example 6:
We identify the two dimensional vector space ℂ² of ordered pairs (z₁, z₂) of two complex numbers with its isomorphic image of column vectors:
We consider a linear form
\[
\varphi \left( z_1 , z_2 \right) = z_1 - 2{\bf j}\, z_2 .
\tag{6.1}
\]
As usual, j denotes a unit imaginary vector in ℂ, so j² = −1.
The kernel of φ consists of all pairs of complex numbers (z₁, z₂) for which z₁ = 2jz₂. Hence, the null space of φ is spanned on the vector
\[
\mbox{ker}\varphi = \left\{ \left[ \begin{array}{c} 2{\bf j}\,y \\ y \end{array} \right] : \quad y \in \mathbb{C} \right\} .
\tag{6.2}
\]
Let z = (z₁, z₂) be an arbitrary vector from ℂ²,and v = (1+j, 1−j) be given. We have to show that there is a unique decomposition:
\[
{\bf z} = \left( z_1 , z_2 \right) = c\left( 1 + {\bf j}, 1 - {\bf j} \right) + \left( 2{\bf j}, 1 \right) y ,
\tag{6.3}
\]
where y∈ℂ is some complex number. Equating two components, we obtain two complex equations:
\[
\begin{split}
z_1 &= c \left( 1 + {\bf j} \right) + 2{\bf j}\, y ,
\\
z_2 &= c \left( 1 - {\bf j} \right) + y .
\end{split}
\tag{6.4}
\]
Since the matrix of system (6.4) is not singular,
\[
{\bf A} = \begin{bmatrix} 1 + {\bf j} & 2{\bf j} \\ 1 - {\bf j} & 1 \end{bmatrix} \qquad \Longrightarrow \qquad \det{\bf A} = -1-{\bf j} \ne 0 ,
\]
1 + I - 2*I*(1 - I)
-1 - I
system (6.4) has a unique solution for any input z = (z₁, z₂):
\[
c = \frac{1 + {\bf j}}{2} \left( z_1 - 2{\bf j}\, z_2 \right) , \qquad y = - z_2 - {\bf j}\, z_1 .
\]
{{c -> (-(1/2) + I/2) (z1 - 2 I z2), y -> -I (z1 - I z2)}}
End of Example 6
■
In other words, any vector v not in the kernel of a linear functional generates a supplement of this kernel: The kernel of a nonzero linear functional has codimension 1. When V is finite dimensional, the rank-nullity theorem claims:
If a vector space V is not trivial, so V ≠ {0}, then its dual V* ≠ {0}.
Theorem 3:
Let φ, ψ be linear functionals on a vector space V. If φ vanishes on ker(ψ), then φ = cψ is a constant multiple of ψ.
If ψ = 0, there is nothing to prove. Let us assume that ψ ≠ 0, and
choose a vector x ∈ V such that ψ(x) ≠ 0. Consider the linear functional ψ(x)φ − φ(x)ψ,
which vanishes on x and on kerψ by assumption. By Corollary 1, this linear functional vanishes identically: We get a linear dependence relation.
Another proof:
∀ ϕ ∈ V*, ∃ x such that ϕ(x) ≠ 0, according to Corollary 1, V = span(x) ⊕ kerϕ ⇒ ∀ v ∈ V, v = cx + y, y ∈ kerϕ ⇒ ψ(v) = ψ(cx + y) = ψ(cx) + ψ(y) = cψ(x) + 0 = c ψ(x)/ϕ(x) ϕ(x) + ϕ(y) = ψ(x)/ϕ(x) ϕ(cx) +ψ(x)/ϕ(x) ϕ(y) = ψ(x)/ϕ(x) ϕ(cx + y) = ψ(x)/ϕ(x ) ϕ(v).
Example 7:
We consider V = ℝ3,3, the set of all real-valued 3×3 matrices, and φ a linear functional on it:
where tr(M) is the trace of matrix M. The dimension of V is 9. Functional φ annihilates matrix M if and only if its trace is zero:
\[
a_{11} + a_{22} + a_{33} = 0 ,
\]
which imposes one condition (linear equation) on the coefficients of matrix M. Therefore, the kernel (null space) of φ is 8-dimensional (9 − 1 = 8). The image of φ is just ℝ, that is, one-dimensional vector space.
Trace is the sum of the diagonal elements in a matrix. Here is how Mathematica computes the Trace.
Theorem 4:
Let φi ∈ V* (1 ≤ i ≤ m)
be a finite set of linear functionals. If a linear functional ψ ∈ V* vanishes on intersection \( \displaystyle \cap_{1 \leqslant i \leqslant m} \mbox{ker}\varphi_i , \) then ψ
is a linear combination of the φi, so ψ ∈ span(φ1, … , φm).
Let us proceed by induction on m. For m = 1, it is the
statement of Theorem 3. Let now m ≥ 2, and consider the subspace
\[
W = \cap_{1 \leqslant i \leqslant m} \mbox{ker}\varphi_i \supset W \cap \mbox{ker}\varphi_1 ,
\]
as well as the restrictions of ψ and φ1 to W. By assumption, \( \displaystyle \left. \psi\right\vert_{W} \) vanishes on
kerφ1|W. Hence there is a scalar c such that
ψ|V = cφ1|V, namely,
ψ − cφ1 vanishes on W. By the induction assumption, ψ − cφ1 is a linear combination of φ2, … , φm, say
Another way to prove this theorem is to to construct a basis according to the following algorithm.
Let β = { }.
Choose φj from the given finite set of linear functionals.
If φj vanishes on the kernel of any linear functional in β, do nothing and start from 2 again. Else, ∃vj∈V, such that φj(vj)≠0, and we can scale vj to make φj(vj)=1. We then add vj to β.
Let's suppose that after going through the given finite set of linear functionals, we have β = {v1, … , vr}, then either β is a basis for the vector space V, or we can expand it to a basis for the vector space V, such as β = {v1, … , vr, ur+1, … , un}, and there is φj in our given finite set of linear functionals such that φj(vj) = 1. Further more, it is not hard to prove that ∩1≤j≤mker(φj) = span(ur+1, … , un).
For such basis β of the vector space V, there is corresponding basis β* = {e1, … , er, er+1, … , en} of dual space V*, in which ej = φj. For any φ ∈ V*, \( \displaystyle \varphi = \sum_{i=1}^n c_i {\bf e}^i . \) Since φ vanishes on ∩1≤j≤mker(φj), cj = 0 for j = r+1, … , n. Then we have
\[
\varphi = \sum_{j=1}^n c_j {\bf e}^j =\sum_{j=1}^r c_j {\bf e}^j = \sum_{j=1}^r c_j \varphi_j .
\]
Example 8:
Let ℝ≤n[x] be a set of polynomials in variable x of degree up to n with real coefficients. As we know, the monomials ek(x) = xk (k = 0, 1, 2, … , n)) form a standard basis in this space.
where \( \displaystyle n^{\underline{j}} = n \left( n-1 \right) \left( n-2 \right) \cdots \left( n-j+1 \right) \) is the j-th falling factorial. As usual, we denote by \( \displaystyle \texttt{D}^0 = \texttt{I} \) the identity operator, and 0! = 1. It is not hard to check that ej form a dual basis for monomials xi. Indeed, substituting p(x) = xk into Eq.(8.1), we get
This polynomial expansion is well-known in Calculus as the Maclaurin formula.
(The Maclaurin formula is represented as a form of Taylor Series in Mathematica.)
Let us consider the linear form
\[
\psi \left( p(x) \right) = p(0) + p' (0) .
\]
This functional vanishes on the intersection of kernels ker(e0) ∩ ker(e1) ⊂ker(ψ). Therefore, ψ = c₀e0 + c<₁b>e1, for some scalars c₀ and c₁.
End of Example 8
■
Since a vector space V and its dual V* have the same dimension, hence, they are isomorphic. In general, this isomorphism depends on a choice of bases. However, a canonical isomorphism (known as Riesz representation theorem) between V and its dual V* can be defined if V carries a
non-degenerate inner product.
Theorem 5:
If v and u are any two distinct vectors of the n-dimensional vector space V, then there exists a linear functional φ on V such that <φ|v> ≠ <φ|u>; or equivalently, to any nonzero vector v ∈ V, there corresponds a ψ ∈ V* such that
<ψ|v> ≠ 0.
This theorem contains two equivalent statements because we can always consider the difference u − v to reduce the first statement to the homogeneous case. So we prove only the latter.
Let β = { e1, e2, … , en } be any basis in vector space V, and let β* = { e1, e2, … , en } be the dual basis in V*. If
\( {\bf v} = \sum_i k_i {\bf e}_i , \) then <ej|v> = ki. Hence, if <ψ|v> = 0 for all ψ, in particular, if <ej|v> = 0 for all j = 1, 2, … , n, then v = 0.
Example 9:
Let V = ℭ[0, 1] be the set of all continuous functions on the interval [0, 1]. Let f and g be two distinct continuous functions on the interval. Then there exists a point 𝑎 ∈ [0, 1] such that f(𝑎) ≠ g(𝑎). Let us consider a linear functional δ𝑎 that maps these two functions f and g into different real values. Indeed, let
\[
\delta_a (f) = f(a) ,
\]
Then δ𝑎 is a required functional that distinguishes f and g. As it is widely used in applications, the functional above can be written with the aid of Dirac delta function:
\[
\int_0^1 \delta (x-a)\,f(x)\,{\text d} x = f(a) .
\]
End of Example 9
From Theorem 5, it follows that a functional on any finite-dimensional vector space is unique; this means that if φ(v) = 0 for every v ∈ V, then φ ≡ 0.
Theorem 6:
Let V be an n-dimensional space with a basis β = { e1, e2, … , en }. If { k1, k2, … , kn } is any list of n scalars, then there is one and only one linear functional φ on V such that φ(ei) = < φ | ei > = ki for i = 1, 2, … , n.
For the existence statement, we can just construct this linear transformation:
\[
\varphi ({\bf e}_i ) = \langle \varphi \mid {\bf e}_i \rangle = k_i , \qquad i=1,2,\ldots , n ,
\]
where β = {e1, e2, … , en} is a basis in V.
Every v in V can be written in the form v = s1e1 + s2e2 + ⋯ + snen in one and only one way because β is a basis. If ψ is any linear functional, then
then ψ is indeed a linear functional and <ψ|ei> = ki.
Example 10:
Let 𝑎1, 𝑎2, … , 𝑎n+1 be distinct points (in
ℝ or ℂ), and let ℝ≤n[x] be the space of polynomials of degree at most n in variable x. Define
n+1 linearly independent functionals
where j in the products runs from 1 to n + 1. The set of Lagrange polynomials β = {p1(x), p2(x), … , pn+1(x)} is a basis in the space ℝ≤n[x]. First, they all are polynomials of degree n, so (3.1) are elements of the vector space ℝ≤n[x]. Second, the Lagrange polynomials are linearly independent. Indeed, suppose that their linear combination is identically zero (for any x)
\[
b_1 p_1 (x) + b_2 p_2 (x) + \cdots + b_{n+1} p_{n+1} (x) = 0 ,
\tag{10.2}
\]
with some scalars b1, b2, … , bn+1 ∈ ℝ. The relation (3.2) must hold for any x because it must be valid in sense of functions (two functions f and g are equal if and only if f(x) = x(g) for any entry x from the domain---ℝ in our case). We evaluate the relation (3.2) at every point 𝑎j and get
\[
b_1 p_1 (a_j ) + b_2 p_2 (a_j ) + \cdots + b_{n+1} p_{n+1} (a_j) = b_j = 0 , \qquad j=1,2,\ldots , n+1 .
\tag{10.3}
\]
Since the numerator of polynomial pk(x) is a product of n simple monomials (x − 𝑎j) that vanish at x = 𝑎j, we have pk(𝑎k) = 1 and
pk(𝑎j) = 0 if j ≠ k, so every coefficient bj in Eq.(3.2) is zero. This proves that the set of Lagrange polynomials is linearly independent. Since there are exactly n + 1 distinct polynomials, we claim that the Lagrange polynomials (10.1) constitute a basis in the space ℝ≤n[x].
Hence,
the system p1(x), p2(x) , … , pn+1(x) is dual to the system φ1, φ2, … , φn+1 in the sense that 〈φk∣pj〉 = δk,j the Kronecker delta.
Now we can construct a polynomial that makes specified values at the given list of points 𝑎1, 𝑎2, … , 𝑎n+1:
Note in the output below, each graphic has a different label for each y axis as the k in Subscript[p, k] increments higher; and the number of points coincide with the value of the k subscript. Accordingly, the product of these approaches the true functional form with added increments.