Although duality is covered in Part 3, we introduce some of its elements here in order to show a different point of view in the context of matrix/vector multiplication.
So far, we have been using a naïve definition of a vector as an ordered (or indexed) finite sequence of scalars; it is usually represented in parenthesis as an n-tuple (x1, x2, … , xn), an element of Cartesian product 𝔽n = 𝔽 × 𝔽 × ⋯ × 𝔽 of n copies of the field of scalars 𝔽. Later, we will extend this definition to include abstract objects; however, for us at this moment, it is sufficient to work with n-tuples. Recall that for scalars we use either rational numbers ℚ or real numbers ℝ or complex numbers ℂ.
Usually, an n-tuple (x1, x2, … , xn), written in parenthesis and a row vector [x1, x2, … , xn], written in brackets, look as the same object to human eyes. One of the pedagogical virtues of any software package is its requirement to pay close attention to the types of objects used in specific contexts. In particular, the computer algebra system Mathematica treats these two versions of a vector differently because
where 𝔽m×n denotes the space of m × n matrices with coefficients from field 𝔽.
Hence, Mathematica considers bracket notation as a matrix with one row.
x = {1, 2}; y = {{1, 2}};
x == y
False
Mathematica considers single curly bracket notation as a list of entries, which is treated as a column vector for convenience. On the other hand, a double curly bracket notation is considered as a list of lists and it is represented as a matrix with one row.
It is a matrix with one row and one column. In order to extract this entry from 1×1 matrix, mathematicians consider this expression as a linear combination. It also can be defined by dot product:
when all entries are real numbers. You need to wait till Part 5 for dot product extension to complex numbers (it is inner product). Although dot product formular above can be used for complex numbers as well, it does not define a metwric (distance) in complex vector spaces. Physicists usually use Dirac's notation (also known as bra-ket notation), which is designated as <a | x> for an alternative to linear combination or dot product.
Every numeric vector a = (𝑎1, 𝑎2, … , 𝑎n) ∈ 𝔽n determines
the corresponding dual vector (known as a covector) a* that defines a linear functional a* : 𝔽n ⇾ 𝔽 via the simple rule
When vector
a = (𝑎1, 𝑎2, … , 𝑎n) is considered as a covector, it is usually called a bra-vector and is denoted by <a∣, while x = (x1, x2, … , xn) ∈ ℂn is called a ket-vector or just a vector and denoted by ∣x>. Then the linear combination \( \displaystyle a_1 x_1 + a_2 x_2 + \cdots + a_n x_n \) (which is dot product for real numbers) defines a linear functional (or linear form), which we denote by a*.
Hence, every n-tupple a = (𝑎1, 𝑎2, … , 𝑎n) defines a covector (or linear functional). In case of real numbers, action of a covector on a vector is just dot product.
Example 1:
The linear functional on ℝ4 given by
\[
f\left( x_1 , x_2 , x_3 , x_4 \right) = 2\,x_1 - 3\, x_2 + x_3 - 5\, x_4
\]
is dual to the numeric vector a = (2, −3, 1, −5). This vector generates a linear functional on ℝ4 by means of dot product
\[
\langle {\bf a} \mid {\bf x} \rangle = {\bf a} \bullet {\bf x} = 2\,x_1 - 3\, x_2 + x_3 - 5\, x_4 ,
\]
for arbitrary vector x = (x₁, x₂, x₃, x₄) ∈ ℝ4. This functional, written in bra-ket notation, can be redefined in standard mathematical form as
\[
{\bf a}^{\ast} \, : \ \mathbb{R} \mapsto \mathbb{R} , \qquad\mbox{with} \qquad
{\bf a}^{\ast} ({\bf x}) = \langle {\bf a} \mid {\bf x} \rangle = {\bf a} \bullet {\bf x} .
\]
On the other hand, the numeric vector a = (2, −3, 1) ∈ ℝ³ is dual to the linear functional (known as a covector) given by
\[
{\bf a}^{\ast} \left( {\bf x} \right) = \left\langle {\bf a} \mid {\bf x} \right\rangle = {\bf a} \bullet {\bf x} = 2\, x_1 -3\, x_2 + x_3 , \quad {\bf x} \in \mathbb{R}^3 .
\]
In particular, actions of the covector 〈a | on some bra-vectors from ℝ³ are given below:
\begin{align*}
{\bf a}^{\ast} (1, 2, 3) &= {\bf a} \bullet \left( 1, 2, 3 \right) = 2\cdot 1 -3\cdot 2 + 3 = -1,
\\
{\bf a}^{\ast} (-1, 2, -5) &= {\bf a} \bullet \left( -1, 2, -5 \right) = 2\cdot (-1) -3 \cdot 2 -5 = -13 .
\end{align*}
End of Example 1
In Linear Algebra, the word duality traditionally refers to the dual
roles of vector and covector associated with a given numeric vector, as
in definition above. The row/column duality just described is one of many
beautiful dualities we highlight
in this tutorial. We will see many other examples of duality, including
image/inverse image, geometric/numeric, and projection/span—problems that
seem difficult from one viewpoint often turn out to be much easier
from the dual viewpoint.
Each of the two perspectives illuminates the
other in a beautiful and useful way, and we shall make good use of this
principle many times.
We can regard an m × n matrix, for instance, as a list of
m (row) vectors in ℝ1×n, or as a list of n (column) vectors in ℝm×1, and
the difference is not superficial, as we shall come to see.
Every m × n matrix
where "T" stands for transposition---a dual operation that transfers rows into columns and vice versa.
Since in the following exposition we utilize matrix multiplication from left, we are forced to consider vectors as columns. Then
In other words, A x is a linear combination of column vectors of matrix A with weights takem from coordinates of vector x.
Alternatively, using dual covectors, we can say the same thing like this: