Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation; with no Invariant Sections, no Front-Cover Texts, and no Back-Cover Texts. A copy of the license is included in the appendix entitled GNU Free Documentation License.
Matrix Transformations
Let V and U be vector spaces. We call a function \( T\,:\,V \to U \) a linear transformation (or operator) from V into U if for all
\( {\bf x} , {\bf y} \in V \) any scalar α, we have
We often simply call Tlinear. The space V is referred to as the domain
of the lienar transformation and the space U is called codomain of T. We summarize
the almost obvious statements about linear transformation in the following proposition.
Theorem: Let V and U be a vector spaces and
\( T\,:\, V\to U \) be linear transformation.
If T linear, then \( T(0) =0. \)
T is linear if and only if \( T(c{\bf x} + {\bf y} ) = c\,T({\bf x}) + T({\bf y}) \)
for all \( {\bf x}, {\bf y} \in V \) and any scalar c.
T is linear if and only if for \( {\bf x}_1 , \ldots , {\bf x}_n \in V \) and
any real or complex scalars \( a_1 , \ldots , a_n \)
\[
T \left( \sum_{i=1}^n a_i {\bf x}_i \right) = \sum_{i=1}^n a_i T \left( {\bf x}_i \right) .
\]
Recall that a set of vectors β is said to generate or span a vector space V if every element from
V can be represented as a linear combination of vectors from β.
Example: The span of the empty set \( \varnothing \) consists of a unique element 0.
Therefore, \( \varnothing \) is linearly independent and it is a basis for the trivial vector space
consisting of the unique element---zero. Its dimension is zero.
Example: Recall from section on Vector Spaces that the set of
all ordered n-tuples or real numbers is denoted by the symbol \( \mathbb{R}^n . \) It is a custom
to represent ordered n-tuples in matrix notation as column vectorsd. For example, the matrix
can be used as an alternative to \( {\bf v} = \left[ v_1 , v_2 , \ldots , v_n \right] \quad\mbox{or}\quad
{\bf v} = \left( v_1 , v_2 , \ldots , v_n \right) . \) The latter is called the comma-delimited form of a
vector and former is called the column-vector form.
In \( \mathbb{R}^n , \) the vectors
\( e_1 [1,0,0,\ldots , 0] , \quad e_2 =[0,1,0,\ldots , 0], \quad \ldots , e_n =[0,0,\ldots , 0,1] \)
form a basis for n-dimensional real space, and it is called the standard basis. Its dimension is n.
For example, the vectors
If f is a function with domain \( \mathbb{R}^m \) and codomain
\( \mathbb{R}^n , \) then we say that f is a transformation from
\( \mathbb{R}^m \) to \( \mathbb{R}^n \) or that f
maps \( \mathbb{R}^m \) into \( \mathbb{R}^n , \) which we
denote by wroting \( f\,:\, \mathbb{R}^m \to \mathbb{R}^n . \) In the special case where
n = m, a transformation is sometimes called an operator on \( \mathbb{R}^n . \)
Although the latter represents a linear system of equations, we could view it instead as a transformation that maps a
vector x from \( \mathbb{R}^m \) into the vector from
\( \mathbb{R}^n \) by multiplying x on the left by A.
We call this a matrix transformation and denote by \( T_{\bf A}:\, \mathbb{R}^m \to \mathbb{R}^n \)
(in case where m=n, it is called matrix operator). This transformation is generated by
matrix multiplication.
Theorem: Let \( T:\, \mathbb{R}^m \to \mathbb{R}^n \) be a
linear transformation. Then there exists a unique matrix A such that
\[
T \left( {\bf x} \right) = {\bf A}\, {\bf x} \qquad\mbox{for all } {\bf x} \in \mathbb{R}^m .
\]
In fact, A is the \( n \times m \) matrix whose j-th column is the
vector \( T\left( {\bf e}_j \right) , \) where \( {\bf e}_j \) is the
j-th column of the identity matrix in \( \mathbb{R}^m : \)
\[
{\bf A} = \left[ T \left( {\bf e}_1 \right) , T \left( {\bf e}_2 \right) , \cdots , T \left( {\bf e}_m \right) \right] .
\]
Write \( {\bf x} = {\bf I}_m {\bf x} = \left[ {\bf e}_1 \ \cdots \ {\bf e}_m \right] {\bf x} =
x_1 {\bf e}_1 + \cdots + x_m {\bf e}_m , \) and usethe linearlity of T to compute
Such representation is unique, which could be proved by showing that for any other matrix representation Bx
of transformation T, it follows that A = B.
Example: The transformation T from \( \mathbb{R}^4 \) to \( \mathbb{R}^3 \)
defined by the equations
so multiplication by zero maps every vector from \( \mathbb{R}^m \) into the zero vector
in \( \mathbb{R}^n . \) Such transformation is called the zero transformation from
\( \mathbb{R}^m \) to \( \mathbb{R}^n . \)
■
Example: If I is the \( n \times n \)
identity matrix, then
■
Theorem: For every matrix A the matrix transformation
\( T_{\bf A}:\, \mathbb{R}^m \to \mathbb{R}^n \) has the following properties for all vectors
v anf u and for every scalar k:
Theorem: \( T\,:\, \mathbb{R}^m \to \mathbb{R}^n \)
is a matrix transformation if and only if the following relationships hold for all vectors v and u
in \( \mathbb{R}^m \) and for every scalar k:
\( T \left( {\bf v} + {\bf u} \right) = T \left( {\bf v} \right) + T \left( {\bf u} \right) \) (Additivity property),
where \( {\bf e}_i , \quad i=1,2,\ldots , m , \) are the staaaaaandard basis vectors for
\( \mathbb{R}^m . \) We know that Ax is a
linear combination of the columns of A in which the successive coefficients are the entries
\( x_1 , x_2 , \ldots , x_m \) of x. That is,
\[
{\bf A}\,{\bf x} = x_1 T \left( {\bf e}_1 \right) + x_2 T \left( {\bf e}_2 \right) + \cdots + x_m T \left( {\bf e}_m \right) .
\]
Theorem: Every linear transformation from \( \mathbb{R}^m \) to
\( \mathbb{R}^n \) is a matrix transformation, and conversely, every matrix transformation from
\( \mathbb{R}^m \) to
\( \mathbb{R}^n \) is a linear transformation.
Theorem: If \( T_{\bf A}\,:\, \mathbb{R}^m \to \mathbb{R}^n \) and
\( T_{\bf B}\,:\, \mathbb{R}^m \to \mathbb{R}^n \) are matrix transformations,
and if \( T_{\bf A} \left( {\bf v} \right) = T_{\bf B} \left( {\bf v} \right) \)
for every vector \( {\bf v} \in \mathbb{R}^m , \) then A = B.
To say that \( T_{\bf A} \left( {\bf v} \right) = T_{\bf B} \left( {\bf v} \right) \) for every
vector in \( \mathbb{R}^m \) is the same as saying that
\[
{\bf A}\,{\bf v} = {\bf B}\,{\bf v}
\]
for every vector v in \( \mathbb{R}^m . \) This will be true, in particular, if v
is any of the standard basis vectors \( {\bf e}_1 , {\bf e}_2 , \ldots , {\bf e}_m \) for
\( \mathbb{R}^m ; \) that is,
Since every entry of ej is 0 except for the j-th, which is 1, it follows
that Aej is the j-th column of A and
Bej is the j-th column of B. Thus,
\( {\bf A}\,{\bf e}_j = {\bf B}\,{\bf e}_j \) implies that coerresponding columns of A
and B are the same, and hence A = B.
The above theorem tells us that there is a one-to-one correspondence between n-by-m
matrices and matrix transformations from \( \mathbb{R}^m \) to
\( \mathbb{R}^n \) in teh sense that every \( n \times m \)
matrix A generates exactly one matrix transformation (multiplication by A) and
every matrix transformation from \( \mathbb{R}^m \) to
\( \mathbb{R}^n \) arises from exactly one \( n \times m \)
matrix: we call that matrix the standard matrix for the transformation, which is given by the formula:
\[
{\bf A} = \left[ T \left( {\bf e}_1 \right) \,|\, T \left( {\bf e}_2 \right) \,|\, \cdots \,| T \left( {\bf e}_m \right) \right] .
\]
This suggests the following procedure for finding standard matrices.
Algorithm for finding the standard matrix of a linear transformation:Step 1: Find the images of the standard basis vectors \( {\bf e}_1 , {\bf e}_2 , \ldots , {\bf e}_m \) for
\( \mathbb{R}^m . \) Step 2: Construct the matrix that has the images obtained in Step 1 as its successive
columns. This matrix is the standard matrix for the transformation.
Example: Find the standard matrix A for the linear transformation:
Example: Over the field of complex numbers, the vector space \( \mathbb{C} \)
of all complex numbers has dimension 1 because its basis consists of one element \( \{ 1 \} . \)
Over the field of real numbers, the vector space \( \mathbb{C} \) of all complex numbers
has dimension 2 because its basis consists of two elements \( \{ 1, {\bf j} \} . \)