Before formulating the Primary Decomposition Theorem, we need to recall some definitions and facts that were explained in other sections.
We remind that the minimal polynomial of a square matrix A (or corresponding lineat transformation)
is the (unique) monic polynomial ψ(λ) of least degree that annihilates the matrix A, that is
ψ(A) = 0. The minimal polynomial\( \psi_u (\lambda ) \) of a
vector \( {\bf u} \in V \ \mbox{ or } \ {\bf u} \in \mathbb{R}^n \) relative to A is the
monic polynomial of least degree such that
\( \psi_u ({\bf A}) {\bf u} = {\bf 0} . \) It follows that \( \psi_u (\lambda ) \)
divides the minimal polynomial ψ(λ) of the matrix A. There exists a vector
\( {\bf u} \in V (\mathbb{R}^n ) \) such that
\( \psi_u (\lambda ) = \psi (\lambda ) . \) This result can be proved by representing
the minimal polynomial as the product of simple terms to each of which corresponds a subspace. Then the original vector
space (or \( \mathbb{R}^n \) ) is the direct sum of these subspaces.
A subspace U of a vector space V is said to be T-cyclic with respect to a linear transformation T : V ⇾ V
if there exists a vector \( {\bf u} \in U \)
and a nonnegative integer r such that \( {\bf u}, T\,{\bf u} , \ldots , T^r {\bf u} \)
form a basis for U. Thus, for the vector u if the degree of the minimal polynomial
\( \psi_u (\lambda ) \) is k, then
\( {\bf u}, T\,{\bf u} , \ldots , T^{k-1} {\bf u} \) are linearly independent and the space
U is spanned by these k vectors is T-cyclic.
Theorem (Primary Decomposition Theorem):
Let V be an
n-dimensional vector space (n is finite) and T is a linear transformation on V.
Then V is the direct sum of T-cyclic subspaces. ■
Let k be the degree of the minimal polynomial ψ(λ) of transformation T (or corresponding matrix
written is specified basis), and let u be a vector in V with
\( \psi_u (\lambda ) = \psi (\lambda ) . \) Then the space U spanned by
\( {\bf u}, T{\bf u} , \ldots , T^{k-1} {\bf u} \) is T-cyclic. We shall prove that if
\( U \ne V \quad (k \ne n), \) then there exists a T-invariant subspace W such
that \( V = U\oplus W . \) Clearly, by induction on the dimension, W will then be the
direct sum of T-cyclic subspaces and the proof is complete.
To show the existence of W enlarge the basis
\( {\bf e}_1 = {\bf u}, {\bf e}_2 = T{\bf u} , \ldots , {\bf e}_k = T^{k-1} {\bf u} \) of
U to a basis \( {\bf e}_1 , {\bf e}_2 , \ldots , {\bf e}_k , \ldots , {\bf e}_n \)
of V and let
\( {\bf e}_1^{\ast} , {\bf e}_2^{\ast} , \ldots , {\bf e}_k^{\ast} , \ldots , {\bf e}_n^{\ast} \)
be the dual basis in the dual space. Recall that the dual space consists of all linear forms on V or, equivalently,
of all functionals on V. To simplify notation, let z = ek*. Then
Consider the dual space U* spanned by
\( {\bf z}, T^{\ast} {\bf z} , \ldots , T^{\ast\, k-1} {\bf z} . \)
Since ψ(λ) is also the minimal polynomial of T*, the space U* is
T*-invariant. Now observe that if
\( U^{\ast} \cap U^{\perp} = \{ 0 \} \) and dim U* = k, then
\( V^{\ast} = U^{\ast} \oplus U^{\perp} , \) where U* and
\( U^{\perp} \) are T*-invariant (since dim \( U^{\perp} \) = n-k).
This in turn implies the desired decomposition \( V = U^{\perp\perp} \oplus U^{\ast\perp} = U \oplus W , \)
where \( U^{\perp\perp} = U \) and \( U^{\ast\perp} = W \)
are T-invariant.
Finally, we shall prove that \( U^{\ast} \cap U^{\perp} = \{ 0 \} \) and
dim U* = k simultaneously as follows. Suppose that
\( a_0 {\bf z} + a_1 T^{\ast} {\bf z} + \cdots + a_s T^{\ast s} {\bf z} \in U^{\perp} , \)
where \( a_s \ne 0 \) and \( 0 \le s \le k-1 . \) Then
This matrix has the characteristic polynomial \( \chi (\lambda ) = \left( \lambda -1 \right)^3 \)
while its minimal polynomial is \( \psi (\lambda ) = \left( \lambda -1 \right)^2 . \)
The matrix A has two linearly independent eigenvectors
Let U and V be one-dimensional subspaces generated by spans of vectors u and v, respectively.
The minimal polynomials of these vectors are the same:
\( \psi_u (\lambda ) = \psi_v (\lambda ) = \lambda -1 \) because
Each of these one-dimensional subspaces U and V are A-cyclic and they cannot form the direct sum of
\( \mathbb{R}^3 . \) We choose a vector \( {\bf z} = \left[ 7, -3, 1 \right]^{\mathrm T} , \)
which is perpendicular to each u and v. matrix A transfers z into the vector
\( {\bf A}\,{\bf z} = \left[ 125, 192, 60 \right]^{\mathrm T} , \) which is perpendicular to
neither u nor v. Next application of A yields