Elementary Matrices
The transformations we perform on a system or on the corresponding augmented matrix, when we attempt to solve the system, can be simulated by matrix multiplication. More precisely, each of the three transformations:
- interchanging two rows;
- multiplication of a row by a nonzero constant k;
- addition of a constant c times one row to another;
An elementary matrix is a matrix which differs from the identity matrix by one single elementary row operation.
In line with the three elementary row operations, we introduce three types of elementary matrices that correspond to these row operations. In a square matrix, we refer to the entries (i, i) as the diagonal entries, and to the entries (i, j), i ≠ j, as the off-diagonal entries. If we let A be the matrix that results from B by performing one of the operations in the above list, then the matrix B can be recovered from A by multiplying it from left by the corresponding elementary matrix from the following list:
-
A square matrix E ∈ 𝔽m×m is of type I if for some choice of i, j ∈ [1..m], i ≠ j,
all its diagonal entries are 1's except i, j, where they are zeroes; on the other hand, the only other non-zero entries are (i, j) and (j, i) where they are 1's. Multiplication from left by matrix of type I corresponds switching rows i and j in the matrix, which we abbreviate as Ri ↕ Rj.
For example,
\[ \begin{bmatrix} 0&0&1 \\ 0&1&0 \\ 1&0&0 \end{bmatrix} \qquad \iff \qquad R_1 \,\updownarrow\, R_3 . \]
-
A diagonal square matrix E ∈ 𝔽m×m is of type II if all diagonal entries are 1's but one is equal to c ≠ 0.
Multiplication from left by matrix of type II corresponds multiplication of every entry in row i by constant c, which we abbreviate as
Ri ← c·Ri.
For example,
\[ \begin{bmatrix} 1&0&0 \\ 0&c&0 \\ 0&0&1 \end{bmatrix} \qquad \iff \qquad R_2 \,\leftarrow\, c\cdot R_2 . \]
-
A square matrix E ∈ 𝔽m×m is of type III if all its diagonal entries are equal to 1, and
exactly one off-diagonal entry (i, j), i ≠ j, is nonzero. Multiplication from left by matrix of type III corresponds adding c times another row to the fixed row. We abbreviate is as
Ri ← Ri + c·Rj.
For example,
\[ \begin{bmatrix} 1&0&0 \\ c&1&0 \\ 0&0&1 \end{bmatrix} \qquad \iff \qquad R_2 \,\leftarrow \, R_2 + c\cdot R_1 . \]
Multiplication from right by matrices of type I and II provide the same effect on column of the matrix. Transposed matrix of type III provides a similar effect on column of the matrix. This assessment follows from the formula
The elementary matrices generate the general linear group of invertible matrices. Left multiplication by an elementary matrix represents elementary row operations, while right multiplication represents elementary column operations. Elementary row operations are used in Gaussian elimination to reduce a matrix to row echelon form. They are also used in Gauss-Jordan elimination to further reduce the matrix to reduced row echelon form.
Theorem 1: Let A ∈ 𝔽m×n be an m × n matrix over field 𝔽. If E ∈ 𝔽m×m is an elementary matrix, then E A is obtained from A by performing an elementary row operation. More specifically,
- If E is type I and 1’s appear in off-diagonal entries (i, j) and (j, i), then E A is obtained from A by switching rows i and j.
- If E is type II and the diagonal entry (k, k) has value c ≠ 0, then E A is obtained from A by multiplying rowk(A) by c.
- If E is type III and the nonzero off-diagonal entry (k, l) has value c, then E A is obtained from A by replacing rowk(A) by rowk(A) + c · rowl(A).
Theorem 2: Let A be n × n matrix. The following statements are equivalent.
- A is invertible.
- A is a product of elementary matrices.
- A is row equivalent to the n × n identity matrix.
- A has n pivot positions.
- The equation A x = 0 has only the solution x = 0; that is, its null space is ker(A) = {0}.
- The columns of A form a linearly independent set.
- The equation A x = b has at least one solution for every b ∈ 𝔽n×1.
-
The columns of A span 𝔽n×1
- There is a n × n matrix C so that C A = In.
- There is a n × n matrix D so that A D = In.
- Its transpose matrix AT is invertible.
(9) ⇒ (5) Suppose A x = 0. Multiplying both sides on the left with C gives C A x = C 0. Since C A = In, we thus get Inx = x = 0.
(5) ⇒ (6) This is due to the definition of linear independence.
(6) ⇒ (4) Since the columns of A form a linearly independent set, there is a pivot in each column of the echelon form of A. Thus A has n pivot positions.
(4) ⇒ (3) Since A has n pivot positions, and A is n × n, the reduced row echelon form of A is In.
(3) ⇒ (2)
Since the reduced row echelon form of A is In, there exist elementary matrices E1, … , Ek so that EkEk-1 ⋯ E
(2) ⇒ (1) If \( \displaystyle {\bf A} = {\bf E}_1 {\bf E}_2 \cdots {\bf E}_k , \) then since each Ej is invertible, we get by repeated use of formula (A B)−1 = B−1A−1 that \( \displaystyle {\bf A}^{-1} = {\bf E}_k^{-1} {\bf E}_{k-1}^{-1} \cdots {\bf E}_1^{-1} . \) Thus A is invertible.
(1) ⇒ (11) This follows from formula \( \displaystyle \left( {\bf A}^{\mathrm T} \right)^{-1} = \left( {\bf A}^{-1}} \right)^{\mathrm T} . \)
(1) ⇒ (10) Since A is invertible, A A−1 = In. So we can chose D = A−1
(10) ⇒ (8) Let b ∈ 𝔽n×1. Then A D b = Inb = b. Thus, with x = D b, we have A x = b. Hence, b belongs to the column space of A. This holds for every b ∈ 𝔽n&thies;1, and thus this column space is 𝔽n&thies;1.
(8) ⇒ (7) This follows directly.
(7) ⇒ (4) Since the column space of A is 𝔽m&thies;1, every row in the row echelon form has a pivot. Since A has n rows, this gives that A has n pivot positions.
Note that C and D in parts (9) and (10) are the same as C = C In = C(A D) = (C A)D = InD = D and (by definition of the inverse). Thus, indeed, to find the inverse of A it suffices to solve the system A B = I;In as indicated in the algorithm for finding A−1.
- The inverse of the elementary matrix which interchanges two rows is itself.
- The inverse of the elementary matrix which simulate Ri ← k · Ri is the elementary matrix which simulates Ri ← k−1 · Ri.
- The inverse of the elementary matrix which simulates Ri ← Ri + k · Rj is the elementary matrix which simulates Ri ← Ri − k−1 · Rj.
Matrices That Interchange Two Rows of a Matrix
Let
Also, the inverse to the elementary matrix Ei,j is the matrix itself, since if the same two rows of a matrix are interchanged twice, the matrix is unchanged.
Multiplying a Row of a Matrix by a Constant
Adding a Multiple of One Row to Another Row
- Anton, Howard (2005), Elementary Linear Algebra (Applications Version) (9th ed.), Wiley International
- Beezer, R.A., A First Course in Linear Algebra, 2017.
- Boelkins, M.R., Goldberg, J.I., Potter, M.C., Differential Equations with Linear Algebra, Mathematica Help, 2009.