This lecture defines the concept of row equivalence and proves some propositions about row equivalent matrices that lie at the heart of many important results in linear algebra.
We start with a definition of row equivalence.
Definition
Let
and
be two
matrices. We say that
is row equivalent to
if and only if there exist
elementary matrices
such
that
Remember that pre-multiplying
by an elementary matrix is the same as performing an
elementary row
operation on
.
Therefore,
is row equivalent to
if and only if
can be transformed into
by performing a sequence of elementary row operations on
.
Row equivalence is an equivalence relation because it is:
symmetric: if
is row equivalent to
,
then
is row equivalent to
;
transitive: if
is equivalent to
and
is equivalent to
,
then
is equivalent to
;
reflexive:
is equivalent to itself.
Suppose
is row equivalent to
.
Since an elementary matrix is invertible and its inverse is an elementary
matrix, we have that
where
are elementary matrices. Therefore,
is equivalent to
.
This proves symmetry. If
is equivalent to
and
is equivalent to
,
then
and
where
and
are elementary matrices. Now, pre-multiply both sides of the first equation by
:
Then,
is equivalent to
,
that is, row equivalence is transitive. Finally, for any elementary matrix
,
we can
write
Since
is elementary, this means that we can transform
into itself by means of elementary row operations. As a consequence, row
equivalence is reflexive.
The next proposition states an important property of row equivalence, known as column correspondence property.
Proposition
Let
and
be two
matrices. Let
be row equivalent to
.
Denote by
and
the
-th
columns of
and
respectively. Then,
for
an
vector
if and only
if
Since
is row equivalent to
,
we have
that
where
is a product of elementary
matrices:
Furthermore,
by the very definition of matrix product (see also
here):
Thus,
we can
substitute
and
in
the
equation
so
as to
obtain
By
pre-multiplying both sides by
,
we
get
Thus,
we have proved that
implies
.
The opposite implication
(
implies
),
can be proved analogously.
In other words, when
and
are row equivalent, the
-th
column of
can be written as a
linear
combination of a given set of columns of
itself, with coefficients taken from the vector
,
if and only if the
-th
column of
is a linear combination of the corresponding set of columns of
,
with coefficients taken from the same vector
.
A useful corollary of the previous proposition follows.
Proposition
Let
and
be two row equivalent matrices. Then, a set of columns of
is linearly independent if
and only if the corresponding set of columns of
is linearly independent.
The proof is by contradiction. Suppose that
a set of columns of
is linearly independent, but the corresponding columns of
are linearly dependent. It follows that a column
can be written as a linear combination of other
columns:
where
.
In particular, there are some non-zero entries of
corresponding to the columns in the set we are considering. But by the
previous proposition, this implies that
In
other words, the set of columns of
is not linearly independent, a contradiction. Therefore, if a set of columns
of
is linearly independent, then the corresponding columns of
must be linearly independent. The opposite implication can be proved
analogously.
This section introduces the concept of dominant columns, which will be used below to study the properties of row equivalent matrices.
Definition
Let
be a
matrix. Denote its
-th
column by
.
We say that
is a dominant column if and only if it cannot be written as a
linear combination of the
columns to its left.
A first simple result about dominant columns follows.
Proposition
Two equivalent matrices
and
have the same set of dominant columns, that is, the set of indices of the
dominant columns of
coincides with the set of indices of the dominant columns of
.
Suppose
is a dominant column of
.
Then, there is no vector
such
that
By
the column correspondence property above, this is possible if and only if
there is no such vector
satisfying
As
a consequence,
cannot be written as a linear combination of the columns to its left. Hence it
is dominant. We have just proved that
is dominant only if
is dominant. The proof of the opposite implication is analogous. This holds
for any
.
Therefore, the columns that are dominant in
are dominant also in
and vice versa.
For instance, if the dominant columns of
are the second, third and fifth, then the dominant columns of
are the second, third and fifth.
The propositions above allow us to prove some properties of matrices in reduced row echelon form.
Remember that a matrix is in reduced row echelon form (RREF) if and only if:
all its non-zero rows contain an element, called pivot, that is equal to 1 and has only zero entries in the quadrant below it and to its left;
each pivot is the only non-zero element in its column;
all the zero rows (if there are any) are below the non-zero rows.
Furthermore, the Gauss-Jordan elimination algorithm can be used to transform any matrix into an RREF matrix by elementary row operations. Therefore, any matrix is row equivalent to an RREF matrix.
Remember that a basic column is a column containing a pivot, while a non-basic column does not contain any pivot.
The basic columns of an RREF matrix are vectors of the
canonical basis, that is, they
have one entry equal to 1 and all the other entries equal to zero.
Furthermore, if an RREF matrix has
basic columns, then those columns are the first
vectors of the canonical basis, as stated by the following proposition.
Proposition
Let
be a matrix in reduced row echelon form. Then, the
-th
basic column of
,
counting from the left, is equal to the
-th
vector of the canonical basis, that is, it has a 1 in position
and all its other entries are equal to 0.
By the definition of RREF matrix, the basic
columns of
are vectors of the canonical basis (they have one entry equal to 1 and all
other entries equal to 0). Furthermore, all non-zero rows contain a pivot.
Therefore, the
-th
basic column contains the
-th
pivot, which is located on the
-th
row. In other words, the pivot, which is equal to 1, is the
-th
entry of the
-th
basic column.
We now state some simple results concerning basic and non-basic columns.
Proposition A basic column of a matrix in reduced row echelon form is a dominant column.
A basic column contains a pivot, equal to 1, and all the entries to the left of the pivot are equal to 0. Therefore, the basic column cannot be written as a linear combination of the columns to its left (no linear combination of 0s can be equal to 1). Hence, it is a dominant column.
Proposition A non-basic column of a matrix in reduced row echelon form is not a dominant column.
If a column
is non-basic, that is, it has no pivot, then it can be written
as
where
is the number of basic columns to its left (the entries below the
-th
must be zero because the
-th
pivot, with
,
has only 0s to its left). Therefore, the non-basic column
can be written as a linear combination of the columns to its left. For
example, if
and the first, third and fourth columns are basic,
then
Thus,
if a column
is non-basic it is not linearly independent from the columns to its left.
Hence, it is not a dominant column.
By combining the two simple propositions above, we get the following one.
Proposition If a matrix is in reduced row echelon form, then one of its columns is basic if and only if it is dominant, and it is non-basic if and only if it is not dominant.
By the previous proposition, if a column is dominant, then it cannot be non-basic. Therefore, it is basic. We have already established the opposite implication (basic implies dominant). Therefore, a column is dominant if and only if it is basic. The proof of equivalence for non-dominant columns is analogous.
Thus, when a matrix is in reduced row echelon form, we can use the concepts of basic and dominant column interchangeably.
We are now ready to state the most important proposition of this lecture.
Proposition Any matrix is row equivalent to a unique matrix in reduced row echelon form.
We have already explained that any matrix
is row equivalent to a matrix in reduced row echelon form which can be derived
by using the Gauss-Jordan elimination algorithm. We need to prove uniqueness.
Suppose that two matrices
and
are in reduced row echelon form and that they are both row equivalent to
.
Since row equivalence is transitive and symmetric,
and
are row equivalent. Therefore, the positions of their dominant columns
coincide. Equivalently, the positions of their basic columns coincide. But we
have proved above that the
-th
basic column of an RREF matrix, counting from the left, is equal to the
-th
vector of the canonical basis. Therefore, not only the basic columns of
and
have the same positions, but their corresponding entries coincide. The
non-basic columns are linear combinations of the basic ones. By the column
correspondence property above, the coefficients of the linear combinations are
the same for
and
.
But also the vectors being combined linearly coincide because the basic
columns of
and
coincide. As a consequence, each non-basic column of
is equal to the corresponding non-basic column of
.
Thus,
,
which proves that the row equivalent RREF of a matrix is unique.
A consequence of this uniqueness result is that if two matrices are row equivalent, then they are equivalent to the same RREF matrix.
Proposition
Let
be row equivalent to
.
Then,
and
are equivalent to the same RREF matrix
.
Denote by
and
the RREF matrices that are row equivalent to
and
respectively:
where
and
are products of elementary matrices. Furthermore,
is row equivalent to
,
so
that
where
is a product of elementary matrices. We pre-multiply both sides of eq. (3) by
,
so as to
get
Since
is a product of elementary matrices,
is an RREF matrix row equivalent to
.
But the RREF row equivalent matrix is unique. Therefore,
.
In this section we present some corollaries of the results we have proved in the previous sections.
Proposition
Let
be a
invertible matrix. Then,
is row equivalent to the
identity matrix
.
By the results above, we know that
is row equivalent to a unique RREF matrix
.
Furthermore,
can be transformed into
by elementary row operations, that is, by pre-multiplying
by an invertible matrix
(equal to the product of the elementary matrices used to perform the row
operations):
But
we know that pre-multiplication by an invertible (i.e.,
full-rank) matrix
does not alter the
rank. Therefore,
is full-rank. As a consequence, all the columns of
are basic (there cannot be non-basic columns because the columns of a
full-rank matrix are all linearly independent from each other). But this means
that the
columns of
are the
vectors of the canonical basis of the space of
-dimensional
vectors. In other words, they are the
columns of the
identity matrix. Hence,
.
Clearly, since the identity matrix
is a matrix in reduced row echelon form, any invertible matrix is equivalent
to the unique RREF matrix
.
An immediate consequence of the previous proposition follows.
Proposition
Let
be a
invertible matrix. Then,
can be written as a product of elementary
matrices:
where
are elementary matrices.
By the previous proposition, the identity
matrix
is row equivalent to
.
So, by the definition of row equivalent matrix, we have
that
where
are elementary matrices.
While the previous two propositions concern square invertible matrices, the following proposition applies to matrices that can be non-square and non-invertible.
Proposition
Let
be an RREF matrix that is row equivalent to a matrix
.
Then
and
have the same rank. The rank is equal to 1) the number of non-zero rows of
or, equivalently, to 2) the number of basic columns of
.
First of all, remember that pre-multiplying
a matrix
by an invertible matrix
does not change the rank of
.
As a consequence, if
(an invertible product of elementary matrices) transforms
into a row equivalent RREF matrix
,
we have
that
The
rank of
is equal to the maximum number of linearly independent columns of
.
The basic columns of
are linearly independent, while the non-basic columns can be written as linear
combinations of the basic ones. Therefore, the rank of
is equal to the number of basic columns of
.
Furthermore, each basic columns contains a pivot and each non-zero row
contains a pivot. As a consequence, the rank is also equal to the number of
non-zero rows of
.
Please cite as:
Taboga, Marco (2021). "Row equivalence", Lectures on matrix algebra. https://www.statlect.com/matrix-algebra/row-equivalence.
Most of the learning materials found on this website are now available in a traditional textbook format.