Matrix inverses
Matrix inverses are analogous to the multiplicative inverse (or reciprocal) in the number system.
Remark.
- by the invertible matrix theorem (proof of its complete version is complicated, and so is skipped), if one of
and
holds, then the other also holds
In the number system, the multiplicative inverse, if it exists, is unique. Indeed, the matrix inverse, if it exists, is also unique similarly, as shown in the following proposition.
Proposition.
(Uniqueness of matrix inverse)
Matrix inverse, if it exists, is unique.
Example.
(Invertible matrix)
The matrix
is invertible and its inverse is
since
(it implies that the matrix product in another order is also
by the invertible matrix theorem)
Example.
(Non-invertible matrix)
The matrix
is non-invertible.
Proof.
Suppose to the contrary, that the matrix is invertible, i.e.
there exists a matrix
such that
But, this equality is equivalent to
which is impossible, causing a contradiction.
Remark.
- inductively, we can have general 'reverse multiplicativity':
is invertible and
Matrix inverse can be used to solve SLE, as follows:
Proof.
Then, we will define the elementary matrix, which is closely related to EROs, and is important for the proof of results related to EROs.
Remark.
- if a matrix needs to be obtained by performing two or more EROs on the identity matrix, it is not an elementary matrix
Example.
The matrix
is an elementary matrix of type I, since
it can be obtained by performing the ERO
on
,
the matrix
is an elementary matrix of type II, since
it can be obtained by performing the ERO
on
, and
the matrix
is an elementary matrix of type III, since
it can be obtained by performing the ERO
on
.
The matrix
is not an elementary matrix, since it needs to be obtained by performing at least two EROs on
, e.g.
, in this order
Proof.
Outline:
case: e.g.
- type I ERO:
- type II ERO:
Remark.
- illustration of the proposition:
Example.
The following EROs
correspond to the matrix multiplication
Proposition.
(Invertibility of elementary matrix)
Elementary matrices are invertible. The inverse of an elementary matrix is an elementary matrix of the same type.
Remark.
- if
is the RREF of
, then
for some elementary matrices 
- since elementary matrices are invertible,
is invertible, and equal to 
- in other words,
for some invertible matrix 
Then, we will state a simplified version of invertible matrix theorem, in which some results from the complete version of invertible matrix theorem are removed.
Proof.
To prove this, we may establish a cycle of implications, i.e.
(i)
(ii)
(iii)
(iv)
(i), then, when we pick two arbitrary statements form the four statements, they are equivalent to each other, which means that the four statements are equivalent.
(i)
(ii): it follows from the proposition about solving SLE, and
(ii)
(iii): since the SLE has a unique solution, the RREF of the augmented matrix of the SLE
has a leading one in each of the first
columns, but not the
st column, i.e. it is
. It follows that the RREF of
is
, since after arbitrary EROs, the rightmost zero column is still zero column.
(iii)
(iv): since RREF of
is
, and RREF of
equals
for some elementary matrices
, it follows that
. By definition and general 'reverse multiplicativity' of matrix inverse, we have
- i.e.
is a product of elementary matrices
(iv)
(i): since
is a product of elementary matrices and an elementary matrix is invertible, it follows that
is invertible by general 'reverse multiplicativity' of matrix inverse.
Remark.
- this theorem provides us multiple ways to prove invertibility of matrix: we can prove it by proving one of the equivalent statements
- this may make the proof easier
- later, when we discuss some results about these equivalent statements, they can be linked to this theorem
Example.
Consider the matrix
.
We can find its RREF by Gauss-Jordan algorithm, as follows:
Since its RREF is
,
by the simplified invertible matrix theorem, we also have the following results:
(i)
is invertible
(ii) the homogeneous SLE
only has trivial solution
(iii)
is a product of elementary matrices
Let's verify them one by one.
(i): Y
(ii): the SLE can be represented by the augmented matrix
, and we can find its RREF by Gauss-Jordan algorithm, as follows:
- Then, we can directly read from the RREF of augmented matrix that the SLE only has the trivial solution. Y
(iii):
Y
Exercise.
Consider the matrix
, and the SLE
.
The following provides us a convenient and efficient way to find the inverse of a matrix.
Example.
Let
.
After performing EROs as follows:
we have
.
We have previously proved that
is non-invertible.
Now, we verify that it is impossible to transform
to
in which
.
We perform EROs as follows:
The last matrix is in RREF.
We can see from the first ERO that, to make the
th entry zero, we will also make the
th entry zero. Thus, it is impossible to have such transformation.
Exercise.
Let
be some elementary matrices of same size
.
Determinants
Then, we will discuss the determinant, which allows characterizing some properties of a square matrix.
Remark.
- minor is determinant of a square submatrix
- the matrix
consisting of all cofactors is called cofactor matrix
- the definition when
is also called the cofactor expansion (or Laplace expansion) along the first row.
- another notation:
, and there are similar notations for matrices with different sizes
- the signs of the cofactors are alternating. the signs of cofactor at the position of each entry of a matrix is shown below:
- which looks like a 'chessboard' pattern.
- we can observe from the above pattern that the sign of cofactors located at the main diagonal are always positive
- this is the case since the row number
equals column number
in main diagonal, and so 
- (some illustrations of deleting the row and column)
Example.
(Formulas of determinants of
and
matrices)
and
For the formula of determinants of
matrices, we have a useful mnemonic device for it, namely the Rule of Sarrus, as follows:
Proposition.
(Rule of Sarrus)
We can compute
matrix as shown in the following image:
, in which red arrows correspond to positive terms, and blue arrows correspond to negative terms. To be more precise,
we can compute the matrix in the image by
Proof.
It follows from the formula in the above example.
Then, we will given an example about computing the determinant of a
matrix, which cannot be computed by the Rule of Sarrus directly.
Example.
Proposition.
(Determinant of zero matrix and identity matrix)
The determinant of the zero matrix is
, and the determinant of the identity matrix is
.
Indeed, we can compute a determinant by the cofactor expansion along an arbitrary row, as in the following theorem.
Remark.
- the first formula is cofactor expansion along the
th row, and the second formula is cofactor expansion along the
th column
Its proof (for the general case) is complicated, and thus is skipped.
Example.
(Illustration of cofactor expansion theorem)
We use cofactor expansion along the 2nd column here.
Exercise.
Let
.
Then, we will discuss several properties of determinants that ease its computation.
Proof.
Outline:
Remark.
- for the property related to type II ERO,
can be zero, and the determinant is multiplied by zero. But, multiplying the row by
is not type II ERO if 
- the determinant of matrix with two identical rows is zero because of the following corollary, which is based on the result about type I ERO
- in view of this proposition, we have some strategies to compute determinant more easily, as follows:
- apply type II EROs to take out common multiples of a row to reduce the numerical value of entries, so that the computation is easier
- apply type III EROs to create more zeros in entries
- apply cofactor expansion along a row or column with many zeros
- apart from the EROs mentioned in the proposition, we can actually also apply elementary column operations (ECOs)
- this is because determinant of matrix transpose equals that of the original matrix (it will be mentioned in the proposition about properties of determinants)
- so, applying elementary column operations is essentially the same as applying EROs, by viewing the operations in different perspectives
- we have similar notations for elementary column operations, with
(stand for row) replaced by
(stand for column)
Example. (Vandermonde matrix)
Corollary.
The determinant of a square matrix with two identical rows is zero.
Then, we will introduce a convenient way to determine invertibility of a matrix.
Before introducing the theorem, we have a lemma.
Lemma.
For each elementary matrix
and matrix
,
Theorem.
(Determining invertibility by determinant)
A square matrix is invertible if and only if its determinant is nonzero.
Proof.
- only if part: by simplified invertible matrix theorem, a matrix
is invertible is equivalent to
is product of elementary matrices. So, if we denote the elementary matrices by
, then
- if part: Let
in which
are elementary matrices and
is the RREF of
. This implies that
- Since
, so
. Thus,
has no zero row (its determinant is zero otherwise). Since
is in RREF, it follows that
(since
is square matrix, if not all columns contain leading ones, then there is at least one zero row lying at its bottom, by definition of RREF). By simplified invertible matrix theorem,
is invertible.
After introducing this result, we will give some properties of determinants which can ease the computation of determinants.
Proof.
- (multiplicativity) let
in which
are elementary matrices and
is the RREF of
. Then,
- and
- then, it remains to prove that

- if
, then 
- if
, then the last row of
is a zero row, so 
- the last row of
is also a zero row, so 
- (invariance of determinant after transpose) we may prove it by induction and cofactor expansion theorem, e.g.
vs. 
- (determinant of matrix inverse is inverse of matrix determinant) using multiplicativity,
(
since
is invertible)
Then, we will introduce adjugate of matrix, which has a notable result related to computation of matrix inverse.
Theorem.
(Relationship between adjugate and determinant)
Let
be an
matrix. Then,
Proof.
The proof is complicated, and so is skipped.
Corollary.
(Formula of matrix inverse)
If
is invertible, its inverse is given by
Proof.
Example.
(Formula of
matrix inverse)
Let
. Then,
That is, we can find the inverse of a
matrix by interchanging the
-th and
th entries, multiplying the
th and
th by
(without interchanging them), and multiplying the matrix by the reciprocal of its determinant.
Example.
(Adjugate of non-invertible matrix)
Let
. Then,
Also, we have
Then, we will introduce a result that allows us to compute the unique solution of SLE directly, namely Cramer's rule.
Theorem.
(Cramer's rule)
Let
be a SLE in which
is an invertible
matrix and
(we use this notation to denote
, a transpose of
matrix).
Let
, and let
be the determinant of the matrix obtained by replacing the
th column of
by the column
for each
. The unique solution to the SLE is given by
Proof.
Since
is invertible, the unique solution of the SLE is
.
Using the formula of matrix inverse, we have
Thus, for each
,
(
are entries at the
th row of
(and at the
th column of the cofactor matrix of
),
so multiplying the entries as above gives the
-th entry of
, namely
)
Example.
Consider the SLE
Since
the unique solution to this SLE is
Exercise.
Solve the SLE
.
Solution.
Since
,
the matrix
is non-invertible.
Thus, we cannot use Cramer's rule.
Instead, we can transform the augmented matrix representing the SLE to RREF, as follows:
Since there is a leading one at the 4th column, the SLE is inconsistent.