In this lecture, we’ll look into the inverse of a matrix, and find out which matrices are invertible. These notes roughly follow Trefethen and Bau, section 1.
Let . Recall that
\text{range}(\mA) = \text{span}(\va_1, \ldots, \va_n),\text{rank}(\mA) = \text{dim}(\text{range}(\mA)).
We call full rank if .
Full rank matrices have the important property that they give rise to one-to-one maps between and . Let’s show this.
Theorem (From Trefethen and Bau, Theorem 1.2) Let , a matrix is full rank if and only if it maps no two distinct vectors to the same vector.
Proof If a matrix is full rank, then it has linearly independent column vectors. These vectors are a basis for . This, in turn, implies that any vector in has a unique representation in this basis. (If not, then and so has linearly dependent columns, which it can’t!) Thus, any vector corresponds to a unique .
We also have to prove the reverse direction, but this is easier to prove via the contrapositive. If is not full rank, then it’s columns are linearly dependent. Hence, there exists a vector such that . Let be any vector in , then and so we have two distinct vectors that give us the same result.
There’s a great picture I could put here, but it’s too tricky. The point is that we have a one-to-one map between and , which is a subset of when . Because this map is one-to-one, it’s invertible! So we can take any vector and find
\mA \vx = \vb
for some .
It’s worth repeating this equation because it’s so fundamental to the rest of the class – and the entire field.
We call
\mA \vx = \vb
a linear system of equations.
Usually these are defined with squares matrices .
Let be a full-rank matrix. What we’ve shown above is that any vector in can be written as for some unique .
Thus, we can find the following vectors:
\mA \vx_i = \ve_i \qquad i = 1, \ldots, n.
If we write this as a matrix equation, we have:
\mA \mX = \mI.
The matrix is called the inverse and usually written .
We’ve shown that and full rank has an inverse such that
\mA \mA^{-1} = \mI.
Let’s study a few properties of this inverse.
First, does too? We’ll show this is the case. Let and let . Then
\mY \mA \mX = (\mY \mA) \mX = \mX,
but also
\mY \mA \mX = \mY (mA \mX) = \mY.
Thus, .
Second, assuming that and are square. The standard way to check that you have the inverse of a particular matrix is just to show that it satifies . In this case:
(\mA \mB) (\mB^{-1} \mA^{-1}) = \mA (\mB \mB^{-1}) \mA^{-1} = (\mA \mA^{-1} = \mI.
where we’ve canceled all the inverse pairs represented by parentheses.
The following set of statements from Trefthen and Bau helps to characterize when a matrix has an inverse. Let , these statements are all equivalent to each other:
Let be full-rank. Then the linear system:
\mA \vx = \vb
has solution
\vx = \mA^{-1} \vb.
Be warned, this is not a good way to find unless is very special. In this class, we will see many superior procedures to find that satifies this linear system. A good way to demonstrate that you have not learned the material is to utilize this idea in your programs.