For a square matrix
, to diagonalize
means to find
a square matrix
and diagonal matrix
with
.
In this case we say that
diagonalizes
. Notice that
has to be nonsingular or it won't have an inverse.
Some matrices are ``diagonalizable'' and others are not.
For
, it turns out that
is a good matrix to diagonalize
.
In fact,
for
.
Notice that
and
are eigenvectors
of
. In fact,
is always made from eigenvectors:
Proposition. For
matrices
and
,
diagonalizes
the columns of
are
linearly independent eigenvectors of
and the diagonal
entries of
are the corresponding eigenvalues.
(The linear independence of columns is needed in order for
to be nonsingular. It wouldn't work, for instance, to have all
columns be the same. For the reasoning behind the Proposition,
see below.)
The Proposition tells how to diagonalize
: Just find the
eigenvalues and corresponding eigenvectors. If there are enough
linearly independent eigenvectors, use them for the columns
of
. The Example could be found that way.
Handy facts:
(1) Eigenvectors that belong to distinct eigenvalues are linearly
independent. Therefore, if
is
and has
distinct eigenvalues, then
is diagonalizable. (``Distinct''
means ``all different from one another''.)
(2) A real symmetric matrix can always be diagonalized, even
if the eigenvalues are not all distinct (for example, if the
characteristic polynomial factors as
), we would say the eigenvalues are
, not
all distinct.
These two facts will be proved in class. Let's go back and prove the Proposition.
Proof of the Proposition. For ``
'': To says that
diagonalizes
means that
. If
we take this equation and multiply both sides by
on the
left, we get
. Now let's try to interpret this second
equation using eigenvalues and eigenvectors. We need two facts:
For example,
and it doesn't matter what the
other entries are.
For example,
.
We still need to prove ``
'' in the Proposition. We start
by assuming we are given
and
such that the columns of
are eigenvectors of
and are linearly independent.
The same two facts as before show that
. Since the columns
of
are linearly independent,
has rank
and is
invertible. Multiply both sides of
on the left by
;
we get
, as required.
Problem
U-12. Diagonalize
. (Use Problem U-
.)
Problem
U-13. Explain how to interpret the equation
in
terms of a change of basis. What is the transformation? The old basis?
The matrix relative to the old basis? The new basis? The transition matrix?
The matrix relative to the new basis?