Chapter 5 - Eigenvalues, Eigenvectors, and Invariant Subspaces
5.A: Invariant Subspaces
Suppose . If we have a direct sum decomposition:
where each is a proper subspace of , then to understand then we only need to understand each , restricting to our domain for a subspace . But the problem is that may not map back to itself, a requirement for being an operator (see Chapter 3 - Linear Maps#Operators as to why this is the case). So we must only consider decompositions of that allow that property to arise.
invariant subspace
Suppose . A subspace of is an invariant under if implies that .
This allows to be invariant when is an operator on .
Some examples of invariant subspaces under over :
since and
since if then already
since if then
since if then .
So the big question is if has invariant subspaces that aren't the ones above, as these are the so-called trivial invariants. Notice that this is because may very well be the first set and likewise is also a possibility.
Eigenvalues/vectors
Let's look at invariant subspaces of dimension 1, the simplest one we know.
Take any where and let be:
Then is a 1-dimensional subspace of . If is invariant under and operator then and thus there is some scalar such that:
Conversely, if for some then is a 1-dimensional subspace of invariant under .
eigenvalue
Suppose . A number is called an eigenvalue of if there exists such that and .
Thus has a 1-dimensional invariant subspace iff has an eigenvalue.
Lemma
Suppose is finite-dimensional and and . Then the following are equivalent.
If and is a subspace of invariant under , then determines two other operators and :
and
Suppose and is a subspace of invariant under
The restriction operator is defined by:
for .
The quotient operator is defined by:
for .
Here we have it that is the set . Notice if then . Suppose then . To be honest, we're not covering these, so see section 3.D and 3.E later on when I get to there to use these in more detail.
5.B: Eigenvectors and Upper-Triangular Matrices
Polynomials Applied to Operators
Suppose and is a positive integer.
is defined by
is defined to be the identity operator on .
If is invertible with inverse then is defined by: .
Suppose and is a polynomial given by:
for . Then is the operator defined by:
If we fix then the function from to given by is linear.
Suppose and is a basis of . The matrix of with respect to this basis is the matrix:
whose entries are defined by:
If the basis is not clear from context, is used instead.
A really simple basis to have is to have the first vector be nothing but zeroes, after a non-zero term , then repeating this process:
If is finite-dimensional and a complex vector space, we know that an eigenvalue like exists, with its associated eigenvector that we use as the first vector of the basis. We can then repeat with the smaller matrix with the first row and column removed, and get the same thing, all the way down!
diagonal of a matrix
The diagonal of a square matrix consists of the entries along the line from the upper left corner to the bottom right corner.
upper-triangular matrix
A matrix is called upper-triangular if all the entries below the diagonal equal 0.
Typically they have the shape of:
Conditions for upper-triangular matrix
Suppose and is a basis of . Then the following are equivalent:
The matrix of () with respect to is upper-triangular
Proof
(1) equals (2) from the definitions. (3) implies (2) is easy to show, so to finish we only prove (2) implies (c).
Suppose (2). Fix . From (2) we know that:
Thus if is a linear combination of then:
So then is invariant under .
☐
Now we really want to show that each operator on a finite-dimensional complex vector space has a matrix of the operator with only 0's below the diagonal.
Over , every operator has an upper-triangular matrix
Suppose is a finite-dimensional complex vector space and . Then has an upper-triangular matrix with respect to some basis of .
See the proof at Lecture 24 - Finishing Eigenstuff#^98fa4c. Notice that to construct such a proof, one usually creates it from the back, then proves it forwards. In this case, seems arbitrary, but notice that it's a good choice because the bottom right matrix at the end of the proof specifically is guaranteed to have 's on the diagonal.
Determination of invertibility from upper-triangular matrix
Suppose has an upper triangular matrix with respect to some basis of . Then is invertible iff all the entries on the diagonal of that upper-triangular matrix are non-zero.
This will be used for a really important lemma, so hold onto your butts.
Determination of eigenvalues from upper-triangular matrix
Suppose has an upper-triangular matrix with respect to some basis of . Then the eigenvalues of are precisely the entries on the diagonal of that upper-triangular matrix.
Suppose and . The eigenspace of corresponding to , denoted , is defined by:
In other words, is the set of all eigenvectors of corresponding to , along with the vector.
Notice that for and , the eigenspace is a subspace of because the null space of each linear map on is a subspace of . Thus, these definitions imply that is an eigenvalue of iff .
For example, consider the matrix above. Here:
Now notice that if we restrict to just , then all that happens is vectors get scaled:
Sum of eigenspaces is a direct sum
Suppose is finite-dimensional and . Suppose also that are distinct eigenvalues of . Then:
An operator is diagonalizable if the operator has a diagonal matrix with respect to some basis of .
For instance, consider where:
The matrix of with respect to the standard basis of is:
which isn't diagonal. But is diagonalizable, since with respect to the basis is:
which is diagonal.
Conditions equivalent to diagonalizability
Suppose is finite-dimensional and . Let denote the distinct eigenvalues of . Then the following are equivalent:
is diagonalizable;
has a basis consisting of eigenvectors of ;
There exist 1-dimensional subspaces of , each invariant under , such that:
;
Proof
(a) = (b). An operator has a diagonal matrix with respect to basis iff for each . Thus, (a) and (b) are equivalent above.
(b) (c). Suppose (b). Then has a basis consisting of eigenvectors of . For each , let . Each is 1-dimensional subspace that is invariant under . Because is a basis of , each vector in can be written uniquely as a linear combination of . Hence, each vector in can be written uniquely as a sum , where each . Thus, , so (b) implies (c).
(c) (b). Suppose (c); so there are 1-dimensional subspaces of , each invariant under , such that . For each , let be a non-zero vector in . Then each is an eigenvector of , as is 1-dimensional. Hence, because each vector in is uniquely written as a sum where each (so each is a scalar multiple of ), we see that is a basis of . So (c) implies (b).
We know that (a,b,c) are all equivalent. We finish showing (b) implies (d), (d) implies (e), and (e) implies (b)
Suppose (b) holds; thus has a basis consisting of eigenvectors of . Hence, every vector in is a linear combinations of eigenvectors of , which means that:
Choose a basis of each , put all these bases together to form a list of eigenvectors of , where via our equation of dimension above. To show is LI, suppose:
For each let denote the sum of all the terms such that . Thus each is in , and:
Because eigenvectors corresponding to distinct eigenvalues are LI via Chapter 5 - Eigenvalues, Eigenvectors, and Invariant Subspaces#^45cb5f, this implies each . Because each where the 's are chosen to be a basis of , this implies that all . Thus is LI, and hence a basis of (since we have the right number of vectors). Thus (e) implies (b).
☐
Okay that was a long proof. Feel free to digest this, then move onto the next thing.
The sad thing though is that not all have diagonalizable matrices . For example, if where:
is not diagonalizable. is the only eigenvalue of and furthermore , so all (b - e) in our theorem above fail, so (a) of it fails and thus isn't diagonalizable.
However, we can guaruntee that we can be diagonalizable if we follow the following lemma:
Enough eigenvalues implies diagonalizablility
If has distinct eigenvalues, then is diagonalizable.
Proof
Suppose has distinct eigenvalues . For each , let be eigenvector corresponding to the eigenvalue . Because eigenvectors corresponding to distinct eigenvalues are LI via Chapter 5 - Eigenvalues, Eigenvectors, and Invariant Subspaces#^45cb5f, then is LI. A LI list of is a basis of , so is a basis of . With respect to this basis consisting of eigenvectors, has a diagonal matrix.
☐
Note that the converse is not true; namely, we could have not all distinct eigenvalues while still being diagonalizable.