Proof
Using equality in the sequence of null spaces will give the final answer. We just have to show . Suppose it wasn't true. Then by the two previous lemmas we have:
The dimension then becomes:
At each step of the way the dimension must increase by at least 1 (because we have instead of 's and thus rather than ). Thus, after steps that implies that which is a contradiction since it's a subspace of with dimension .
☐
While in general it's not true that the next could be used as a substitute:
is a direct sum of and
Suppose . Let . Then:
Proof
First we show the intersection is trivial. Suppose . So then and where . Apply to both sides to get that . Thus by the fact that null spaces stop growing. Thus so then the intersection is trivial.
Since the intersection is trivial, then the sum is a direct sum. Also:
More often than we desire, we don't have enough eigenvectors to lead to diagonalization. We need something more general. Let's examine this issue by fixing . We which we can describe by direct sums of simpler spaces:
where each is a subspace of invariant under . The simplest 's would be 1-dimensional, but this is only possible iff has a basis consisting of eigenvectors of iff has an eigenspace decomposition:
where are all distinct eigenvalues of . The Chapter 7 - Operators on Inner Product Spaces#^6995f7 says if is an inner product space then a decomposition of the form of the equation above holds for every normal operator if and for every self-adjoint operator if as there's enough eigenvectors in these operators to form a basis of .
But the form above may not always hold, so we need to expand our definition of eigenvectors.
generalized eigenvector
Suppose and is an eigenvalue of . A vector is called a generalized eigenvector of corresponding to if and:
for some .
Although is arbitrary, we will soon prove that every generalized eigenvector satisfies this equation with .
generalized eigenspace,
Suppose and . The generalized eigenspace of corresponding to , denoted , is defined to be the set of all generalized eigenvectors of corresponding to , along with the .
Because every eigenvector of is a generalized eigenvector of (take in the definition), each eigenspace is contained in the corresponding eigenspace (ie: ). The next result implies that is a subspace of since the null space of each linear map on is a subspace of .
Description of generalized eigenspaces
Suppose and . Then .
Proof
Suppose . Using the definitions where , so then we have .
Using the definition of eigenvalue shows that 's eigenvalues are 0 and 5. The corresponding eigenspaces are and . There's clearly not enough eigenvectors to span the vector space .
An operator is called nilpotent if some power of it equals .
For instance, the operator defined by:
is nilpotent because .
Nilpotent operator raised to dimension of domain is
Suppose is nilpotent. Then
Proof
Because is nilpotent, . Thus by the description of generalized eigenvectors then .
☐
Given an operator on we want to find a basis of such that the matrix of with respect to this basis is a simple as possible, so hopefully has a lot of 0's.
Matrix of a nilpotent operator
Suppose is a nilpotent operator on . Then a basis of w.r.t. which the matrix of has the form:
so all entries on and below the diagonal are 0's.
Proof
First choose a basis of . Then extend this to a basis of . then extend to a basis of . Continue in this fashion, eventually getting a basis of since .
Now let's think about the matrix of w.r.t. this basis. The first column consists of all 0's since the corresponding basis vectors are in . Applying to any such vector we get a vector in which is a vector that's a linear combination of the previous basis vectors. Thus all nonzero entries in these columns lie above the diagonal. The next set of columns comes from the basis vectors in . Applying to any such vector, we get a vector in ; in other words we get a vector that is a linear combination of the previous basis vectors. Thus once again, all nonzero entries in these columns lie above the diagonal. Continue in this fashion to get the result.
☐
8.B: Decomposition of an Operator
Description of Operators on Complex Vector Spaces
We saw prior that we may lack enough eigenvectors to have a deconstruction of onto those eigenspaces. But we observed that has it that and are invariant under via the nullspace proof and the range proof. Now we can show that the null space and the range of each polynomial of is also invariant under :
The null space and range of are invariant under
Suppose and . Then and are invariant under
Proof
Suppose , so then . Thus:
Hence so it's invariant under .
Suppose , so where . Then:
So clearly , so it's invariant under .
☐
The following result shows that every operator on a complex vector space can be though of as composed of pieces, each of which is a nilpotent operator plus a scalar multiple of the identity.
Description of operators on complex vector spaces
Suppose is a complex vector space and . Let be the distinct eigenvalues of .
Suppose is a complex vector space and . Then there is a basis of consisting of generalized eigenvectors of .
Proof
Choose the basis of each via our previous lemma. Put all the bases together to form a basis of consisting of generalized eigenvectors of . This is more specifically covered via this in-depth look.
☐
Multiplicity of an Eigenvalue
If is a complex vector space and , then the decomposition of provided by the description of operators on complex vector spaces can be a power tool. The dimensions of the subspaces involved are sufficiently important. So much so that they get their own name:
multiplicity
Suppose . The multiplicity of an eigenvalue of is defined to be the dimension of the corresponding generalized eigenspace . In other words, the multiplicity of an eigenvalue of equals .
To interpret our results in matrix form, we make the following definition, generalizing the notion of a diagonal matrix. In the case all are then we actually have a diagonal matrix.
block diagonal matrix
A block diagonal matrix is a square matrix of the form:
where are square matrices lying along the diagonal and all the other entries of the matrix equal 0.
For instance, the matrix:
is a block diagonal matrix of the form:
Block diagonal matrix with upper-triangular blocks
Suppose is a complex vector space and . Let be the distinct eigenvalues of , with multiplicities . Then there is a basis of with respect to which has a blcok diagonal matrix like seen above:
Putting the bases of the 's together gives a basis of via operators on vector spaces (a). The matrix of with respect to this basis has our desired form.
☐
For example, suppose is defined by:
We have:
We found that:
We saw that the basis of was:
The matrix of with respect to this basis is:
Square Roots
Recall a square root of an operator is an operator such that via its definition. Every complex number has a square root, but not every operator on a complex vector space has a square root.
For example, the operator in given by this problem has no square root. The non-invertibility has something to do with it. But first, we'll show that the identity plus any nilpotent operator has a square root.
Identity plus nilpotent has a square root
Suppose is nilpotent. Then has a square root.
Proof
Consider the Taylor series for the function :
We will not find an explicit formula for the coefficents or worry about convergence because we just want to use this equation only for motivation.
Because is nilpotent, then for some positive integer . In the equation for up top, suppose we replace with and with . Then the infinite sum on the right side becomes finite since all :
Having made this guess, we try to choose such that the operator above has its square equal to . Just apply this squaring process:
We want the right side of the equation to equal . Thus, notice we should choose where . Next, choose such that . Then choose such that . Continue in this manner for .
We don't actually care the formula for each , we just need to know that for each we can make a correct choice of to make a square root of .
☐
The previous result works on real as well as complex vector spaces. However, the next result holds only on complex vector spaces. For example, the operator of multiplication by on the -dimensional real vector space has no square root.
Over , invertible operators have square roots
Suppose is a complex vector space and is invertible. Then has a square root.
for each . Clearly is nilpotent, so has a square root using our freshly proved lemma. Multiplying a square root of the complex number by a square root of we obtain a square root of (see the equation above). A typical vector can be written uniquely in the form:
The operators on the right side of the equation all commute, so move out the factor to the last term, and since we can conclude that as desired.
☐
The Minimal Polynomial
monic polynomial
A monic polynomial is a polynomial whose highest-degree coefficient equals 1.
For instance, the polynomial is a monic polynomial of degree 7 (since the highest degree coefficient is 1).
Minimal Polynomial
Suppose . Then there is a unique monic polynomial of smallest degree such that .
Proof
Let . Then the list:
is not linearly independent in because the vector space has dimension Chapter 3 - Linear Maps#^ee27e8 while we have a list of length . Let be the smallest positive integer such that:
is LD. The Linear Dependence Lemma implies that one of the operators above is a linear combination of the others, and because was chosen to be the smallest positive integer where the above list is LD, then we conclude that is a linear combination of , so where:
Define the monic polynomial by:
Plugging in into the polynomial and using our our equation above, we conclude that .
To show uniqueness of , note that choice of implies that no monic polynomial with degree smaller than can satisfy . Suppose is a monic polynomial with degree and . Then and . The choice of now implies that .
☐
minimal polynomial
Suppose . Then the minimal polynomial of is the unique monic polynomial of smallest degree such that .
Using the fact that minimal polynomials have at most this implies that each operator on has a minimal polynomial of, at most, . The Cayley-Hamilton Theorem tells us that if is a complex vector space then the minimal polynomial of each operator on has degree at most . This remarkable improvement also holds on real vector spaces.
Programming a computer
Suppose you are given , and thus . You can program a computer to find the minimal polynomial of by considering that:
for values until the system of equations has a solution which then are the coefficients of the minimal polynomial of . This process itself can be done via Gaussian elimination or similar methods.
Example
Let be the operator on 𝟝 whose matrix is w.r.t. the standard basis:
The minimal polynomial can be calculated using the powers of of increasing : To save you some trouble, there's no solution until in this case, where then:
where then solving quickly gives that so then the characteristic polynomial is .
implies is a multiple of the minimal polynomial.
Suppose and . Then iff is a polynomial multiple of the minimal polynomial of .
Proof
Let denote the minimal polynomial of .
First we prove . Suppose is a polynomial multiply of . Thus such that . Thus:
Which implies that since otherwise dividing by its highest degree coefficient would produce a monic polynomial that, when applied to , gives ; this polynomial would have a smaller degree than the minimal polynomial, creating a contradiction.
Thus so is a polynomial multiply of as desired.
☐
Characteristic polynomial is a multiple of minimal polynomial
Suppose and . Then the characteristic polynomial of is a polynomial multiple of the minimal polynomial of .
Proof
Use Cayley-Hamilton to get as the characteristic polynomial where , then use our proved lemma to show that is a multiple of the minimal polynomial of .
☐
Eigenvalues are the zeroes of the minimal polynomial
Let . Then the zeroes of the minimal polynomial of are precisely the eigenvalues of .
Proof
Let:
be the minimal polynomial of .
First to prove suppose is a zero of . Then can be written as:
for some monic polynomial . Because then:
Because then such that . This is an eigenvector of the transformation above as when applied above. Thus must be an eigenvalue of .
To prove () suppose is an eigenvalue of . Thus where . Apply to both sides to show that for all . Thus:
Since then as desired.
☐
Let's do some examples:
Example
Find the minimal polynomial of given:
Proof
We found that had multiplicity and had multiplicity , so then the characteristic polynomial is:
The minimal polynomial is either or just the characteristic polynomial itself. To check, we check if when we reduce degree. Notice:
Thus the minimal polynomial is just .
☐
Example
Find the minimal polynomial of the operator defined by .
Proof
Clearly with and where . For later computations notice that:
Thus consider if is the minimal polynomial:
Thus is the minimal polynomial.
☐
8.D: Jordan Form
We know if is a complex vector space then all has a basis for where is upper triangular. For this section, we can add more 0's to that diagonal. To show what we mean, consider the matrix for a nilpotent operator where:
Where we have as our basis, giving:
The idea here is that we can always have a diagonal of eigenvalues (in this case all 0's) and then block diagonal "submatrices" that are also block diagonal (have only eigenvalues with submatrices that are block diagonal). As another example, the block diagonal matrix:
Here all the submatrices are block diagonal, making the whole matrix still block diagonal. To interpret this in the general case, we use the following proof:
Basis corresponding to a nilpotent operator
Suppose is nilpotent. Then and such that:
is a basis of
Proof
While Axler does an induction over , I do like the usage of quotient spaces used in Lecture 28 - Continuing Jordan Block Decomposition. I feel it works at the whole "process" of how this gets derived, rather than going through the motions of an induction proof. While it's not a "proof" per say, the findings are easily generalized.
☐
We want to define this "form" of a matrix, as well as the basis required to make it.
Jordan Basis
Suppose . A basis of is called a Jordan basis for if w.r.t. this basis then has a block diagonal matrix:
where each is an UT matrix of the form:
Jordan Form
Suppose is a complex vector space. If then there is a basis of that is a Jordan basis for .
Proof
First consider nilpotent operator and the vectors are as they are given given the basis for a nilpotent operator. For each note that sends the first vector in the list:
to and that sends each vector in this list other than the first to the previous vector. In other words, the basis:
is a basis for which is now block diagonal.
Now suppose . Let be the distinct eigenvalues of . We have the generalized eigenspace decomposition:
where each is nilpotent via the description of operators on complex vector spaces. Thus, some basis of each is a Jordan basis for this transformation by reapplying our logic above, using . Thus, putting the bases together gets a basis of that is a Jordan basis for .
☐