HW 6 - Finishing UT Matrices, Eigenspaces and Diagonal Matrices

5.B: Eigenvalues/Vectors and UT Matrices

14

We'll give an example of an operator who's matrix with respect to some basis contains only 0's on the diagonal, while the operator is still invertible. By the question's own admittance, we should consider non-U.T matrices for this.

Consider:

[0110]

Here all the diagonals are 0's. For the sake of simplicity let V=R2, and β is the standard basis of i^,j^. Then since M(T) is as described above then:

T(i^)=0i^+1j^=j^T(j^)=1i^+0j^=i^

So define T(v) for v=(v1,v2)R2 as:

T(v1,v2)=(v2,v1)

notice that M(T) is as described above, but the inverse matrix M(T)1 I claim is:

M(T)1=M(T)

Note that I can show this because:

M(T)1=M(T1)

and I can show that:

T1(a,b)=(b,a)=T(a,b)

since if I claim this I can verify that:

(TT1)(a,b)=T(b,a)=(a,b)=I(a,b)TT1=I

and:

(T1T)(a,b)=T1(b,a)=(a,b)=I(a,b)T1T=I

so then T is invertible.

15

For simplicity let V=R2 and let β be the standard basis. Consider the matrix:

M(T)=[1111]

Here the entries in the diagonal are all non-zero, but T is not invertible as we'll show. Since TL(V) we need to show that T isn't surjective or injective (just one). We'll do injectivity. To show it's not injective, we need to find two vectors v,uV such that Tu=Tv but uv. Notice first that:

T(1,0)=1(1,0)+1(0,1)=(1,1)T(0,1)=1(1,0)+1(0,1)=(1,1)=T(1,0)

Hence, choose u=(1,0) and v=(0,1). By the above showing, notice that Tu=(1,1) and:

Tv=T(0,1)=(1,1)=Tu

but uv, showing that T isn't invertible.

16

Operators on complex vector spaces have an eigenvalue

Every operator on a finite-dimensional, nonzero, complex vector space has an eigenvalue.

We'll prove the above lemma using the below lemma:

Lemma

If V,W is finite dimensional where dim(V)>dim(W) then no linear map from VW is injective.

Proof
Let TL(V) be arbitrary, where V is:

Consider the transformation τ:Pn(C)V, where τ is defined by:

τ(p(z))=p(T)vV

We should show that τ is itself linear. Let p,qPn(C) be arbitrary. Notice that for additivity:

τ(p+q)=(p(T)+q(T))v=p(T)v+q(T)=τ(p)+τ(q)

and similarly for homogeneity:

τ(αp)=(αp)(T)v=αp(T)v=ατ(p)

thus τ is linear. Notice that dim(V)=n<n+1=dim(Pn(C)) so then we can use the aforementioned lemma to say that τ isn't injective. Thus, there's some vectors p,qPn(C) such that:

τp=τqpq

Thus implying that a new polynomial r=pq is where:

τ(pq)=0=τr

where notice that pq so r0. Namely, notice that r is a polynomial of degree n or less. Denote deg(r)=mn. Then:

r(z)=c(zλ1)(cλm)

But then:

τ(r(z))=r(T)vV

as we defined τ above. Therefore:

0=τ(r)=r(T)v=c(Tλ1I)(TλmI)vV

But then clearly v is an eigenvector for at least one (TλiI) where i{1,...,m}. So then v has some eigenvalue λi associated with the transformation T, completing the theorem.

20

Theorem

Suppose V is a finite-dimensional complex vector space and TL(V). Then T has an invariant subspace of dimension k for each k=1,...,dim(V).

Proof
For simplicity, let dim(V)=n. Fix k=1,...,dim(V) as arbitrary. Let v1,...,vn be the basis vectors that spans V.

Since V is a finite-dimensional complex vector space and TL(V), then T has an upper triangular matrix with respect to some basis of V, which we use as β where β={v1,...,vn}. As such, then the matrix of M(T) with respect to v1,...,vn is U.T.

But this is equivalent to saying that span(v1,...,vj) is invariant under T for each j=1,...,n. As such, for whatever k we've fixed, have:

U=span(v1,...,vk)

as per the reasoning above, then U is invariant under T because of our conditions under V, so then we can always find U for each k we've chosen. Furthermore, clearly U is k-dimensional as v1,...,vk are all LI, so then any uU must be composed of these LI vectors, namely k of them.

5.C: Eigenspaces and Diagonal Matrices

1

Theorem

Suppose TL(V) is diagonalizable. Then V=null(T)range(T).

Proof
Let TL(V) be diagonalizable. As such, then V has a basis consisting of eigenvectors of T, we denote as v1,...,vn (where dim(V)=n is implicitly defined). As such, consider each Tvj for any j. Since each vj is an eigenvector of T, we always get:

Tvj=λjvj

Now, consider the cases where λj=0 or not. If λj=0, then Tvj=0 so vjnull(T). Notice though that vjrange(T) since vj is an eigenvector, so the only vector where Tw=vj would be 1λjvj since:

T(1λjvj)=vj

But that would contradict λj=0.

If instead λj0 then Tvj=λjvj so then clearly vjrange(T)={Tv:vV}. Notice though that Tvj0 since both λj0 as well as vj0 since it's an eigenvector.

Since v1,...,vn is a basis and T is linear, then any vector v where v is in the nullspace cannot be in the range, and vice versa, of T. As such, then both sets are disjoint. It's clear that V=null(T)+range(T) by the FTOLM, but furthermore we know that since they're disjoint subsets whose sum makes up V that then this must be a direct sum.

3

Theorem

Suppose V is finite-dimensional and TL(V). The following are equivalent:

  1. V=null(T)range(T)
  2. V=null(T)+range(T)
  3. null(T)range(T)={0}

Proof
It's clear that if (1) is true then (2, 3) are both true via the definition of a direct sum as well as Chapter 1 - Vector Spaces#^038942.

Suppose (2). Notice if we prove (3) then (1) comes for free. By the FTOLM then:

dim(V)=dim(null(T))+dim(range(T))

But since null(T),range(T) are subspaces of V, then we also have it that:

dim(V)=dim(null(T))+dim(range(T))dim(null(T)range(T))

Thus subtracting both equations, it's clear that dim(null(T)range(T))=0 so then null(T)range(T)={0}, proving (3), thus also getting (1) as well.

Suppose (3). Notice that if we prove (2) then (1) comes for free. Since null(T) and range(T) are subspaces of V then:

dim(V)=dim(null(T))+dim(range(T))dim(null(T)range(T))

But via our supposition then we know that the dimension of the negative term is 0, so:

dim(V)=dim(null(T))+dim(range(T))

As such, then we can represent any vector vV by a set of basis vectors u1,...,um,w1,...,wk where m is the dimension of the null space, and k is the dimension of the range (of T). Thus:

v=α1u1++αmum+β1w1+βkwk=u+w

where unull(T) and wrange(T) so then since v was arbitrary it follows that V is at least a normal sum of the null-space and range (of T).

6

Theorem

Suppose V is finite-dimensional, TL(V) has dim(V)=n distinct eigenvalues, and SL(V) has the same eigenvectors as T (not necessarily with the same eigenvalues). Then ST=TS.

Proof
Suppose the suppositions in the theorem. For clarity, there's λ1,...,λn distinct eigenvalues for T and thus has corresponding eigenvectors v1,...,vn. Furthermore, since S has the same eigenvectors as T, then v1,...,vn are eigenvectors of S. We may have different λ's so then these have corresponding eigenvalues λ1,...,λn.

Note here that since we have v1,...,vn that these can form a basis of V, as there's n vectors and they are all LI with each other (since they're eigenvectors, otherwise they'd share λ's and thus λ wouldn't be distinct). And note that for each j that Tvj=λjvj and furthermore Svj=λjvj.

Consider any vector vV. It can be composed into a linear combination of it's basis vectors:

v=α1v1++αnvn

Thus notice that:

STv=S(α1Tv1++αnTvn)=S(α1λ1v1++αnλnvn)=α1λ1Sv1++αnλnSvn=α1λ1λ1v1++αnλnλnvn

and likewise:

TSv=T(α1Sv1++αnSvn)=T(α1λ1v1++αnλnvn)=α1λ1Tv1++αnλnTvn=α1λ1λ1v1++αnλnλnvn=α1λ1λ1v1++αnλnλnvn=STv

Thus since v was arbitrary it follows that ST=TS.

7

Theorem

Suppose TL(V) has a diagonal matrix A with respect to some basis of V and that λF. Then λ appears on the diagonal of A precisely dim(E(λ,T)) times.

Proof
Since M(T)=A is a diagonal matrix, then

A=diag(λ1,...,λn)

where λi may or may not be distinct. Clearly since A is n×n then:

n=dim(V)=dim(E(λ1,T))++dim(E(λm,T))

where we take λ1,...,λn and consider just the distinct eigenvalues λ1,...,λm eigenvalues.

Let's prove this via induction over mn., where we consider the arbitrary λ{λ1,...,λm}. Consider m=1. Then:

n=dim(λ1,T)

This implies that λ=λ1 is the only distinct eigenvalue from λ1,...,λn, so then:

A=diag(λ1,...,λ1)

and since λ1 appears n times (since A is n×n), then that's the same as dim(λ1,T) as we needed.

Now suppose that the theorem holds for any k<m. Consider the k-th case. Here we have:

n=dim(λ1,T)++dim(λk,T)

Notice that since the k1 case is true, then if λ{λ1,...,λk1} then by the inductive hypothesis then we have λ on the diagonal dim(E(λ,T)) times as expected. If instead λ=λk then we know that:

dim(λk,T)=ni=1k1dim(λi,T)

clearly we know that λλi for i=1,...,k1, so then by the inductive hypotheses for each λi where λiλk that we will get dim(λi,T) times on the diagonal for A. As such, if a total of n times will appear for all λ's and we subtract the other dim(λi,T) times, then we get the equality above, showing that therefore for λk we expect an appearance of dim(λk,T) times, as expected.

Thus, via the principle of mathematical induction, the theorem holds.

8

Theorem

Suppose TL(F5) and dim(E(8,T))=4. Then T2I or T6I is invertible.

Proof
Notice that dim(V)=5 in this case, and since we have the dimension above then:

dim(V)=5=dim(E(8,T))++dim(E(λ,T))5=4++dim(E(λ,T))1=+dim(E(λ,T))

Notice that dim(E(λ,T))0 since if that's not the case then we have E(λ,T)={0} which is a contradiction as 0 itself isn't an eigenvector. As such, then we must have it that:

1=dim(E(λ,T))

for some eigenvalue λF. As such, we have one other eigenvalue λ8 for T, with some eigenvector vF5.

Now assume for contradiction that T2I and T6I are both not invertible. That means that both 2,6 are eigenvalues of T. But that's a clear contradiction since if that's the case then we have it that instead:

1=dim(E(2,T))+dim(E(6,T))++dim(E(λ,T))

but clearly since none of these dimensions can be 0, then we cannot share the 1 across all values of the dimensions of our eigenspaces above. Hence, then we must have a contradiction, so then the opposite of our assumption is true. Hence, T2I or T6I must be invertible.

10

Theorem

Suppose dim(V)=n and TL(V). Let λ1,...,λm denote the distinct nonzero eigenvalues of T. Then:

dim(E(λ1,T))++dim(E(λm,T))dim(range(T))

Proof

First, we'll prove a nice lemma, that E(λi,T)range(T), as this will be useful later on. Let vE(λi,T) be arbitrary. Since λi0 (see the theorem) then notice that we can say that Tv=λivT(λi1v)=v so then by definition then vrange(T). So no matter what vrange(T). Thus, we've proved our lemma.

Now consider:

E(λ1,T)E(λm,T)range(T)

which is true via a quick inductive proof, using our lemma as our base case. Notice then that using the idea of the dimension of direct sums gives the left side, and it being a subset implies the part:

dim(E(λ1,T))++dim(E(λm,T))dim(range(T))

11

Theorem

TL(R2) is diagonalizable since the matrix M(T) with respect to the vectors (1,4),(7,5) is:

(690046)

where:

T(x,y)=(41x+7y,20x+74y)

Proof
Notice that:

T(1,4)=(411+74,201+744)=(69,276)=69(1,4)

and:

T(7,5)=(417+75,207+745)=(322,230)=46(7,5)

thus then if β is the basis respective to these new vectors then:

M(T,β)=(690046)

as expected.

12

Theorem

Suppose R,TL(F3) each have λ=2,6,7 as eigenvalues. Then there is an invertible operator SL(F3) such that R=S1TS.

Proof
Notice that since dim(F3)=3 then, if e1,e2,e3 are the basis w.r.t. T and e1,e2,e3 is the basis w.r.t. R then we know that:

Te1=2e1,Te2=6e2,Te3=7e3

and similarly:

Re1=2e1,Re2=6e2,Re3=7e3

Choose S such that Sei=ei for i=1,2,3. Namely:

Se1=e1,...,Se3=e3

Notice then that S1 exists since:

S1ei=eiS1S=SS1=I

and furthermore, SL(F3) by the similar properties for our bases.

Since ei's form a basis for our vector space, then for any vF3 then:

v=α1e1+α2e2+α3e3

Thus:

S1TSv=S1TS(α1e1+α2e2+α3e3)=S1T(α1e1+α2e2+α3e3)=S1(α12e1+α26e2+α37e3)=α12e1+α26e2+α37e3=Rv

Thus S1TS=R.

14

Theorem

Find TL(C3) such that 6,7 are eigenvalues of T while T does not have a diagonal matrix with respect to any basis of C3.

Proof
The basis we are respective to doesn't matter, other than that there is some basis e1,e2,e3 as we have dim(C3)=3.

We'll want to find eigenvectors for our T such that the sum of the dimensions of our eigenspaces per λ doesn't add up to 3. We can just just force one eigenvector to occur per λ. Hence, we have:

{Te1=6e1Te2=7e2Te3=e1+6e3

Notice here we have e1 is an eigenvector for λ=6 and similarly we have λ=7 with eigenvector e2. Furthermore, notice that if β={e1,e2,e3} then:

M(T,β)=[601070006]

Notice that this is what we want! That's because we are always forced to have an UT matrix (and hence the diagonal tells what λ's there are), so we have to use 6 or 7 in there on the diagonal, and we must have other vector components since if not then e3 is an eigenvector of eigenvalue λ=6 or λ=7.

We can show that if vV then v=α1e1+α2e2+α3e3 and then:

Tv=λvTiαivi=λiαiviiαiTei=iλαivi6α1e1+7α2e2+α3(e1+6e3)=λα1e1+λα2e2+λα3e3(6α1+α3λα1)e1+(7α2λα2)e2+(6α3λα3)e3=0

Creates equations:

{6α1+α3λα1=07α2λα2=06α3λα3=0

Notice if λ=6 then (3) is satisfied (α3 is arbitrary), and from (2) we have α2=0, and from (1) we have α1=0. Thus, any multiple of e1 (α1 is now arbitrary) is an eigenvector with that eigenvalue.

A similar process shows that if λ=7 then any multiple of e2 is an eigenvector.

If λ6,7 then (2,3) force use to have it that α2,α3=0. Hence (1) becomes:

6α1λα1=0

where since λ6 then we have it that α1=0 so we'd have the zero vector. Hence, there's no other eigenvalues or vectors.

As such, then dim(E(6,T))=1=dim(E(7,T)). Their sum is less than the dimension of F3, hence T must not be diagonalizable.

15

Theorem

Suppose TL(C3) such that 6,7 are eigenvalues of T while T does not have a diagonal matrix with respect to any basis of C3. Then there exists (x,y,z)F3 such that:

T(x,y,z)=(17+8x,5+8y,2π+8z)

Proof
We claim that (17,5,2π)C3 is this vector. Notice that if λ=8 was an eigenvalue for T, then T would have 3 distinct λ's, and since dim(C3)=3 then that would imply that T is diagonalizable which is a contradiction. Thus, λ=8 must not be an eigenvalue.

As a result, then T8I must be surjective (if it wasn't, then since TL(C3) then it would imply that 8 is an eigenvalue which is a contradiction), so then there is some (x,y,z)C3 is non-zero such that (T8I)(x,y,z)=(17,5,2π). Then:

(T8I)(x,y,z)=T(x,y,z)8(x,y,z)(17,5,2π)=T(x,y,z)8(x,y,z)(17+8x,5+8y,2π+8z)=T(x,y,z)

completing the proof.