Lecture 22 (online) - Invariant Subspaces

Invariant Subspaces

Recall our usual setup:

The idea of invariant subspaces asks whether the subspace mapping from T:UU maps into U only. In the picture above we have an invariant, but if we had:

Then we don't have an invariant.

Invariant

Suppose TL(V), and we have subspace U of V. The subspace U is an invariant under T if, for all uU we have TuU.

As some examples of invariant subspaces:

Some cooler ones are as follows. Here null(T) is an invariant since all items get mapped to the zero vector which is in null(T):

Likewise, the range(T) is also an invariant subspace of T:

Why are we interested?

Suppose V=U1U2Um, where Ui are all invariant under T. Suppose any v=u1+u2++umV. Then Tv=Tu1++Tum where each Tui are in Ui. Hence, we can reduce our space into our corresponding smaller subspaces to talk about the ultimate higher ones.

As an example, consider TL(V), using R as our field. Let vV with v0V, where Tv=2v. We claim that span(v) is an invariant subspace of V.

Proof
We need to show that for any uspan(v) and show that Tuspan(v). Let uspan(v) be arbitrary. Then:

u=αv

for some αR. Then:

Tu=2u=2(αv)=(2α)v

but notice that β=2αR:

Tu=βuspan(v)

thus span(v) is an invariant under T.

Notice that this transformation merely scaled our vector. We're going to hence define the related eigenvector, and the amount it gets scaled by is its eigenvalue.

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors

Given TL(V):

  1. A number λF is called an eigenvalue of T if there exists vV such that v0 and Tv=λv.
  2. A vector vV is called an eigenvector of T corresponding to eigenvalue λ if v0V and Tv=λv.

For instance, let V be the vector space having basis B={e2x,xe2x} and let D be the derivative map. Note that DL(V) already, and we can see what it does to our vectors:

D(e2x)=2e2x

notice here that λ=2 is an eigenvalue of D. Furthermore, v=e2x is its corresponding eigenvector.

Note that:

D(0v)=0v=λ0v

for all λF. Does this mean that 0v is an eigenvector? Or that any λ is an eigenvalue? We don't really want this for any other transformations or spaces, hence why we remove these cases in the definition.

Note

We say that 0 isn't an eigenvector, but 0 could be a possible eigenvalue.

Notice the following. If v is a (non-zero) eigenvector with e-val λ then:

T(v)=λvT(v)λv=0T(v)λIVv=0(TλIV)v=0

so then vnull(T) so then all eigenvectors are in the nullspace of T.

This gives rise to the following theorem:

Equivalences of Eigenvalues/Eigenvectors

The following are equivalent:

  1. λ is an eigenvalue of T
  2. TλIV is not injective.
  3. TλIV is not surjective.
  4. TλIV is not invertible.
    where IV:VV where for any vV we have Iv(v)=v.

Proof
This is a baby proof, which is more outlined in the book. Suppose (1). Then there is some non-zero vector v such that Tv=λv. Subtracting on both sides gives Tvλv=0v, so then (TλIV)v=0v. We've found something that maps a non-zero vector v to the zero-vector, so then we've gotten (2).

If we have (2) we get (3, 4) as we've shown before. If you work backwards, you can suppose non-injectivity and extract the λ as an eigenvalue thing.

Here we note that null(TλIV) is the eigenspace of T on V.

Theorem

Suppose we have distinct eigenvalues λ1,,λm are distinct eigenvalues of TL(V) and v1,...,vm are corresponding eigenvectors, where all T(vi)=λivi. Then v1,...,vm is LI in V.

Proof
We prove this by contradiction. Assume instead that v1,...,vm is LD. Note that v1 itself is just LI always, and as we add more vectors, we then get LD. Hence, let k be the smallest positive integer such that v1,...,vk is LD. Then vk can be written as a linear combination of v1,...,vk1. Then:

Tvk=α1Tv1++αk1Tvk1

where since all vi's are eigenvectors, then:

λkvk=α1λ1v1++αk1λk1vk1

multiply our initial definition of vk by λk to get:

λkvk=α1λkv1++αk1λkvk1

But then equating coefficients says that λ1=λk,,λk1=λk, as subtracting both equations above gives

0=α1(λkλ1)v1++αk1(λkλk1)vk1

But since v1,...,vk1 is LI, then all ai(λkλi)=0. So either αi=0 or λkλi=0. We cannot have the latter since all the λi's are unique, so then αi=0, so then vk=0 is a contradiction as vk is non-zero.

This is handy since we can construct bases from our corresponding eigenvalues and vectors.

Theorem

Suppose V is a finite-dimensional vector space. Then each linear transformation T on V has at most dim(V) distinct eigenvalues and corresponding eigenvectors.

Proof
Assume you had more than dim(V) eigenvalues. Then that would give you more than that n number of eigenvectors, which are all LI. But since the dimension is n, then we have a LI list of longer than n eigenvectors, which is a contradiction.

Intro to Polynomial Operators

Given some polynomial p=a0+a1z++anzn and some operator TL(V), we can construct a new operator from this polynomial, namely p(T) is the linear operator in L(V) such that:

p(T)(v)=(a0I+a1T++anTn)(v)

where

Tk=TTTk times

we'll talk more about this in the next lecture.