Lecture 23 - Polynomial Operator

Let's just start with an example. Say:

p(x)=x22x+2

and D is the derivative map. Then:

P(D)=D22D+2I

notice we are just "plugging in" x=D here. So then:

P(D)(D22D+2I)(1+x+x2)=D2(1+x+x2)2D(1+x+x2)+2(1+x+x2)

which simplifies to:

(2)2(2x+1)+2(1+x+x2)=...

the rest really doesn't matter. What's important is that P(D) is a polynomial operator, derived from D. Notice that it's an operator, so we operated it on something that D can operate on.

As another example, let p(x)=x2 and q(x)=2x+1, and P(T)=T2I and Q(T)=2T+I, where TL(V). So then:

P(T)Q(T)v=P(T)(2Tv+v)=(T2I)(2Tv+v)=2T2v4Tv+Tv2v=2T2v3Tv2v

which also equals:

(2T23T2I)v

it shouldn't surprise you that the multiplication of polynomials give the polynomial operator where x=T:

p(x)q(x)=2x23x2

Therefore, it follows that:

(pq)(T)=p(T)q(T)

as an interesting property. Is it always true? But of course, and it's because that you can foil out polynomial operators just like normal polynomials, so the properties of polynomials carry over to these new objects.

As an example, since (pq)(x)=(qp)(x) then by our theorem above then:

p(T)q(T)=(pq)(T)=(qp)(T)=q(T)p(T)

THIS IS BIGGGGGG!!!! Not always fggf, but it does here. The thing here is that because we are only dealing with one operator T and that:

TT=TT

And that we have distributivity and all the other good properties of TL(V) which is a vector space, and therefore has those properties.

Theorem

Every operator on a finite-dimensional complex vector space V has an eigenvalue, and associated eigenvector.

Note that this isn't true on real vector spaces. This is why. Think of the rotation transformation. Here in a real vector space you only have complex-valued eigenvalues, so then it won't work for real vector spaces.

Proof
Suppose that V is a finite-dimensional complex vector space V with dim(V)=n1. Further, let's suppose that TL(V). We'll need to show that T has an eigenvalue.

Choose v0V. Make a list:

v,Tv,T2v,,Tnv

eventually, we'll have a LD list, since there's n+1 vectors, so this list is definite LD. Hence, there's αiC (not all 0) where:

α0Iv+α1Tv++αnTnv=0

But we can convert our operators into a polynomial operator!:

(α0I+α1T++αnTn)v(α0+α1z++αnzn)(T)v=0

And we can factor our polynomial into a product of linear terms. But αn may be zero, and so on. But we have at least 1 that's non-zero, so the degree of our polynomial is at least degree 1 (really between 1 and n), so call it degree m. Thus:

c(zλ1)(zλm)(T)v=0

Thus, we can write this as a product of operators:

c(Tλ1I)(TλmI)v=0

since v0. Recall that if (TλI)v=0 then vnullspace(TλI) so then v is an eigenvector and λ is it's eigen value. Go in the order of the polynomials. If (TλmI)v=0 then by the prior theorem then v is an eigenvector with eigenvalue λm. If (Tλm1I)(TλmI)v=0 then we know that (TλmI)v is an eigenvector with eigenvalue λm1. Repeat forever. We know that we must have one of these has to be it since if we get to the end:

c(Tλ1I)(TλmI)v=0

where since c0 (we're dealing with a degree m polynomial) so then if we get to that point then (TλmI)v is an eigenvector with eigenvalue λ1.

Boom!

We can now always use an eigenvalue to simplify certain processes! An application is now detailed below:

Computing Higher Powers of Matrices

When we have 0's in matrices, we normally can diagonalize it, and thus compute higher powers of matrices. Let An×n. We know that the diagonal is on the ai,i entries for all 1in.

[a11a12a1na21a22a2nan1an2ann]

Upper Triangular

A matrix was upper triangular if there's 0's in the lower triangle (ie: there's data only in the diagonal and in the upper part):

[a11a12a1n0a22a2n00ann]

for example, suppose that V is a vector space of dimension 3, and v1,v2,v3 is its corresponding basis β. Let TL(V) where:

M(T,β)=[a11a12a130a22a2300a33]

Therefore:

T(v1)=a11v1T(v2)=a12v1+a22v2T(v3)=a13v1+a23v2+a33v3

but notice this! If I look at T(vi) we can always write this as a linear combination of all vj's where j<i. Namely:

T(vi)span(v1,...,vi)

That's big! If I were to look at:

T(span(vi))span(v1,...,vi)

So then span(v1,...,vi) is a T invariant subspace for all i.

We tie everything together into one neat theorem:

Upper Triangular Transformation Properties

The following are equivalent:

  1. T has an upper triangular matrix representation w.r.t. some basis β={v1,...,vn} for V.
  2. T(vi)span(v1,...,vi) for all 1in
  3. span(v1,...,vi) is T-invariant for all i.

Finishing Up

For the next lecture, we'll talk about the following theorem:

Theorem

Every operator TL(V), where V is a finite-dimensional complex vector space has a matrix representation which is upper-triangular.

We'll talk about this next time.