Lecture 13 - Linear Transformations

Studying vectors on their own is boring. Now, we'll look at mappings from vectors to other vectors.

For example, consider V=P(R), the space of polynomials with real coefficients. This is a countably-infinitely-dimensional vector space.

Now let D denote the calculus derivative map. So then D:VV such that it's defined by:

D(pV)=p

and if you remember power rule, we know how to take the derivative of a polynomial. In fact, we know that taking this derivative would be closed under derivation, as we get a polynomial as an output.

D actually has some really nice properties:

  1. D(p+q)=Dp+Dq. This comes from calculus where (p+q)=p+q.
  2. D(αp)=αDp. This comes from calculus where (αp)=αp.
    Because these 2 properties are satisfied, then we say D is a linear map or a linear transformation. Sometimes you'll hear that D is "linear", and that's that this means.

These two properties come up so often that we'll generalize it:

Linear Transformation

Given a vector space V and a vector space W, where V,W are over the same field F, then the function T:VW is a linear transformation iff:

  1. T(v1+v2)=Tv1+Tv2 for all v1,v2V
  2. T(αv)=αTv for all αF,vV

Note that there are two different + signs in (1). The v1+v2 adds vectors in V, while Tv1+Tv2 are adding vectors in W.

Some notation: sometimes you want to have the collection of linear maps from V to W. We denote it as L(V,W) to denote this collection of linear maps from V to W. Furthermore, if V is W, then instead of saying L(V,V) we can just say L(V).

So then, as an example, DL(P(F)).

But, okay this is mind blowing, L(V,W) itself is a vector space. That's because:

Okay let's see this in more detail. If we had some function f where f(v)=w then the additive inverse f would be the function where (f)(v)=w. Or, given a function TL(V,W), define S by S(v)=T(v). Since S+T=Z then T is the additive inverse of S and vice versa.

Properties of Linear Maps

Here are some properties of linear maps:

Properties of Linear Maps

For the following, we say TL(V,W), or TL(V,W).

  1. T(0V)=0W. This is because T(vv)=TvTv=0W, or T(0v)=0T(v)=0W. Notice that the zero vectors may be different, as they may come from different vector spaces.
  2. T(civi)=ciT(vi).
  3. T(v1v2)=T(v1)T(v2).

Let's look at an example. Consider R2, the transformation C rotates an input vector CCW by α radians. We can show that this is linear transformation, via a picture:

You could check property 2 (scalar multiplication) too and get that that holds too.

Let's look at another example. Let's work in C(R), the space of continuous functions with domain R. We'll define a function T:C(R)R, defined by:

T(f)=abf(x)dx

where a,b are fixed, real numbers. We claim that T is linear:

  1. T(f+g)=ab(f+g)(x)dx=abf(x)dx+abg(x)dx=T(f)+T(g).
  2. T(αf)=ab(αf)(x)dx=αabf(x)dx=αT(f).

So if we look at L(C(R),R), then TL(C(R),R). But notice that here a,b were arbitrary. If we have another one where we have b,c defined in T then it's still a linear transformation. But if we consider T+T then that itself is still a linear transformation! You can see how meta this can get...

LT's Defined by Their Bases

Linear Transformations Defined on the Actions of their Basis

Given a vector space V with basis v1,...,vn, then given any vectors w1,...,wn, in a vector space W (all over the same field F), there is a unique transformation T:VW satisfying:

T(vi)=wi

for all i.

We'll do part of the proof today:

Proof
This is an existence/uniqueness proof. We'll have to define a linear transformation. Suppose T were a LT with the property that T(vi)=wi for all i. Let vV be an arbitrary vector. Then since v1,...,vn is a basis, there's a linear combination of the basis elements that construct v:

a1v1++anvn=v

where each ai exists and is unique. Then:

T(v)=T(a1v1++anvn)=a1T(v1)++anT(vn)

but since T(vi)=wi:

v=a1w1++anwn

so if there were two T1,T2, doing this process yeilds that they are the same.

But is there one? We can just define T:VW as:

T(a1v1++anvn)=a1w1++anwn

and all we would have to show is that T is linear and T(vi)=wi for all i. The second just comes from the definition, but the linear part requires a bit more work.