Lecture 13 - Applications of Quadratic Forms

This lecture is much more of a lecture on the applications of what we've learned so far.

Quadratic Forms

Quadratic Forms

A quadratic form in n commuting variables is a homogeneous polynomial of degree 2 in terms of those variables.

Here homogeneous here just says that every monomial in the polynomial is of degree 2. As an example:

3x2+y2z2+4xz5xy

is a quadratic form of 3 commuting variables x,y,z. This is homogeneous since each term is of degree 2 or less. If we added a +5y then it wouldn't fit the form since we need exactly the second degree of a combination of our x,y,z.

If we have a quadratic form Q(x1,...,xn) then:

Q(x1,...,xn)=i,j=1naijxixj

We don't really need to do the entire summation, as xixj=xjxi. As such, we may restrict to have ij. Notice:

Q(x1,...,xn)=i=1nxi(j=1naijxj)the i-th entry of Ax=i=1nxi(Ax)i=x,AxDot Product=xTAx(x,y=xTy for dot product)

You may guess that since xixj is commuting that we have some freedom. This will be the case for A.

As an instance of this, consider:

3x12+2x1x2+x1x33x32=[x1x2x3][3a1b1a200b203][x1x2x3]

We want to convert the LHS to the matrix representation we noticed before. We got the entries in the matrix by doing the actual matrix transformations and confirming these things. This works as long as:

a1+a2=2,b1+b2=1

We want to make A symmetric, so we can do that by splitting our terms! Thus have a1=a2=1 and b1=b2=12. Thus:

A=[31121001200]

is a symmetric matrix. Notice in general for all the 2 terms we have the entries on the diagonal, and the antidiagonal is the sums of the other entries for Q.

Symmetry in A

Every quadratic form Q can be put in the form:

Q(x)=xTAx

where A is a symmetric matrix.

The operator T where M(T)=A then must be normal and self-adjoint (if we restrict AM(Rn)). Namely, T:RnRn is defined by Tx=Ax. Thus T is self-adjoint since M(T)=A=A. Using the Spectral Theorem, then we must have an orthonormal eigenbasis β={u1,...,un} for which M(T,β) is diagonal.

If we let:

P=[111u1u2un111]PT=[1u111u211un1]

Thus:

PTP=IPT=P1

In particular, we have PTAP=D, or A=PDPT, where D=diag(λ1,...,λn) for each u1,...,un eigenvector.

So given a quadratic form:

Q(x)=xTAx(Let y=PTx so x=Py)Q(x)=Q(Py)=(Py)TA(Py)=yTPTAPy=yTDyD=diag(λ1,...,λn)=λ1y12++λnyn2

This is a nice theorem that we just proved:

Theorem

Given a symmetric matrix A, there is an orthogonal change of variables x=Py or (y=PTx) that transforms the quadratic form xTAx into yTDy where there are no cross terms. Here D=diag(λ1,...,λn).

Using these!

When you graph quadratic forms, ellipses, hyperbolas, etc. are all different shapes that had constrained properties. The theorem says we can make a change of variables to allow for an elimination of other potential cases.

For instance, try to graph:

2x24xy+5y2=1

We want to try to graph this out. Here we have a quadratic form:

[xy][2225][xy]

So we have an orthonormal basis for A. We can go through our Linear Analysis I methods to find our eigenvectors. Doing so yields that v1=(1,2) and v2=(2,1) with λ1=6 and λ2=1. Notice their dot product is 0, as expected as they are orthogonal (although they are not normal) but we can normalize. As such:

u1=15[12],u2=15[21]

so then define:

P=15[1221]

notice that $P^T = P here. So then we introduce our new coordinates:

x=[xy]=PTx=PT[xy]=15[x+2y2x+y]

Thus:

Q(x)=xT[6001]x=6(x)2+(y)2

So we can graph the ellipse in the x,y coordinate system, noting that it's equivalent to its standard form:

(x)2(16)2+(y)212=1

Then we just convert from x,y back to x,y. Recall that we had:

x=PTxx=Px,P=[1/52/52/51/5]

Thus we just transform our x to x in a transforming way:

Here $P$ reflects across $y = x$ and rotates CCW by some angle $\alpha$. Thus, we transform into the new axes:

Taylor Series with Multiple Variables

In Calc IV you learn about the second-derivative test. We now have all the tools using Linear Algebra to show why that works entirely.

We first look at the extremal values of a quadratic form. For instance if we have:

Q(x)=xTAx

then:

Q(αx)=(αx)TA(αx)=α2(xTAx)=α2Q(x)

If we know what Q does to unit vectors, then we know what it does to all vectors. So we can just look at the unit sphere for these vectors. Suppose x=1, so it lives there. The question we'll look at (tomorrow) is: what are the extreme values of Q(x) when x=1?