Variance and Standard Deviation of X

Sometimes even though the Expected Value of some distribution equals, they are more/less variant:

Pasted image 20241003164949.png

Although both distributions in the figure above have the same expected value, they clearly are different distributions, where the second is more varied than the first. We have a quantitative way to describe this variance:

Variance, standard deviation

Let X have pmf p(x) and expected value μ. Then the variance of X, denoted by Var(X) or σX2 or σ2 is:

Var(X)=xD[(xμ)2p(x)]=E[(Xμ)2]

The standard deviation of X denoted SD(X) or σX or σ is just:

σX=Var(X)
Where does this come from

The idea is that σ is the average distance from μ. Think of the σ2 function as doing the norm, so then square rooting will find the distance from μ.

See [[handout05-DiscreteRVs-350F24_annotated.pdf#page=4]] for some examples of computing these.

Chebyshev's Inequality

Let X be a discrete rv with Expected Value μ and standard deviation σ. Then for any k1:

P(|Xμ|kσ)1/k2

That is, the probability X is at least k standard deviations away from the mean μ is at most 1/k2.

Proof
Let A denote the event |Xμ|kσ. Begin by rewriting out the definition of Var(X):

Var(X)=D[(xμ)2p(x)]=A[(xμ)2p(x)]+A[(xμ)2p(x)]A[(xμ)2p(x)]the other sum 0A[(kσ)2p(x)](xμ)2(kσ)2=(kσ)2Ap(x)=(kσ)2P(A)=k2σ2P(|Xμ|kσ)

Solving Var(X)=σ2 and simplifying gives the desired result. The middle step uses the variance shortcut.

Variance Shortcut

Var(X)=σ2=E(X2)μ2

Proof
See [[handout05-DiscreteRVs-350F24_annotated.pdf#page=5]].

If h(X) is a linear function, then we get the following:

Linear h implies linear transformation of Var,μ,σ

If h(X)=Y is a linear function, then:

Var(aX+b)=σaX+b2=a2σX2σaX+b=|a|σX