Proof of the Cauchy-Schwarz Inequality in Mathematical Statistics
📂LemmasProof of the Cauchy-Schwarz Inequality in Mathematical Statistics
Theorem
For a random variable X,Y, the following holds.
Cov(X,Y)≤VarXVarY
The necessary and sufficient condition for the equality to hold is as follows.
∃a=0,b∈R:aX+b=Y
Proof
Let’s denote the population means of X,Y as μX and μY, respectively.
h(t):===E([(X−μX)t+(Y−μY)]2)t2E[(X−μX)2]+2tE[(X−μX)(Y−μY)]+[(Y−μY)2]VarXt2+2Cov(X,Y)t+VarY
According to the root determination method of the quadratic formula, for the root of h to exist at most once, the following must be true.
(2Cov(X,y))2−4VarX⋅VarY≤0
Rearranging this leads to the following Cauchy-Schwarz inequality.
Cov(X,Y)≤VarXVarY
For the equality to hold, it must be that h(t)=0, and when a:=−t and b:=μXt+μY are set, it is equivalent to the following.
⟺⟺P([(X−μX)t+(Y−μY)]2=0)=1P((X−μX)t+(Y−μY)=0)=1P(Y=aX+b)=1
■
Explanation
At first glance, because the variance Var and covariance Cov appear, it may seem different from the commonly known Cauchy-Schwarz inequality, but upon closer inspection, there is no reason not to call it the Cauchy-Schwarz inequality. When considering the application in mathematical statistics, the inequality itself, as well as the necessary and sufficient condition for the equality to hold
∃a=0,b∈R:aX+b=Y
are very useful.