Proof of Cochran's Theorem
📂Mathematical StatisticsProof of Cochran's Theorem
Theorem
Let Sample X=(X1,⋯,Xn) be iid and follow a Normal distribution like X1,⋯,Xn∼iidN(0,σ2). For a symmetric matrix A1,⋯,Ak∈Rn×n with rank rj, suppose the random variable Q1,⋯,Qk is expressed as a random vector quadratic form Qi:=XTAiX, and the sum of squares of the sample is given as ∑i=1nXi2=∑j=1kQj. Then the following holds.
∀j,σ2Qj∼χ2(rj)∧∀j1=j2,Qj1⊥Qj2⟺j=1∑krj=n
In other words, Qj are mutually independent and their equivalence condition with the Chi-square distribution χ2(rj) implies that the sum of ranks rj equals the size of the sample n.
Explanation
This theorem provides the theoretical framework supporting the analysis of variance where F-test is used.
Proof
Assume that (⟹) and Qj are mutually independent and that Qj/σ2∼χ2(rj) holds.
Addition of random variables: If Xi∼χ2(ri) then
i=1∑nXi∼χ2(i=1∑nri)
Since Qj/σ2 follows a Chi-square distribution with degree of freedom rj, the sum of these variables also follows a Chi-square distribution as below.
j=1∑kσ2Qj∼χ2(j=1∑krj)
Derivation of the Chi-square distribution from the standard normal distribution: If X∼N(μ,σ2) then
V=(σX−μ)2∼χ2(1)
Since X1,⋯,Xn follows a Normal distribution, Xi2/σ2∼χ2(1) holds, and their sum follows a Chi-square distribution as shown below.
i=1∑nσ2Xi2∼χ2(n)
Given that ∑i=1nXi2=∑j=1kQj was stated in the major premise, n=∑j=1krj must hold.
Assume that (⟸) and ∑j=1krj=n hold.
j=1∑kQj===XT(A1+⋯+Ak)XXTXi=1∑nXi2
Since ∑i=1nXi2=∑j=1kQj holds as discussed above, it can be noticed that In=∑j=1kAj is true. By defining the matrix Bj=In−Aj, Bj is equal to the sum excluding Aj, leaving the remainder A1,⋯,Ak.
Subadditivity of matrix rank: The rank of a matrix exhibits subadditivity. That is, for two matrices A,B, the following holds.
rank(A+B)≤rankA+rankB
By defining the rank of Rj0 as Bj0, the rank of the matrix sum is less than or equal to the sum of the individual ranks, yielding the following inequality.
Rj0=rankBj0≤rank(In−Aj0)=j=1∑krj−rj0=n−rj0
However, considering that In=Aj0+Bj0, it follows that n≤rj0+Rj0⟹n−rj0≤Rj0, and exactly Rj0=n−rj0 holds.
This implies that Bj0 has exactly 0 non-zero eigenvalues. The eigenvalues λ of Bj0 must satisfy det(Bj0−λI)=0, allowing us to rewrite it in the following manner given Bj0=In−Aj0.
det(In−Aj0−λIn)=det(Aj0−(1−λ)In)=0
Thus, the eigenvalues of Aj0 differ by 1 from the eigenvalues of Bj0, which exactly counted rj0 zero eigenvalues; hence Aj0 has exactly rj0 eigenvalues of 1, and the rest are all 0.
Symmetric real matrix with only eigenvalues 0 and 1: If a symmetric matrix A∈Rn×n has all eigenvalues as 0 or 1, then A is an idempotent matrix.
Conditions for the chi-square distribution of quadratic forms of normal random vectors: Let sample X=(X1,⋯,Xn) be iid following a normal distribution as X1,⋯,Xn∼iidN(0,σ2). For a symmetric matrix A∈Rn×n with rank r≤n, if we define the random vector quadratic form as Q=σ−2XTAX, the following holds.
Q∼χ2(r)⟺A2=A
All symmetric real matrices A1,⋯,Ak are idempotent matrices since their eigenvalues are only 0 and 1, and since their rank is rj, Qj/σ2 follows a Chi-square distribution χ2(rj).
Hogg-Craig theorem: Let sample X=(X1,⋯,Xn) be iid following a normal distribution as X1,⋯,Xn∼iidN(0,σ2). For a symmetric matrix A1,⋯,Ak∈Rn×n, assume that the random variable Q1,⋯,Qk is expressed as a random vector quadratic form Qi:=XTAiX, define the symmetric matrix A and random variable Q as follows.
A=Q=A1+⋯+AkQ1+⋯+Qk
If Q/σ2 follows a Chi-square distribution χ2(r), satisfies i=1,⋯,k−1 for Qi/σ2∼χ2(ri), and if Qk≥0 holds, then Q1,⋯,Qk is independent and Qk/σ2 follows a Chi-square distribution χ2(rk) with degrees of freedom rk=r−r1−⋯−rk−1.
By the Hogg-Craig theorem, Q1,⋯,Qk are mutually independent.
■