Proof of the Hogg-Craig Theorem
📂Mathematical StatisticsProof of the Hogg-Craig Theorem
Theorem
Let sample X=(X1,⋯,Xn) follow iid Normal distribution like X1,⋯,Xn∼iidN(0,σ2). Consider a symmetric matrix A1,⋯,Ak∈Rn×n and a random variable Q1,⋯,Qk represented as a random vector quadratic form Qi:=XTAiX. Define symmetric matrix A and random variable Q as follows:
A=Q=A1+⋯+AkQ1+⋯+Qk
If Q/σ2 follows a Chi-squared distribution χ2(r) and for i=1,⋯,k−1 it is Qi/σ2∼χ2(ri), and Qk≥0 is true, then Q1,⋯,Qk is independent, and Qk/σ2 follows a Chi-squared distribution with degrees of freedom rk=r−r1−⋯−rk−1 χ2(rk).
Explanation
It might initially seem strange that not Q/nσ2 but Q/σ2 follows a Chi-squared distribution, but it is precise to discuss Q/σ2 because the addition involved is not of samples, but of a matrix as follows:
Q====Q1+⋯+QkXTA1X+⋯+XTAkXXT(A1+⋯+Ak)XXTAX
This theorem is used in the proof of Cochran’s theorem.
Proof
We will prove it by mathematical induction. First, let k=2.
Equivalence condition for Chi-squaredness of normal distribution random vector quadratic forms: Let sample X=(X1,⋯,Xn), and follow iid Normal distribution like X1,⋯,Xn∼iidN(0,σ2). For a symmetric matrix with rank r≤n A∈Rn×n, if we set the random vector quadratic form as Q=σ−2XTAX, the following holds.
Q∼χ2(r)⟺A2=A
Since Q/σ2 follows a Chi-squared distribution, A is an idempotent matrix.
Eigenvalues of an idempotent matrix: The eigenvalues of an idempotent matrix are only 0 or 1.
Since A is symmetric and a real matrix, it is diagonalizable, and the eigenvalues of A are only 0 and 1, so there exists an orthogonal matrix which satisfies the following regarding the sizes of identity matrix Ir∈Rr×r and zero matrix O.
ΓTAΓ=[IrOOO]
If you derive A=A1+A2, it is as follows.
[IrOOO]=ΓTA1Γ+ΓTA2Γ
Positive definitiveness and eigenvalues: The necessary and sufficient condition for A to be positive definite is that all eigenvalues of A are positive.
Since it was assumed Q2≥0, the matrix A2 is positive semidefinite, and since A and A1 are idempotent, the eigenvalues are only 0 and 1, so they are also positive semidefinite according to the equivalence condition of semidefiniteness. Naturally, ΓTAΓ, ΓTA1Γ, and ΓTA2Γ, which have orthogonal matrices multiplied on the front and back, are also positive semidefinite.
Properties of diagonal elements of positive definite matrices: Given a positive definite matrix A=(aij)∈Cn×n. The sign of diagonal elements aii of A is the same as the sign of A. Suppose a symmetric matrix A∈Rn×n made up of real numbers is positive semidefinite.
When a positive semidefinite matrix is composed of real numbers and has symmetry, if any among the diagonal elements is 0, all rows and columns of that element would be 0. According to this, the following expression is possible for certain Gr∈Rr×r and Hr∈Rr×r.
ΓTAΓ=⟹[IrOOO]=ΓTA1Γ+ΓTA2Γ[GrOOO]+[HrOOO]
Since Q1/σ2∼χ2(r1), A1 is also an idempotent matrix, and the following is obtained.
(ΓTA1Γ)2=ΓTA1Γ=[GrOOO]
Multiplying both sides of ΓTAΓ=ΓTA1Γ+ΓTA2Γ by ΓTA1Γ results in the following:
[IrOOO]=⟹[IrOOO]ΓTA1Γ=⟹[IrOOO][GrOOO]=⟹[GrOOO]=⟹[OOOO]=⟹GrHr=⟹ΓTA1ΓΓTA2Γ=⟹A1A2=[GrOOO]+[HrOOO]ΓTA1Γ⋅ΓTA1Γ+[HrOOO]ΓTA1ΓΓTA1Γ+[HrOOO][GrOOO][GrOOO]+[GrHrOOO][GrHrOOO]OOO
Craig’s theorem: Let sample X=(X1,⋯,Xn) follow iid Normal distribution like X1,⋯,Xn∼iidN(0,σ2). For a symmetric matrix A,B∈Rn×n, if random variables Q1 and Q2 are defined as random vector quadratic forms Q1:=σ−2XTAX and Q2:=σ−2XTBX, the following holds.
Q1⊥Q2⟺AB=On
Addition of random variables: Xi∼χ2(ri) then
i=1∑nXi∼χ2(i=1∑nri)
According to Craig’s theorem, Q1 and Q2 are independent, and Q2 follows a Chi-squared distribution with degrees of freedom (r−r1).
It suffices to show that it holds for k=3. Suppose A3 is a positive semidefinite matrix satisfying the following:
A=A1+(A2+A3)=A1+B1
By factoring B1:=A2+A3, B1 is still a positive semidefinite matrix, and if we apply the result when k=2 to A=A1+B1, we obtain A1B1=O as follows:
A=A2===⟹B12=(A1+B1)2A12+A1B1+B1A1+B12A1+O+B12A−A1=B1
On the other hand, applying the result when k=2 to B1=A2+A3 itself gives A2A3=O and A32=A3. Repeating this factoring process for B1 applied to A=A2+(A1+A3) gives A1A3=O, and continuing this completes the proof.
■