logo

The Variance of an Unbiased Estimator Given a Sufficient Statistic is Minimized 📂Mathematical Statistics

The Variance of an Unbiased Estimator Given a Sufficient Statistic is Minimized

Theorem 1

Let’s say we have a parameter θ\theta. UU is an unbiased estimator, T1T_{1} is a sufficient statistic, and T2T_{2} is a minimal sufficient statistic, defined as follows: U1:=E(UT1)U2:=E(UT2) \begin{align*} U_{1} :=& E \left( U | T_{1} \right) \\ U_{2} :=& E \left( U | T_{2} \right) \end{align*} it holds that: VarU2VarU1 \operatorname{Var} U_{2} \le \operatorname{Var} U_{1}

Explanation

Whether T1T_{1} or T2T_{2} is given, UU being an unbiased estimator means it hits θ\theta in expectation, but roughly speaking, it does so with less fluctuation when the minimal sufficient statistic is given. It’s easy to remember that the minimality of the sufficient statistic leads to the minimality of the variance of the unbiased estimator.

Proof

Definition of the Minimal Sufficient Statistic: A sufficient statistic T(X)T \left( \mathbf{X} \right) is called a minimal sufficient statistic if it can be represented as a function of every other sufficient statistic T(X)T ' \left( \mathbf{X} \right), denoted by T(x)T \left( \mathbf{x} \right) being a function of T(x)T ' \left( \mathbf{x} \right).

According to the definition of the minimal sufficient statistic, since T2T_{2} can be represented as a function of T1T_{1}, E(U1T2)=E(E(UT1)T2)=E(UT2)=U2 \begin{align*} E \left( U_{1} | T_{2} \right) =& E \left( E \left( U | T_{1} \right) | T_{2} \right) \\ =& E \left( U | T_{2} \right) \\ =& U_{2} \end{align*}

Property of Conditional Variance: Var(X)=E(Var(XY))+Var(E(XY)) \operatorname{Var}(X) = E \left( \operatorname{Var}(X | Y) \right) + \operatorname{Var}(E(X | Y))

Following the property of conditional variance, for U1U_{1} and T2T_{2} we have

VarU1=EVar(U1T2)+VarE(U1T2)=EVar(U1T2)+VarU2 \begin{align*} \operatorname{Var} U_{1} =& E \operatorname{Var} \left( U_{1} | T_{2} \right) + \operatorname{Var} E \left( U_{1} | T_{2} \right) \\ =& E \operatorname{Var} \left( U_{1} | T_{2} \right) + \operatorname{Var} U_{2} \end{align*}

This holds for any other sufficient statistic T1T_{1}, so the variance of the expected value of the unbiased estimator UU given the minimal sufficient statistic T2T_{2} is minimized.


  1. Casella. (2001). Statistical Inference(2nd Edition): p305. ↩︎