Sufficient Statistics and Maximum Likelihood Estimators for the Geometric Distribution
📂Probability DistributionSufficient Statistics and Maximum Likelihood Estimators for the Geometric Distribution
Theorem
Given a random sample X:=(X1,⋯,Xn)∼Geo(p) that follows a geometric distribution, the sufficient statistic T and the maximum likelihood estimator p^ for p are as follows.
T=p^=k=1∑nXk∑k=1nXkn
Proof
Sufficient Statistic
f(x;p)====k=1∏nf(xk;p)k=1∏np(1−p)xk−1pn(1−p)∑kxk−npn(1−p)∑kxk−n⋅1
Neyman factorization theorem: Suppose a random sample X1,⋯,Xn has the same probability mass/density function f(x;θ) with respect to the parameter θ∈Θ. A statistic Y=u1(X1,⋯,Xn) being a sufficient statistic for θ implies the existence of two non-negative functions k1,k2≥0.
f(x1;θ)⋯f(xn;θ)=k1[u1(x1,⋯,xn);θ]k2(x1,⋯,xn)
However, k2 must not depend on θ.
According to the Neyman factorization theorem, T:=∑kXk is a sufficient statistic for p.
Maximum Likelihood Estimator
logL(p;x)===logf(x;p)logpn(1−p)∑kxk−nnlogp+k=1∑nxklog(1−p)
The log-likelihood function of the random sample is as above, and for the likelihood function to be maximized, the partial derivative with respect to p needs to be 0. Therefore,
⟹⟹⟹0=np1−1−p1(k=1∑nxk−n)pn+1−pn=1−p1k=1∑nxkp(1−p)n=1−p1k=1∑nxkp1=n1k=1∑nxk
Thus, the maximum likelihood estimator p^ for p is as follows.
p^=∑k=1nXkn
■