logo

Definition of Likelihood Ratio Test in Mathematical Statistics 📂Mathematical Statistics

Definition of Likelihood Ratio Test in Mathematical Statistics

Definition 1

H0:θΘ0H1:θΘ0c \begin{align*} H_{0} :& \theta \in \Theta_{0} \\ H_{1} :& \theta \in \Theta_{0}^{c} \end{align*}

For the hypothesis test described above, the statistic λ\lambda is called the Likelihood Ratio test statistic. λ(x):=supΘ0L(θx)supΘL(θx) \lambda \left( \mathbf{x} \right) := {{ \sup_{\Theta_{0}} L \left( \theta \mid \mathbf{x} \right) } \over { \sup_{\Theta} L \left( \theta \mid \mathbf{x} \right) }}

A hypothesis test that has a rejection region {x:λ(x)c}\left\{ \mathbf{x} : \lambda \left( \mathbf{x} \right) \le c \right\} for a given c[0,1]c \in [0,1] is called a Likelihood Ratio Test and is often abbreviated as LRT.


  • LL is a likelihood function.

Explanation

In the definition of λ\lambda, the numerator involves finding the supremum in supΘ0\sup_{\Theta_{0}}, and the denominator involves finding the supremum in supΘ\sup_{\Theta}. The parameter space under the null hypothesis Θ0\Theta_{0} is a subset of the entire parameter space Θ0Θ\Theta_{0} \subseteq \Theta, and naturally, 0λ(x)10 \le \lambda \left( \mathbf{x} \right) \le 1 holds. The closer this ratio is to 00, the less plausible the parameters are under the null hypothesis.

Reflecting back on when we first encountered statistics, starting from basic probability distribution theory, studying test statistics like the t-distribution, F-distribution, chi-squared distribution separately seemed much cleaner. Of course, there is a certain motivation behind the Likelihood Ratio Test, but it makes sense without any buildup unlike the other tests mentioned.

Example: Normal Distribution

For practical applications of LRT, the supremum sup\sup must be reflected. Since the denominator becomes largest over the entire parameter space Θ\Theta, the maximum likelihood estimator is used, and the numerator is set to be maximized under the null hypothesis. H0:θ=θ0H1:θθ0 \begin{align*} H_{0} :& \theta = \theta_{0} \\ H_{1} :& \theta \ne \theta_{0} \end{align*} Consider a hypothesis test for a known-variance normal distribution N(θ,σ2)N \left( \theta , \sigma^{2} \right) with a random sample X1,,XnX_{1} , \cdots , X_{n}. In this case, the denominator should use the sample mean xˉ\bar{\mathbf{x}}, which is the maximum likelihood estimator for the population mean θ\theta, and the numerator should directly use θ0\theta_{0} since the parameter space of the null hypothesis is a singleton set Θ0={θ0}\Theta_{0} = \left\{ \theta_{0} \right\}. It can be mathematically derived as follows. λ(x)=supΘ0L(θx)supΘL(θx)=L(θ0x)L(xˉx)=(2π)n/2exp((xkθ0)2/2)(2π)n/2exp((xkxˉ)2/2)=exp(n(xˉθ0)2/2) \begin{align*} \lambda \left( \mathbf{x} \right) =& {{ \sup_{\Theta_{0}} L \left( \theta \mid \mathbf{x} \right) } \over { \sup_{\Theta} L \left( \theta \mid \mathbf{x} \right) }} \\ =& {{ L \left( \theta_{0} \mid \mathbf{x} \right) } \over { L \left( \bar{\mathbf{x}} \mid \mathbf{x} \right) }} \\ =& {{ (2\pi)^{-n/2} \exp \left( - \sum \left( x_{k} - \theta_{0} \right)^{2} / 2 \right) } \over { (2\pi)^{-n/2} \exp \left( - \sum \left( x_{k} - \bar{\mathbf{x}} \right)^{2} / 2 \right) }} \\ =& \exp \left( -n \left( \bar{\mathbf{x}} - \theta_{0} \right)^{2} / 2 \right) \end{align*} Note that λ(x)\lambda (\mathbf{x}) is exactly xˉ=θ0\bar{\mathbf{x}} = \theta_{0} when 11, and the larger the difference, the closer it gets to 00. Of course, it is understood that in the mathematical definition of LRT, it ranges from 00 to 11, but seeing the calculated result intuitively allows one to perform hypothesis testing based on how similar the sample variance is to the population mean under the null hypothesis.


  1. Casella. (2001). Statistical Inference(2nd Edition): p375. ↩︎