logo

Polynomial Distribution 📂Probability Distribution

Polynomial Distribution

Definition

Let a random vector composed of nNn \in \mathbb{N} and kNk \in \mathbb{N} counts of random variables be denoted as (X1,,Xk)\left( X_{1} , \cdots , X_{k} \right). i=1kXi=n&i=1kpi=1 \sum_{i=1}^{k} X_{i} = n \qquad \& \qquad \sum_{i=1}^{k} p_{i} = 1 For p=(p1,,pk)[0,1]k\mathbf{p} = \left( p_{1} , \cdots , p_{k} \right) \in [0,1]^{k} that satisfies this, a multivariate probability distribution Mk(n,p)M_{k} \left( n, \mathbf{p} \right) with the following probability mass function is called the Multinomial Distribution. p(x1,,xk)=n!x1!xk!p1x1pkxk,x1,,xkN0 p \left( x_{1} , \cdots , x_{k} \right) = {{ n! } \over { x_{1} ! \cdots x_{k}! }} p_{1}^{x_{1}} \cdots p_{k}^{x_{k}} \qquad , x_{1} , \cdots , x_{k} \in \mathbb{N}_{0}


  • [0,1]k=[0,1]××[0,1][0,1]^{k} = [0,1] \times \cdots \times [0,1] is a kk-cell.
  • N0={0}N\mathbb{N}_{0} = \left\{ 0 \right\} \cup \mathbb{N} is a set that includes natural numbers and 00.

Description

To interpret the definition as it is, (X1,,Xk)\left( X_{1} , \cdots , X_{k} \right) is a random vector indicating how many elements are actually in each category when nn elements have a probability pip_{i} of falling into the ii category among kk categories, having a probability mass function of p(x1,,xk)=P(X1=x1,,Xk=xk)=n!x1!xk!p1x1pkxk \begin{align*} p \left( x_{1} , \cdots , x_{k} \right) =& P \left( X_{1} = x_{1} , \cdots , X_{k} = x_{k} \right) \\ =& {{ n! } \over { x_{1} ! \cdots x_{k}! }} p_{1}^{x_{1}} \cdots p_{k}^{x_{k}} \end{align*} Especially, when k=2k = 2, it becomes a generalization of the binomial distribution itself.

Basic Properties

Mean and Covariance

  • [1]: If X:=(X1,,Xk)Mk(n,p)\mathbf{X} := \left( X_{1} , \cdots , X_{k} \right) \sim M_{k} \left( n, \mathbf{p} \right), the expected value of the ii component XiX_{i} is E(Xi)=npi E \left( X_{i} \right) = n p_{i} and the covariance matrix is as follows. Cov(X)=n[p1(1p1)p1p2p1pkp2p1p2(1p2)p2p2pkp1pkp2pk(1pk)] \operatorname{Cov} \left( \mathbf{X} \right) = n \begin{bmatrix} p_{1} \left( 1 - p_{1} \right) & - p_{1} p_{2} & \cdots & - p_{1} p_{k} \\ - p_{2} p_{1} & p_{2} \left( 1 - p_{2} \right) & \cdots & - p_{2} p_{2} \\ \vdots & \vdots & \ddots & \vdots \\ - p_{k} p_{1} & - p_{k} p_{2} & \cdots & p_{k} \left( 1 - p_{k} \right) \end{bmatrix}

Theorem

Lumping Property

For iji \ne j, Xi+XjX_{i} + X_{j} follows the binomial distribution Bin(n,pi+pj)\text{Bin} \left( n , p_{i} + p_{j} \right). Xi+XjBin(n,pi+pj) X_{i} + X_{j} \sim \text{Bin} \left( n , p_{i} + p_{j} \right) This is called the Lumping Property.

Proof

Mean

Looking at each component XiX_{i} alone, it’s essentially a binomial distribution regarding whether it falls into category ii with a probability pip_{i} or not, hence XiBin(n,pi)X_{i} \sim \text{Bin} \left( n , p_{i} \right), and its expected value is E(Xi)=npiE \left( X_{i} \right) = n p_{i}.

Covariance

It is directly deduced using the lumping property.

Lumping Property 1

In the case of n=1n = 1, that is, when considering only a single trial, Xi+XjX_{i} + X_{j} is exactly 11 when the outcome of that trial belongs to either the ii or jj category, and follows a Bernoulli distribution Bin(1,pi+pj)\text{Bin} \left( 1, p_{i} + p_{j} \right) which is 00 in all other cases.

Addition of Binomial Distributions: Let’s assume that the probability variables X1,,XnX_{1} , \cdots , X_{n} are mutually independent. In the case of binomial distributions, if XiBin(ni,p)X_i \sim \text{Bin} ( n_{i}, p) is true, i=1mXiBin(i=1mni,p) \displaystyle \sum_{i=1}^{m} X_{i} \sim \text{Bin} \left( \sum_{i=1}^{m} n_{i} , p \right)

Since nn trials are conducted independently, the following is obtained according to the addition of binomial distributions. Xi+XjBin(j=1n1,pi+pj)=Bin(n,pi+pj) X_{i} + X_{j} \sim \text{Bin} \left( \sum_{j=1}^{n} 1 , p_{i} + p_{j} \right) = \text{Bin} \left( n , p_{i} + p_{j} \right)