F-distribution
📂Probability DistributionF-distribution
Definition
The continuous probability distribution F(r1,r2), which has the following probability density function for degrees of freedom r1,r2>0, is called the F-distribution.
f(x)=B(r1/2,r2/2)1(r2r1)r1/2xr1/2−1(1+r2r1x)−(r1+r2)/2,x∈(0,∞)
- B(r1/2,r2/2) refers to the beta function.
Basic Properties
Moment Generating Function
- [1]: The F-distribution does not have a moment-generating function.
Mean and Variance
- [2]: If X∼F(r1,r2), then
E(X)=Var(X)=r2−2r2r1(r2−2)2(r2−4)2r22(r1+r2−2),r2>2,r2>4
Theorem
Let two random variables U,V be independent with U∼χ2(r1) and V∼χ2(r2).
- [a]: If d2>2k, then F:=V/r2U/r1 exists as the kth moment
EFk=(r1r2)kEUkEV−k
- [b]: V/r2U/r1∼F(r1,r2)
- [c]: A random variable X∼F(r1,r2) that follows the F-distribution with degrees of freedom r1,r2 is defined as Y, which follows the beta distribution Best(2r1,2r2).
Y:=1+(r1/r2)X(r1/r2)X∼Beta(2r1,2r2)
- [d]: A random variable X∼t(ν) that follows the t-distribution with degrees of freedom ν>0 is defined as Y, which follows the F-distribution F(1,ν).
Y:=X2∼F(1,ν)
Reciprocality
- [e]: If X∼F(r1,r2), then the distribution of its reciprocal is as follows.
X1∼F(r2,r1)
Explanation
Just as the t-distribution is called the Student t-distribution, the F-distribution is also referred to as the Snedecor F-distribution, named after the statistician George Snedecor.
The probability density function of the F-distribution may seem incredibly complex at first glance, but in reality, there is little need to manipulate the formula itself. Understanding the relationship with the chi-squared distribution is of utmost importance. Just as the chi-squared distribution can be used for goodness-of-fit tests, the F-distribution can be used to compare the variances of two populations. As can be directly seen in theorem [b], since the F-distribution is expressed as a ratio of data following the chi-squared distribution, if this statistic deviates too much from 1, it can be inferred that the variances of the two distributions are different.
Proof
[1]
The existence of a moment-generating function for a random variable means that the kth moment exists for all k∈N. However, as per theorem [a], the kth moment of the F-distribution exists when k<d2/2, thus a moment-generating function cannot exist.
■
[2]
Use the moment formula stated in [a].
■
[a]
Substituting with t=r2r1x results in dt=r2r1dx, so
EFk=====∫0∞xkB(r1/2,r2/2)1(r2r1)r1/2xr1/2−1(1+r2r1x)−(r1+r2)/2dxB(r1/2,r2/2)1(r2r1)r1/2∫0∞xk+r1/2−1(1+r2r1x)−(r1+r2)/2dxB(r1/2,r2/2)1(r2r1)r1/2∫0∞(r1r2t)k+r1/2−1(1+t)−(r1+r2)/2r1r2dtB(r1/2,r2/2)1(r2r1)r1/2(r1r2)k+r1/2∫0∞tk+r1/2(1+t)−r1/2−r2/2dtB(r1/2,r2/2)1(r1r2)k∫0∞tk+r1/2(1+t)−(r1/2+k)−(r2/2−k)dt
Representation of the beta function as a definite integral:
B(p,q)=∫0∞(1+t)p+qtp−1dt
Relationship between the beta and gamma functions:
B(p,q)=Γ(p+q)Γ(p)Γ(q)
EFk====B(r1/2,r2/2)1(r1r2)kB(2r1+k,2r2−k)(r1r2)kΓ(r1/2)Γ(r2/2)Γ(r1/2+r2/2)Γ(r1/2+k+r2/2−k)Γ(r1/2+k)Γ(r2/2−k)(r1r2)kΓ(r1/2)Γ(r2/2)11Γ(r1/2+k)Γ(r2/2−k)(r1r2)kΓ(r1/2)Γ(r1/2+k)2kΓ(r2/2)2−kΓ(r2/2−k)
Moment of the chi-squared distribution: Let’s say X∼χ2(r). If k>−r/2, then the kth moment exists
EXk=Γ(r/2)2kΓ(r/2+k)
EFk=(r1r2)kEUkEV−k
■
[b]
Derive directly from the joint density function.
■
[c]
Derive directly from the variable change.
■
[d]
Circumvent as a ratio of the chi-squared distributions.
■
[e]
Since the numerator and the denominator are reversed, it is trivial according to theorem [b]. From a practical statistician’s point of view, defining the F-distribution according to theorem [b] and deriving the probability density function accordingly is more natural.
■
See Also