logo

Probability Variables and Probability Distributions Defined by Measure Theory 📂Probability Theory

Probability Variables and Probability Distributions Defined by Measure Theory

Definition 1

Let’s assume a Probability Space (Ω,F,P)( \Omega , \mathcal{F} , P) is given.

  1. A function X:ΩRX : \Omega \to \mathbb{R} that satisfies X1(B)FX^{-1} (B) \in \mathcal{F} for every Borel Set BB(R)B \in \mathcal{B} (\mathbb{R}) is called a Random Variable.
  2. FX\mathcal{F}_{X} defined as follows is called the Sigma Field generated by XX. FX:=X1(B)=σ(X)={X1(B)Ω:BB(R)} \mathcal{F}_{X} := X^{-1} ( \mathcal{B} ) = \sigma (X) = \left\{ X^{-1} (B) \in \Omega : B \in \mathcal{B}( \mathbb{R} ) \right\}
  3. Measure PXP_{X} defined as follows is called the Probability Distribution of XX. PX(B):=P(X1(B)) P_{X} (B) := P ( X^{-1} (B) )

  • If you haven’t encountered measure theory yet, you can ignore the term probability space.

Explanation

Just like the probability space, a random variable can also be rigorously defined within Measure Theory.

  1. Saying X1(B)FX^{-1} (B) \in \mathcal{F} means that XX maps elements of Ω\Omega to real numbers allowing the use of relations like P(aXb)P(a \le X \le b) while ensuring that the pre-images of Borel sets belong to the Sigma Field, thus limiting what is considered an Event to reasonable sets only. At first glance, it might seem overly abstract, but paradoxically, its goal is to counter excessive abstraction. According to the definition, a random variable XX is not only a real function but also a Measurable Function, and if Ω=R\Omega = \mathbb{R}, then F=B(R)\mathcal{F} = \mathcal{B} \left( \mathbb{R} \right), making it a Borel function X:RRX : \mathbb{R} \to \mathbb{R}. Basic theorems in mathematical statistics are sufficient at this level. Beyond this, generalization to multivariate random variables is simply done by defining X:ΩRpX : \Omega \to \mathbb{R}^{p} that satisfies X1(B)FX^{-1} (B) \in \mathcal{F} for every Borel set BB(Rp)B \in \mathcal{B} (\mathbb{R}^{p}). Naturally, XX can be expressed as a vector X=(X1,,Xp)X = ( X_{1}, \cdots , X_{p}) for each random variable Xi:ΩRX_{i} : \Omega \to \mathbb{R} and is called a Probability Vector. When this leads to a sequence of random variables, it is called a Stochastic Process, and more generally, a Random Element.
  2. For a Sigma Field G\mathcal{G}, if Y1(B)GY^{-1} ( \mathcal{B} ) \in \mathcal{G}, then YY is G\mathcal{G}-measurable, and naturally, according to the definition of FX\mathcal{F}_{X}, XX is FX\mathcal{F}_{X}-measurable.
  3. It might be confusing with all the definitions, but if you think about it step by step, it’s not difficult at all. Since X1(B)FX^{-1} (B) \in \mathcal{F}, you can think of it as if reversing the function, which leads to X1:B(R)FX^{-1} : \mathcal{B} (\mathbb{R}) \to \mathcal{F}. This way, PX:=(PX1)P_{X} : = ( P \circ X^{-1} ) can be understood as PX:B(R)F[0,1] P_{X} : \mathcal{B} (\mathbb{R}) \to \mathcal{F} \to [0,1] and is merely a composite function that maps any values between 00 and 11 for a given Borel set BB. For example, [3,2][-3,-2] is naturally a Borel set of R\mathbb{R}, and depending on how the random variable YY is defined, it enables calculations like PY([3,2])=0.7P_{Y} ( [-3,-2] ) = 0.7.

See Also


  1. Capinski. (1999). Measure, Integral and Probability: p66~68. ↩︎