logo

Conditional Expectation of Random Variables Defined by Measure Theory 📂Probability Theory

Conditional Expectation of Random Variables Defined by Measure Theory

Definition

Let’s assume a probability space (Ω,F,P)( \Omega , \mathcal{F} , P) is given.

If G\mathcal{G} is a sub sigma-field of F\mathcal{F} and the random variable XL1(Ω)X \in \mathcal{L}^{1} ( \Omega ) is integrable, for all AGA \in \mathcal{G}, AYdP=AXdP \int_{A} Y d P = \int_{A} X d P a G\mathcal{G}-measurable random variable YY uniquely exists satisfying the above, then Y:=E(XG)Y := E ( X | \mathcal{G} ) is defined as the conditional expectation of XX given G\mathcal{G}.


  • Even though I’d like to say you can ignore the term probability space if you have not encountered measure theory yet, understanding the content of this post is nearly impossible without any knowledge of measure theory at all.
  • That G\mathcal{G} is a sub sigma-field of F\mathcal{F} means both are sigma-fields of Ω\Omega, with GF\mathcal{G} \subset \mathcal{F} being true. That YY is a G\mathcal{G}-measurable function means for every Borel set BB(R)B \in \mathcal{B}(\mathbb{R}), Y1(B)GY^{-1} (B) \in \mathcal{G} holds true.

Explanation

In the mathematical definition, G\mathcal{G} becomes a probability space (Ω,G,P)( \Omega , \mathcal{G} , P), which is not as vast as the original probability space (Ω,F,P)( \Omega , \mathcal{F} , P) but has a bit more information given. Therefore, AXdP=AYdP=AE(XG)dP \int_{A} X d P = \int_{A} Y d P = \int_{A} E ( X | \mathcal{G} ) d P indicates that the calculation remains the same within that reduced space, which means we’ve successfully pulled down the probability PP from F\mathcal{F} to G\mathcal{G} while keeping its properties intact.

Also, from the format of the definition, since E(XG)E ( X | \mathcal{G} ) exists as a G\mathcal{G}-measurable random variable, it should be naturally accepted that the expected value is a variable random, and measurable with respect to the given sigma field.

While the definition may not be intuitively difficult to grasp, its expression might feel somewhat unfamiliar. The σ(X):={X1(B):BB(R)}\sigma (X) := \left\{ X^{-1} (B) : B \in \mathcal{B}(\mathbb{R}) \right\} regarding the random variable XX becomes the smallest sigma field σ(X)F\sigma (X) \subset \mathcal{F} generated by XX, which is often stated in familiar terms as: E(YX)=E(Yσ(X)) E(Y|X) = E \left( Y | \sigma (X) \right) However, although this can be stated, getting accustomed to this approach is much easier if one intends to continue studying measure-based probability theory. Considering it, E(YX)E(Y|X) was conceptually intuitive but was a painfully complex notation for handling formulas or making direct calculations. It’s time to let it go without regret.

The existence of the conditional expectation is warranted by the Radon-Nikodym Theorem. Understanding the theorem is crucial, but the proof itself is not difficult.

Proof

Case 1. X0X \ge 0

PG(A):=AXdP P_{\mathcal{G}}(A) := \int_{A} X d P If defined as PGP_{\mathcal{G}} for all AGA \in \mathcal{G}, PGP_{\mathcal{G}} becomes a measure on G\mathcal{G}, and PGPP_{G} \ll P holds.

If the two sigma-finite measures ν\nu, μ\mu on the measurable space (Ω,F)( \Omega , \mathcal{F} ) satisfy νμ\nu \ll \mu according to the Radon-Nikodym Theorem, for all AFA \in \mathcal{F}, μ\mu-almost everywhere h0h \ge 0 and ν(A)=Ahdμ \nu (A) = \int_{A} h d \mu a F\mathcal{F}-measurable function ff uniquely exists in accordance to μ\mu.

Following the theorem, if we define ν=PG\nu = P_{\mathcal{G}} and μ=P\mu = P, for all AGA \in \mathcal{G}, PG(A)=AYdP P_{\mathcal{G}} (A) = \int_{A} Y d P a Y0Y \ge 0 uniquely exists, satisfying the above. According to the initial definition of PGP_{\mathcal{G}}, YY becomes the conditional expectation of XX given G\mathcal{G}.


Case 2. General case

You can decompose XX into two parts X+,X0X^{+} , X^{-} \ge 0 and use the same method as in Case 1..