Definition of Conditional Expectation and its Uniqueness

311 Views Asked by At

there are many questions on Mathexchange, e.g. Definition of conditional expectation, discussing the definition of conditional expectation. In most cases the random variable which is later named "conditional expectation" is assumed to be a measurable $\Omega\rightarrow\mathbb R$ function. In some lecture notes (that are maybe too advanced for me) conditional expectation is defined to be a $\Omega\rightarrow\mathbb R^n$ measurable function -- but imposes in the definition that this function is a.e. unqiue. For me it feels that we do not need a.e. uniqueness in the definition but that it can be proven (I failed though). So how is uniqueness of conditional expectation proven in $\mathbb R^n$ or even more general, in Banach spaces?

Edit: Interestingly, this five-year-old question receives some attention now. The conditional expectation $Z$ of a random variable $X:\Omega\rightarrow S$ given a sub-$\sigma$-algebra $\mathcal G$ satisfies $$\int_G Z\,\mathrm dP = \int_G X\,\mathrm dP$$ for all $G\in\mathcal G$. The common notation is $\mathbb E[X\mid\mathcal G] := Z$.

It can be shown in the case $S = \mathbb R$ that if such a random variable exists, it is unique. The proof can be found e.g. here: Proof that conditional expectation is defined uniquely almost everywhere? This proof does not work for $S = \mathbb R^n$, or more general $S = $ some Banach space. The question is: how can I prove uniqueness in these cases?

1

There are 1 best solutions below

1
On

The additional imposed condition (of the uniqueness up to a set of measure zero) in the definition for $\mathbb{R}^n$ might be just a way of phrasing it and condensing several statements into one. I am not aware of any principal difference between conditional expectations of variables taking values in $\mathbb{R}$ and $\mathbb{R}^n$. There could be some room for a formal subtlety of whether you define the expectation as a random variable or as a class of equivalence of random variables. I believe it is most convenient to define the conditional expectation as any representative of the equivalence class, and then proceed to establishing that if the measurable spaces are nice, then we can choose representatives from the classes of equivalence in such a way that the family of functions $$\Sigma\ni A \mapsto P(A|t):=\mathbb{E}[1_{A}| T=t],\quad \forall t\in \mathrm{range}(T)$$ has the properties we expect from a conditional probability measure $P(\cdot|t)$. Here $\mathbb{E}[1_{A}| T=t]$ is a measurable function from $\mathrm{range}(T)$, which are defined using Doob-Dynkim lemma, as described in this Wikipedia article. Note that $\mathbb{E}[1_{A}| T]$ and $\mathbb{E}[1_{A}| T=t]$ are different objects. See e.g. Chapter 2.5 in Testing Statistical Hypotheses by Lehmann and Romano for details on choosing the right representatives from the equivalence classes.

Your question about proving the uniqueness up to a set of measure zero is answered here. [Edit]

I don't feel qualified to talk about general Banach spaces, but the starting point is the Bochner Integral.