Let $X$ and $Y$ be two bounded random variables in a probability space $(\Omega,\mathcal{F},P)$. $G$ is a sub $\sigma$-algebra on $\Omega$. I have to prove that: $$E[XE[Y\mid G]] = E[YE[X\mid G]]$$ As I found in this question, it is quite easy to use the property "taking out what it is known" if $X$ or $Y$ are $G$-measurable. But in my case, $X$ and $Y$ are neither necessarily $G$-measurable nor independent, so I can't use those properties. I tried to follow the steps on page 6 in this pdf but it doesn't say why this equality holds: $$\lim_n E[Z_nX] = \lim_n E[Z_nE[X\mid G]]$$ It is supposed to be understood from number 6 in the same page: $$E[XY1_A] = E[YE[X\mid G]1_A] \space\space\space\space \forall A \in G $$ But I don't know how to drop the indicator function $1_A$. What I think is that the author is using this property: $$\text{if } E[XY1_A] = E[YE[X\mid G]1_A] \space\space\space\space \forall A \in G \Rightarrow E[XY] = E[YE[X\mid G]]$$ but I'm not 100% sure if my reasoning is correct.
2026-04-03 13:47:31.1775224051
On
Proving that $E[XE[Y\mid G]] = E[YE[X\mid G]]$
2.9k Views Asked by Bumbble Comm https://math.techqa.club/user/bumbble-comm/detail At
2
There are 2 best solutions below
0
On
We have $E[XE(Y | G)] = E[E(XE(Y | G)|G)] = E[E(Y | G)E(X | G)]$.
And $E[YE(X | G)] = E[E(YE(X | G)|G)] = E[E(X | G)E(Y | G)]$.
The first equalities follow immediately from the definition of conditional expectation (as noted in my comments), and the second equalities follow from the "taking out what is known" principle that you mention.
Here are two ways to explain this.
If $Z$ is a bounded ${\cal G}$-measurable random variable, then $$ E[ X Z ] = E [ E[X|{\cal G}] Z ].$$
Therefore, letting $Z=E[Y|{\cal G}]$, we have $$ E [ X E[Y |{\cal G}]] = E [ E[ X | {\cal G}] E [ Y |{\cal G}]]\quad (1)$$
But the same holds for $Y$. Namely, if $Z$ is bounded and ${\cal G}$-measurable, then $$ E [ Z Y ] = E [ Z E [Y |{\cal G}]]\quad(2).$$ Letting $Z = E[X |{\cal G}]$, (2) gives $$ E[ E [X|{\cal G}] Y ] = E [ E[X|{\cal G}] E[Y|{\cal G}]]\quad (3)$$ The result now follows by comparing (1) and (3).
Recall that the mapping $f \to E[f|{\cal G}]$ is an orthogonal projection from $L^2 (\Omega,{\cal F},P)$ to its (closed) subspace of ${\cal G}$-measurable random variables. The inner product in $L^2(\Omega,{\cal F},P)$ is given by
$$ (f,g) = E [ f g] .$$
Now since $T_{\cal G}$ is an orthogonal projection, it satisfies $T_{\cal G}^* = T_{\cal G}$. In particular,
$$ (T_{\cal G} f,g)= (f, T_{\cal G} g).$$
Set $f= X$ and $g=Y$, and the result follows.
(note that what we did in the probabilistic approach was exactly moving $T_{\cal G}$ from one variable to the other, so these actually represent the same idea).