drop independent random variables in conditional expectation expression

664 Views Asked by At

Is the following true for discrete random variables $K,Y,X_1,...$ where $K$ is independent of $\sigma(Y,X_1,...)$?

$$E[\sum_{i=1}^K X_i | K,Y] = \sum_{i=1}^K E[X_i|K,Y] = \sum_{i=1}^K E[X_i|Y]$$

Intuitively it seems ok, but I am lacking the rigorous argument. We only know when to drop a condition in conditional probability expressions like $P[X_i|K,Y] = P[X_i|Y]$.

Thinking of independence as "$K$ doesn't tell us anything about $Y,X_1,...$" should then also allow to drop the condition in the conditional expectation. Is it also rigorously speaking true?


draft (author is unsure about the following try):

Let's consider the factorization $E[\sum_{i=1}^K X_i\ |\ K,Y] = \underbrace{E[\sum_{i=1}^K X_i\ |\ (K,Y) = (\cdot,\cdot)]}_{(1)} \circ (K,Y)$.

Where we obtain (1) by

$$E[\sum_{i=1}^K X_i | K=k,Y=y] = \sum_{i=1}^k E[X_i|K=k,Y=y] = \sum_{i=1}^k E[X_i|Y=y]$$

Taking together we have $$E[\sum_{i=1}^K X_i\ |\ K,Y] = \sum_{i=1}^K E[X_i|Y]$$

2

There are 2 best solutions below

7
On BEST ANSWER

It might help to rewrite the random sum as $$ S=\sum_{k=1}^\infty\mathbf 1_{K\geqslant k}\,X_k. $$ Since every $\mathbf 1_{K\geqslant k}$ is measurable with respect to $\sigma(K,Y)$, $$ E(S\mid K,Y)=\sum_{k=1}^\infty\mathbf 1_{K\geqslant k}\,E(X_k\mid K,Y)=\sum_{k=1}^KE(X_k\mid K,Y). $$ This does not use the independence argument. All that is needed is that, for every $U$ measurable with respect to the sigma-algebra $\mathcal G$ and every $V$, $$ E(UV\mid\mathcal G)=UE(V\mid\mathcal G). $$ To go further, note that, for every sigma-algebras $\mathcal G$ and $\mathcal K$, if the sigma-algebras $\mathcal K$ and $\sigma(\sigma(X),\mathcal G)$ are independent, then $E(X\mid\mathcal G,\mathcal K)=E(X\mid\mathcal G)$.

Under the hypothesis that $K$ is independent of $(Y,(X_k)_{k\geqslant1})$, one can apply this general result to $X=X_k$, $\mathcal G=\sigma(Y)$ and $\mathcal K=\sigma(K)$, getting $$E(X_k\mid K,Y)=E(X_k\mid Y),$$ for every $k$. Thus, $$ E(S\mid K,Y)=\sum_{k=1}^KE(X_k\mid Y). $$ As a final caveat, note that if the random variables $(X_k,Y)$ are identically distributed, then $E(X_k\mid Y)$ does not depend on $k$, for example $E(X_k\mid Y)=E(X_1\mid Y)$ for every $k$ hence $$E(S\mid K,Y)=K\,E(X_1\mid Y),$$ but that the hypothesis that the sequence $(X_k)_{k\geqslant1}$ is i.i.d. is not enough to guarantee this.

5
On

What is true (and it's basically nonsensical to give a 'proof' unless you have had the measure theoretic definition of conditional expectation) is that under your hypotheses, $E(X_i|K,Y) = E(X_i|Y)$, which your intuitive grasp of the concepts involved should 'justify'.

With this in hand, we can prove the following result, which is similar to the one above (but I will assume all the $X_i$'s are identically distributed to get the nice-looking result). I think this might interest you as it has a similar form to your conjecture above.

$E(\sum_{i=1}^K X_i | K,Y) = E(\sum_{k=1}^{\infty}1_{K=k}\sum_{i=1}^k X_i | K,Y)$

$= \sum_{k=1}^{\infty}E(1_{K=k}\sum_{i=1}^k X_i | K,Y)$

$= \sum_{k=1}^{\infty}E(1_{K=k}) \sum_{i=1}^k E(X_i | K,Y)$

$= \sum_{k=1}^{\infty}P(K=k) \cdot k \cdot E(X_i|Y) $

$= \sum_{k=1}^{\infty}kP(K=k)E(X_1|Y) $

$= E(X_1|Y) \sum_{k=1}^{\infty}kP(K=k)$

$= E(K)E(X_1|Y).$

In the first equality is just partition the sample space over possible values of $K$; in the second a take a deterministic sum outside the expectation (this would require some justification using dominated convergence); in the third I use the fact that the random variable $1_{K=k}$ is $K$-measurable (and I use a property of conditional expectation called 'taking out what is known'); in the fourth I rewrite the expectation of an indicator as a probability, and use the independence result mentioned above; in the fifth I use the identical distributions of the $X_i$; in the sixth I take out the constant factor $E(X_1)$; and in the sixth I use the definition of expectation for a random variable $K$ taking values in $\{1,2,3,...\}$.