Expected value of $X_N$ with smallest index $N$ for which $\sum_{i=1}^N X_i$ exceeds $1$ when $X_i$ are uniformly distributed

155 Views Asked by At

From an interview book, where the answer is not so clear I believe. You keep generating $\mathcal U_{[0,1]}$ iid random variables until their sum exceeds 1, then compute the expected value of the last random variable, i.e. the one responsible for letting the sum of rvs overflow 1.

My idea (not working):

The $i$-th draw from $\mathcal U_{[0,1]}$ is called $X_i$, and $S_N:=\sum_{i=1}^N X_i$.

I aim to compute: $$\mathbb E\left[X_{N}\right], N:=\min \left[i:\sum_i X_i > 1\right].$$

Rewrite it as: $$\mathbb E\left[X_{N}\right] = \sum_{i=2}^\infty \mathbb E\left[X_{N}|N=i\right]\mathbb P[N=i].$$

From this question I know that $\mathbb P[N=i] = (i-1)/i!$.

I know that $X_N$ takes positive values between 0 and 1, so I use the expectation of the tail function: $$\mathbb E\left[X_{N}|N=i\right]=\int_0^1 \mathbb P[X_N>t|N=i]\ \text d t= 1-\int_0^1 \mathbb P[X_N\leq t|N=i]\ \text d t.$$

Now, some relabeling, using $X$ for the generic $\mathcal U_{[0,1]}$ and $Y$ for $S_{i-1}$: $$\mathbb P[X_N\leq t|N=i]=\mathbb P[X_i\leq t|S_{i-1}<1 \cap S_{i-1}+X_i> 1]=\mathbb P[X\leq t|Y<1 \cap X> 1-Y].$$

Now reversing the conditioning: $$\mathbb P[X\leq t|Y<1 \cap X> 1-Y] = \frac{\mathbb P[X\leq t\cap X> 1-Y|Y<1]}{\mathbb P[X> 1-Y|Y<1]}.$$

Now, from the same interview book I know that $\mathbb P[S_N\leq y|S_N < 1]=y^N$, so the density function of $Y|Y<1$ ends up being $(i-1)y^{i-2}\ \text dy$.

By total probability, conditioning over the value of $Y$, I write: $$\mathbb P[X\leq t|Y<1 \cap X> 1-Y] = \frac{\int_0^1\mathbb P[X\leq t\cap X> 1-Y|Y<1, Y=y](i-1)y^{i-2}\ \text dy}{\int_0^1\mathbb P[X> 1-Y|Y<1, Y=y](i-1)y^{i-2}\ \text dy}.$$

The numerator leads to the integral: \begin{equation}\tag{error is here!} \int_0^1\mathbb P[X\leq t\cap X> 1-Y|Y<1, Y=y](i-1)y^{i-2}\ \text dy=\\ \int_0^1(t-1+y)(i-1)y^{i-2}\ \text dy=\dots=t-1+\frac{i-1}i. \end{equation} The denominator similarly: $$\int_0^1y(i-1)y^{i-2}\ \text dy=\dots=\frac{i-1}i.$$

Substituting: $$\mathbb P[X>t|Y<1 \cap X> 1-Y] =1 - \mathbb P[X_N\leq t|N=i]=i\frac{1-t}{i-1},$$

$$E\left[X_{N}|N=i\right]=\int_0^1 \mathbb P[X_N>t|N=i]\ \text d t=\int_0^1 i\frac{1-t}{i-1}\ \text d t=\frac{i}{2(i-1)}$$

$$\mathbb E[X_{N}] = \sum_{i=2}^\infty \mathbb E[X_{N}|N=i]\mathbb P[N=i]=\sum_{i=2}^\infty \frac{i}{2(i-1)}(i-1)/i!=\dots=\frac{e-1}2.$$

The answer should be (verified via MC) $2-\frac e2$.

Would you mind checking my procedure and letting me know where it is wrong?


EDIT: Found the problem thanks to two answers below, adding for completeness.

The probability $P[X\leq t\cap X> 1-Y|Y<1, Y=y]$ should actually be written as (answer by Noble Mushtak):

$$P[X\leq t\cap X> 1-Y|Y<1, Y=y]=\\ =P[X\leq t\cap X> 1-Y\cap 1-Y<t|Y<1, Y=y]=\\=P[1-Y<X\leq t|1-t<Y<1, Y=y]P[1-t<Y<1|Y<1, Y=y]=\\=(t-1+y)\int_{1-t}^1(i-1)y^{i-2}\ \text dy=\dots$$

If so, then (also found by Amir):

$$\mathbb E[X_{N}|N=i]=\frac{i+2}{2(i+1)}.$$

3

There are 3 best solutions below

1
On BEST ANSWER

This step is wrong:

$$\int_0^1\mathbb P[X\leq t\cap X> 1-Y|Y<1, Y=y](i-1)y^{i-2}\ \text dy= \int_0^1(t-1+y)(i-1)y^{i-2}\ \text dy$$

You can't say $\mathbb P[X\leq t\cap X> 1-Y|Y<1, Y=y]=t-1+y$ because sometimes $t-1+y$ is negative. Instead, for this probability to be nonzero, we need $1-y \geq t$, which is the same as $y \geq 1-t$, so we need to adjust the integral limits accordingly. We then get:

$$ \begin{align*} \int_0^1\mathbb P[X\leq t\cap X> 1-Y|Y<1, Y=y](i-1)y^{i-2}\ \text dy &= \int_{1-t}^1(t-1+y)(i-1)y^{i-2}\ \text dy \\ &=(t-1)(1-(1-t)^{i-1})+\frac{i-1}{i}(1-(1-t)^i) \end{align*} $$

You also have a second mistake later on: You are substituting $(it-1)/(i-1)$ for $\mathbb{E}[X_N \mid N=i]$, but this isn't right, you should use your integral from before to compute this expectation: $$\mathbb E[X_{N}|N=i]= 1-\int_0^1 \mathbb P[X_N\leq t|N=i]\ \text d t.$$ In fact, I'm not sure how you figured out the value of that infinite sum, since that infinite sum still uses $t$ even though $t$ does not mean anything in that context.


Unfortunately, I don't know how to correct this reasoning, I think the problem can be solved in the way you are doing, but I am not good at infinite sums. Instead, I'll use differential equations.

For any constant $r$, I define the following program $P(r)$.

  1. Let the variable $s$ be $r$.
  2. Generate a random quantity $q$ from $\mathbb{U}_{[0,1]}$.
  3. Subtract $s$ by $r$.
  4. If $s < 0$, then output $q$ and exit.
  5. Otherwise, go back to step 2.

This is the same random process that you are describing, except I am start with a target sum $r$ and then subtract from there, whereas you start with $0$ and then keep adding until you get to the target sum of $1$. However, these processes are clearly equivalent: Subtracting from $r$ down to some number less than $0$ is the same as adding from $0$ up to some number greater than $r$. I am also using a parameter $r$ for the target sum, whereas in your program, the target sum is $1$. Let $f(r)$ be the expected value of the output of $P(r)$. We will solve for $f(r)$, and then the final answer is just $f(1)$.

Now, in this program, there are basically two cases:

  1. The first number $q$ we generate is less than $r$, in which case the process becomes equivalent to $P(r-q)$.
  2. The first number $q$ we generate is at least $r$, in which case the output is just $q$.

Using these two cases, we get this equation for the expected output of $f(r)$:

$$f(r)=\int_0^r f(r-q)dq + \int_r^1 qdq$$

Use $u=r-q$ in the first integral and simplify the second integral.

$$f(r)=\int_0^r f(u)du+\frac{1-r^2}{2}$$

Take the derivative of both sides:

$$f'(r)=f(r)-r$$

By guess and check, we find that one solution to this differential equation is $f(r)=r+1$. The general solution to the homogeneous equation $f'(r)=r$ is $f(r)=Ce^r$ for arbitrary $C$, so the general solution to this differential equation is $f(r)=Ce^r+r+1$ for arbitrary $C$.

Now, we can find $C$ using $f(0)$: Clearly, $f(0)$ just generates $q$ and always outputs $q$, so $f(0)$ is just the expectation of $q$, which is $1/2$. Ergo, we have: $$ f(0)=Ce^0+0+1=\frac{1}{2} $$ so $C=-1/2$. Ergo, $$ f(1)=-\frac{1}{2}e^1+1+1=2-\frac{e}{2} $$

3
On

Your idea can work as follows.

For $i\in \{2,3,\dots \}$, we can show that for $\color{blue}{t\in (0,1)}$:

$$\color{blue}{F_{X_{N}|N=i}(t)=1-\left ( 1+ \frac{1}{i-1} \right ) (1-t) +\frac{1}{i-1} (1-t)^{i}} \tag{1}.$$

This gives

$$\color{blue}{\mathbb E[X_{N}|N=i]}=\int_0^1 \left [1- F_{X_{N}|N=i}(t) \right ] \text d t \color{blue}{=\frac{i+2}{2(i+1)}}.$$

Finally, we have

$$\color{blue}{\mathbb E[X_{N}] }= \sum_{i=2}^\infty \mathbb E[X_{N}|N=i]\mathbb P[N=i]\\=\sum_{i=2}^\infty \frac{i+2}{2(i+1)}\times \frac{i-1}{i!}\color{blue}{=2-\frac{e}2}.$$

As an aside, using (1), the cdf of $X_N$ can be also obtained as $\color{blue}{t\in (0,1)}$

$$\color{blue}{F_{X_{N}} (t)}= \sum_{i=2}^\infty F_{X_{N}|N=i}(t) \times \frac{i-1}{i!}=\color{blue}{e^{1-t}-e(1-t)}$$

with the pdf

$$\color{blue}{f_{X_{N}} (t)=e-e^{1-t}, \, t\in (0,1)}.$$


Why (1) holds?

For $0<t<1$,

$$F_{X_{N}|N=i}(t)=\mathbb P (X \le t |N=i)=\frac{\mathbb P (X \le t, N=i)}{ \mathbb P (N=i)} \tag{2}.$$

From the provided link, we know $$\mathbb P (N=i)=\frac{i-1}{i!}, i=2,3,\dots \tag{3}. $$

Using the equivalence $ N=i \, \equiv \, S_{i-1}<1\le S_i=S_{i-1}+X_i$, we obtain

$$\mathbb P (X \le t, N=i)=\mathbb P (X \le t, S_{i-1}<1\le S_{i-1}+X_i). $$

From the provided link, we also know that the pdf of $S_{i-1}$ for $0<s<1$ is given by

$$f_{S_{i-1}}(s)=\frac{1}{(n-2)!}s^{i-2},$$

so considering that $X_i\sim \mathcal U (0,1)$ and $S_{i-1}$ are independent, we have

$$\mathbb P (X \le t, S_{i-1}<1\le S_{i-1}+X_i)=\int_{1-t}^1 \mathbb P (1-s \le X_i \le t) \frac{1}{(n-2)!}s^{i-2} \text d s=\int_{1-t}^1 (t-1+s) \frac{1}{(n-2)!}s^{i-2} \text d s. \tag{4} $$

From (2), (3), and (4), we obtain (1).

0
On

Another way: \begin{gather*} f_{S_{N-1}|N}(s|n)=ns^{n-2}\int_1^{1+s}dx=ns^{n-1},\quad E[S_{N-1}|N]=\frac{n}{n+1}\\ f_{S_N|N}(s|n)=\int_{s-1}^1nx^{n-2}dx=\frac{n(1-(s-1)^{n-1})}{n-1},\quad E[S_N|N]=\frac{3n+2}{2n+2}\\ E[X_N|N]=E[S_N|N]-E[S_{N-1}|N]=\frac{3n+2}{2n+2}-\frac{n}{n+1}=\frac{n+2}{2n+2}\\ E[X_N]=\sum_2^\infty\frac{n+2}{2n+2}\frac{n-1}{n!}=2-\frac{e}{2} \end{gather*}