I am trying to solve the following exam practice problem:
Let $X_1, X_2$ be independent exponential random variables with parameter $1$.
- Find the conditional PDF of $X_1+X_2$ given that $\frac{X_1}{X_2}=1.$
- Find the conditional PDF of $X_1+X_2$ given that $X_1-X_2=0.$
- The events $\frac{X_1}{X_2}=1$ and $X_1-X_2=0$ are the same. Does this mean that conditioning on either of these two events should give the same answer?
Here is my approach:
For part 1, I computed the joint PDF of $Y= X_1+X_2$ and $Z=\frac{X_1}{X_2}$ and my result was $$f_{Y,Z}(y,z) = \frac{ye^{-y}}{(z+1)^2}\mathbf{1}_{\{y\geq 0, z \geq 0\}}.$$ This shows that $Y$ and $Z$ are independent. So, conditioned on $Z=1$, we get that $$f_{Y|Z}(y|1) = ye^{-y}\mathbf{1}_{\{y\geq 0\}}.$$
For part 2, I computed the joint PDF of $U=X_1+X_2$ and $V=X_1-X_2$ and got $$f_{U,V}(u,v) = \frac{1}{2}e^{-u}\mathbf{1}_{\{u\geq |v|\}}.$$ I then used Bayes' theorem to get that $$f_{U|V}(u|0) = e^{-u}\mathbf{1}_{\{u\geq 0\}}.$$
My question regards part 3. Firt of all, part 3 states that the events $X_1-X_2=0$ and $\frac{X_1}{X_2}=1$ are the same. But, I disagree because the event $X_1-X_2=0$ contains the measure-zero event $X_1=X_2=0$, which is not contained in the event $\frac{X_1}{X_2}=1$. Is this not the case? Second of all, if these two events were indeed the same, then why would conditioning on them produce a different result?
I tend to feel that conditioning on $X_1=X_2$ is not really consistent with the assumption that $X_1$ and $X_2$ are independent, so seeing things like the Borel–Kolmogorov paradox should not be a surprise.
You might say that conditioning on an event of probability $0$ is meaningless in itself as it leads to $\frac00$, but can only be interpreted as the limit of conditioning on an event of positive probability. Here
as $\epsilon_a,\epsilon_b,\epsilon_c,\epsilon_d \to 0_+$, and there is no reason to expect the two limiting distributions to be the same.
The first constraint tends to cut out small values of pairs of $X_1$ and $X_2$ which have large relative differences but small absolute differences and to leave in large values of such pairs which have small relative differences but large absolute differences, while the second constraint does the opposite, so it is not particularly a surprise that for small $\epsilon$s the first has a mean for $X_1+X_2$ near $2$ while the second has a mean near $1$, each consistent with your earlier answers.
Using the following R code to simulate this with $\epsilon=0.1$
you see the different simulated means are near $2$ and $1$ and the different empirical cumulative distribution functions are shown below in red and blue, broadly consistent with your answers.
The next chart of some of the simulated pairs helps illustrate how the red constraint tends to select more larger values and fewer smaller values than the blue constraint
To me it is curious that your first answer is a $\text{Gamma}(2,1)$ distribution while your second is an exponential distribution with mean $1$, while I would have had an intuitive attraction to an answer which was an exponential distribution with mean $2$, i.e. the distribution of $2X_1$