
I can figure out the basics to the question, that is the mean and variance of Y:
- E(Y) = 1-2p
- Var(Y) = 4p(1-p)
I don't understand parts (i) and (ii). I dont understand the question itself, that is what is X=Y1 + Y2 + Y3, and how to use it for the latter of the question.
- Why is E(X) = nE(Y) and Var(X)=nVar(Y)
Can someone please explain the logic, and working out process for i and ii. How does the textbook conclude this.
The first relationship follows from the linearity of expectation: for any random variables $Y_1, Y_2, \ldots, Y_n$, $$\mathrm{E}[Y_1 + Y_2 + \cdots + Y_n] = \mathrm{E}[Y_1] + \mathrm{E}[Y_2] + \cdots + \mathrm{E}[Y_n].$$ They need not be independent nor identically distributed.
The second relationship is true only if the variables in the sum are all mutually independent; that is to say, for each $i \ne j$, $\mathrm{E}[Y_i Y_j] = \mathrm{E}[Y_i]\mathrm{E}[Y_j].$ Then it is easy to show that $$\mathrm{Var}[X] = \mathrm{E}\left[(X - \mathrm{E}[X])^2\right] = \mathrm{E}[X^2] - \mathrm{E}[X]^2,$$ and then $$\mathrm{E}[X^2] = \mathrm{E}\left[\sum_{i=1}^n Y_i^2 + \sum_{i \ne j} Y_i Y_j\right] = \sum_{i=1}^n \mathrm{E}[Y_i^2] + \sum_{i \ne j} \mathrm{E}[Y_i Y_j],$$ by the linearity of expectation. We also have $$\mathrm{E}[X]^2 = \biggl( \sum_{i=1}^n \mathrm{E}[Y_i] \biggr)^2 = \sum_{i=1}^n \mathrm{E}[Y_i]^2 + \sum_{i \ne j} \mathrm{E}[Y_i]\mathrm{E}[Y_j].$$ But if $Y_1, Y_2, \ldots, Y_n$ are independent, then the second term in each of the above expressions are equal to each other, so their difference is $$\mathrm{E}[X^2] - \mathrm{E}[X]^2 \underset{\mathrm{ind}}{=} \sum_{i=1}^n \mathrm{E}[Y_i^2] - \mathrm{E}[Y_i]^2 = \sum_{i=1}^n \mathrm{Var}[Y_i].$$ Then if $Y_1, Y_2, \ldots, Y_n$ are also identically distributed, then they each have a common variance, so $\mathrm{Var}[X] = n \mathrm{Var}[Y_1]$ as claimed.