I am reading some notes on Conditional Expectations as Orthogonal Projections (Page 60 - Section 3.5.2.3). There, they have an example of two biased coin, which I paraphrase.
Let $X$ denote a random variable that is $1$ if the first coin is heads (out of two unbiased coin tosses). I.e., the sample space is $\langle TT, TH, HT, HH\rangle$, and $X$ can be seen as a vector $\langle 0, 0, 1, 1\rangle$ in $\mathbb{R}^4$ corresponding to these outcomes. Now, they say $E[X]$ is a projection of $X$ onto subspace of constants, that is, $E[X] = E[\langle0, 0, 1, 1\rangle] = \langle\frac{1}{2}, \frac{1}{2}, \frac{1}{2}, \frac{1}{2}\rangle$. We can see that this is correct because $X - E[X]$ is orthogonal to the subspace of constants: $(X - E[X]) \cdot \langle 1, 1, 1, 1\rangle = 0$.
I don't understand what the subspace of constants is, and how they arrive at the conclusion that $E[X] = \langle \frac{1}{2}, \frac{1}{2}, \frac{1}{2}, \frac{1}{2} \rangle$. For example, when I don't understand how to generalize this example for two biased coin tosses with $\Pr(\text{Heads}) = p$.
The key to extending the ideas in that section of the notes to other probability spaces is hinted at in the last paragraph, where the authors write “… for example, $\mathbf E[XY]$ acts like an inner product for random variables…” In fact, you have to use this inner product to compute the orthogonal projections. Relative to the basis that consists of the indicator variables for the elementary events and treating the random variables as vectors now, this inner product can be computed as the matrix product $Y^TQX$, where $Q$ is a diagonal matrix with the probabilities of the elementary events along its main diagonal. If we represent these probabilities as $p_i$, this expression can be written as $\sum_i p_ix_iy_i$.
The authors are a bit cavalier about the inner product used in their examples: they use the dot product of the vectors to define orthogonality, which can be misleading. Per the previous paragraph, the matrix of the inner product for this space is actually $Q=\operatorname{diag}\left(\frac14,\frac14,\frac14,\frac14\right)$, yielding the inner product $\frac14(X\cdot Y)$. This is zero iff $X\cdot Y$ is zero, so the factor of $\frac14$ can be ignored when checking orthogonality. (In fact, this factor cancels when computing the angle between two vectors.)
For your unfair coin with $\Pr(\text{Heads})=p$, the matrix of the inner product is instead $$Q=\begin{bmatrix}(1-p)^2 &0&0&0 \\ 0& p(1-p) &0&0 \\ 0&0& p(1-p) &0 \\ 0&0&0& p^2 \end{bmatrix}.$$ You can then compute the corresponding orthogonal projections using well-known formulas: $\mathbf E[X]=\langle p,p,p,p \rangle$, as we’d expect, while $\mathbf E[X+Y \mid X] = \langle p,p,p+1,p+1 \rangle$, which jibes with the computation $\mathbf E[X+Y\mid X] = X+\mathbf E[Y]$ from a previous section. Using this inner product, $X-\mathbf E[X]$ is indeed orthogonal to $\mathbf 1 = \langle1,1,1,1\rangle$.