I'm going through the proof of the symmetrization lemma (Vershynin), which says
$$ \frac{1}{2} E \left\| \sum_i \epsilon_i X_i \right\| \leq E \left\| \sum_i X_i \right\| \leq 2 E \left\| \sum_i \epsilon_i X_i \right\| $$ where $X_1,\ldots,X_N$ are independent mean-zero random vectors in a normed space. The $\epsilon_i$'s are i.i.d taking on $1$ or $-1$ with equal probability.
However, I don't see why independence or the mean-zero assumption is required anywhere for this to hold because the steps in the proof below don't seem to use either assumption.
Anyone know why this is?
References:
Vershynin, Roman, High-dimensional probability. An introduction with applications in data science (to appear), ZBL06872475.



Here's why independence is needed. Let's simplify our situation to one-dimensional, and all the $X_i$'s are just Rademacher variables. We'll do something extreme and make everything perfectly correlated. So as $n \to \infty$, the term $E \left\|\sum_i \epsilon_i X_i\right\|$ is the expectation of the absolute value of a Gaussian variable with variance $n$ and is roughly $O(\sqrt{n})$. The middle term on the other hand grows as $O(n)$.
We can show that mean zero is needed in a similar fashion. Consider $X_i = 1$ for all $i$ (I know it's silly but they are independent!) The middle term will now grow as $O(n)$ and the term on the right grows as $O(\sqrt{n})$.