I came across this question that kind of motivated comparing this $\sigma^2$ estimator $$S^2 = \frac {\sum_{i=1}^n (Y_i - \bar{Y})^{2}}{n-1}$$ with this other $\sigma^2$ estimator $$S'^2 = \frac {\sum_{i=1}^n (Y_i - \bar{Y})^{2}}{n}$$
where $Y_1, \ldots, Y_n$ is a random sample.
First, it asked for $\operatorname{E}(S^2)$ and $\operatorname{E}(S'^2)$, and then it asked to compare $\operatorname{Var}(S^2)$ and $\operatorname{Var}(S'^2)$.
To find the expectations, I feel like it can go two ways here:
One potential way is using that $\frac{(n-1)S^2}{\sigma^2} \sim \chi^2(n-1)$ and somehow deriving a distribution for $S^2$ and then the expectation. But I feel like this won't work for $S'^2$.
The other potential way is to just apply
$$\operatorname{E}(S'^2) = \operatorname{E}\Biggl(\frac {\sum_{i=1}^n (Y_i - \bar{Y})^{2}}{n}\Biggr) = \frac{1}{n} \operatorname{E} \biggl({\sum_{i=1}^n (Y_i - \bar{Y})^{2}}\biggr) = \frac{1}{n} \biggl({\sum_{i=1}^n \operatorname{E} \bigr((Y_i - \bar{Y})^{2}\bigr)}\biggr) = \frac{1}{n} {\sum_{i=1}^n \operatorname{E}(Y_i^2) - 2 \bar{Y} \mu + \bar{Y}^2}$$
If I let $n$ be large, I can simplify it further using $\bar{Y} = \mu$:
$$= \frac{1}{n} {\sum_{i=1}^n \operatorname{E}(Y_i^2) - 2\mu^2 + \mu^2} = \frac{1}{n} {\sum_{i=1}^n \operatorname{E}(Y_i^2) - \mu^2} = \frac{1}{n} {\sum_{i=1}^n \operatorname{Var}(Y_i)} = \frac{1}{n} {\sum_{i=1}^n \sigma^2} = \frac{n\sigma^2}{n} = \sigma^2$$
I am not sure about this derivation though, because it does not really work for $S^2$.
Any help is appreciated.
You might want to consider this;
$\sum_{i=1}^n(Y_i-\overline Y)^2\\=\sum_{i=1}^n(Y_i-\mu+\mu-\overline Y)^2\\=\sum_{i=1}^n[(Y_i-\mu)-(\overline Y-\mu)]^2\\=\sum_{i=1}^n[(Y_i-\mu)^2-2(Y_i-\mu)(\overline Y-\mu)+(\overline Y-\mu)^2]\\=\sum_{i=1}^n(Y_i-\mu)^2-\sum_{i=1}^n2(Y_i-\mu)(\overline Y-\mu)+\sum_{i=1}^n(\overline Y-\mu)^2\\=\sum_{i=1}^n(Y_i-\mu)^2-2(\overline Y-\mu)\sum_{i=1}^n(Y_i-\mu)+(\overline Y-\mu)^2\sum_{i=1}^n1\\=\sum_{i=1}^n(Y_i-\mu)^2-2(\overline Y-\mu)(\sum_{i=1}^nY_i-\sum_{i=1}^n\mu)+n(\overline Y-\mu)^2\\=\sum_{i=1}^n(Y_i-\mu)^2-2(\overline Y-\mu)(n\overline Y-n\mu)+n(\overline Y-\mu)^2 \\=\sum_{i=1}^n(Y_i-\mu)^2-2n(\overline Y-\mu)^2+n(\overline Y-\mu)^2\\=\sum_{i=1}^n(Y_i-\mu)^2-n(\overline Y-\mu)^2$ $(*)$
Now we know that $Var(Y_i)=E[(Y_i-\mu)^2]=\sigma^2$ and $Var(\overline Y)=E[(\overline Y-\mu)^2]=\frac {\sigma^2}{n}$ by definition of variance.
Taking the expected value of $(*)$,
$E[\sum_{i=1}^n(Y_i-\overline Y)]\\=E[\sum_{i=1}^n(Y_i-\mu)^2-n(\overline Y-\mu)^2]\\=E[\sum_{i=1}^n(Y_i-\mu)^2]-E[n(\overline Y-\mu)^2]\\=\sum_{i=1}^nE(Y_i-\mu)^2-nE(\overline Y-\mu)^2\\=n\sigma^2-n.\frac {\sigma^2}{n}\\=(n-1)\sigma^2$
So clearly, $E(S^2)=E[\frac {\sum_{i=1}^n(Y_i-\overline Y)^2}{n-1}]=\frac {(n-1)\sigma^2}{n-1}=\sigma^2$
This explains why the first formula is a better approximation for $\sigma^2$. Had you used the second formula, then the expected value, $E(S^{'2})=\frac {(n-1)\sigma^2}{n}$, which clearly shows that it is an underestimation.