Note: I've asked this question on "cross validated stackexchange" with no biters. I guess this question belongs here instead. (I've deleted my question there.)
In a basic statistics course we see CLT-like theorems appear for three cases of $\hat \theta$:
- For $\hat \theta=\frac{\sigma_X}{\sqrt{n}}$, this is classical CLT.
- For $\hat \theta=\frac{s}{\sqrt{n}}$, where $s$ is the sample variance.
- In the case of hypothesis testing, with $H_0: \mu_U=\mu_W$, let $X=U-W$. Then use $\hat \theta=\sqrt{\frac{s_1^2}{n}+\frac{s_2^2}{m}}$, where $s_1$ is the sample variance for $U$, and $s_2$ is the sample variance of $W$.
So this makes me wonder. Is the statement in the title of the question, which generalizes all of these cases, correct?
References (or counter-examples) would be ideal. If the statement in the title of the question incorrect, then what is the right generalization of CLT that captures cases 1-3 above?
As you said, the classical CLT stating that
$$\sqrt{n} \frac {\bar{X} - \mu_X} {\sigma_X} \stackrel {d} {\to} \mathcal{N}(0,1)$$
For any consistent estimator $\hat{\theta}$ of $\sigma_X$, we have $$ \hat{\theta} \stackrel {p} {\to} \sigma_X$$
Then by Slutsky Theorem, $$\sqrt{n} \frac {\bar{X} - \mu_X} {\hat{\theta}} = \sqrt{n} \frac {\bar{X} - \mu_X} {\sigma_X} \frac {\sigma_X} {\hat{\theta}}\stackrel {d} {\to} \mathcal{N}(0,1)$$
So this is a common result which statisticians are using everywhere - using a consistent estimator to "replace" the parameter.