I would appreciate either one of the following, or both:
- A source (if a book, with page numbers) where I can find this result proven
- A proof of the result
The question at hand:
Let $X_{11}, \dots, X_{1n_1}$ be independent and identically distributed random variables with mean $\mu_1$ and variance $\sigma_1^2$, and let $X_{21}, \dots, X_{2n_2}$ be independent and identically distributed random variables with mean $\mu_2$ and variance $\sigma_2^2$. Assume $n_1 \neq n_2$ and $\sigma_1^2 \neq \sigma_2^2$.
Denote $\bar{X}_{1, n_1} = \dfrac{1}{n_1}\sum_{i=1}^{n_1}X_{1i}$ and $\bar{X}_{2, n_2} = \dfrac{1}{n_2}\sum_{i=1}^{n_2}X_{2i}$. Also, let $S_1^2 = \dfrac{1}{n_1 - 1}\sum_{i=1}^{n_1}(X_{1i} - \bar{X}_{1, n_1})^2$ and $S_2^2 = \dfrac{1}{n_2 - 1}\sum_{i=1}^{n_2}(X_{2i} - \bar{X}_{1, n_2})^2$.
As $n_1 \to \infty$ and $n_2 \to \infty$, does the statistic $$T = \dfrac{\bar{X}_{1, n_1} - \bar{X}_{2, n_2}}{\sqrt{\dfrac{S_1^2}{n_1} + \dfrac{S_2^2}{n_2}}}$$ converge in distribution to a random variable; and if so, with what distribution?
Context. The test statistic $T$ is that which arises from Welch's t-test. Conventional statistical wisdom (e.g., here, here) is that regardless of the population distributions of $X_{1i}$ and $X_{2j}$ (i.e., they are not iid normal), the Central Limit Theorem (CLT) may be used so as to justify that $T$ is approximately $\mathcal{N}(0, 1)$. I haven't seen a proof of this, and I am doubtful the classical CLT may be used.
My Efforts. I demonstrated that for a single population, with obvious notational extensions, it holds that $\dfrac{\bar{X} - \mu}{S/\sqrt{n}}$ converges in distribution to a random variable with an $\mathcal{N}(0, 1)$ distribution. However, this result cannot be used in this question.
For one thing, $\bar{X}_{1, n_1} - \bar{X}_{2, n_2}$ cannot be written as a single arithmetic mean as in the result above. For another thing, while $S_1^2 \to \sigma_1^2$ and $S_2^2 \to \sigma_2^2$ in probability, since $\sigma_1^2 \neq \sigma_2^2$ they cannot be "factored out" like in my demonstration, prohibiting direct use of the CLT.
Additional assumption: $\frac{n_{2}}{n_{1}}\to c$ and $n_{1}\land n_{2}\to \infty$,
Denote $n\equiv n_{1}\lor n_{2}$. Construct the following triangular array
$$\begin{matrix}Y_{1,1}\\Y_{2,1}&Y_{2,2}\\Y_{3,1}&Y_{3,2}&Y_{3,3}\\ \cdots&\cdots&\cdots&\cdots\\ Y_{n,1}&Y_{n,2}&Y_{n,3}&\cdots&Y_{n,n}\\\cdots&\cdots&\cdots&\cdots&\cdots&\cdots\end{matrix}$$
with $Y_{n,i}\equiv \frac{\sqrt{n_{2}}}{n_{1}}X_{1,i}1_{\{i\leq n_{1}\}}-\frac{1}{\sqrt{n_{2}}}X_{2,i}1_{\{i\leq n_{2}\}}$. Then it remains to prove $$\frac{\sum_{i=1}^{n}Y_{n,i}}{\sqrt{\frac{n_{2}}{n_{1}}\sigma_{1}^{2}+\sigma_{2}^{2}}}\to_{d}\mathcal{N}(0,1) \quad\text{Under }\mathbb{H}_{0}:\mu_{1}=\mu_{2}.$$
By construction $Y_{n,i}$ is row-wise independent, also we have $$\mathbb{E}[Y_{n,i}]=\frac{\sqrt{n_{2}}}{n_{1}}\mu_{1}1_{\{i\leq n_{1}\}}-\frac{1}{\sqrt{n_{2}}}\mu_{2}1_{\{i\leq n_{2}\}},$$ and $$\mathrm{Var}(Y_{n,i})=\frac{n_{2}}{n_{1}^{2}}\sigma_{1}^{2}1_{\{i\leq n_{1}\}}+\frac{1}{n_{2}}\sigma_{2}^{2}1_{\{i\leq n_{2}\}}.$$ This gives $$\sum_{i=1}^{n}\mathbb{E}[Y_{n,i}]=\sqrt{n_{2}}\mu_{1}-\sqrt{n_{2}}\mu_{2}=0\quad \text{(under the null)},$$ and $$\mathrm{Var}\biggl(\sum_{i=1}^{n}Y_{n,i}\biggr)=\sum_{i=1}^{n}\mathrm{Var}(Y_{n,i})=\frac{n_{2}}{n_{1}}\sigma_{1}^{2}+\sigma_{2}^{2}.$$
The desired convergence is guaranteed by triangular array CLT
Note that $$ \begin{align*} Y_{n,i}^{2}&=\Bigl(\frac{\sqrt{n_{2}}}{n_{1}}X_{1,i}1_{\{i\leq n_{1}\}}-\frac{1}{\sqrt{n_{2}}}X_{2,i}1_{\{i\leq n_{2}\}}\Bigr)^{2}\\ &\leq 2\Bigl(\frac{n_{2}}{n_{1}^{2}}X_{1,i}^{2}1_{\{i\leq n_{1}\}}+\frac{1}{n_{2}}X_{2,i}^{2}1_{\{i\leq n_{2}\}}\Bigr), \end{align*}$$ separate terms in the summation and by dominated convergence theorem, it's easy to verify the Lindeberg condition.