I have a question relating to convergence in probability and distribution for least squares estimators. Frequently, I see in textbooks that $\hat{\beta} \rightarrow^p b$. Where $b$ is the population parameter, and $\hat{\beta}$ is the Least Squares estimator of that parameter; demonstrating that LS estimators are consistent.
I also often see that, $\hat{\beta} \rightarrow^d N(b,\frac{1}{N}(X'X)^{-1})$, showing that $\hat{\beta}$ tends in distribution to a Normal centered around $b$.
I just wanted to check that my thinking here is correct. Convergence in probability always implies convergence in distribution as far as I understand it. Hence convergence in probability of LS estimators to a constant $b$ should imply convergence in distribution to the same constant $b$. Hence we should have that $\hat{\beta} \rightarrow^d b$. Is this satisfied above with $N(b,\frac{1}{N}(X'X)^{-1})$ since the variance tends to zero (since $N\rightarrow \infty$), meaning the distribution is itself that of a constant? Or is there some other reason that means that LS estimators converge in distribution to a constant?
Many thanks,
Ben
Many thanks for your response. You were correct I actually meant the asymptotic distribution of $\hat{\beta}$ is given by:
$\hat{\beta} \rightarrow^d N(b,\sigma^2 (1/n)Q^{-1})$
Where $\left[(1/n)(X'X)\right]^{-1}\rightarrow^p Q^{-1}$ as in the following notes: http://www3.grips.ac.jp/~yamanota/Lecture_Note_6_Asymptotic_Properties.pdf
I understand your reasoning, so am I correct in thinking that the above (distributional convergence) implies that because as $n \rightarrow \infty$, that really:
$\hat{\beta} \rightarrow^d b$
Since the variance goes to zero. Is this logic not correct?
However, as I understand it we can magnify the difference between $\hat{\beta}$ and b, by multiplying the difference between them by $n^{1/2}$, and it turns out that this only just balances the rate at which their difference goes to zero. Meaning this difference tends to a finite distribution:
$n^{1/2}(\hat{\beta}-b) \rightarrow^d N(0,\sigma^2 Q^{-1})$
I have done some research and it is often stated that:
$\hat{\beta} \rightarrow^a N(b,\sigma^2 (1/n) Q^{-1})$
Am I right in thinking that the above stands only for moderate n, since when n gets large the variance tends to zero? Does the 'a' stand for asymptotic distributed in the above?
Many many thanks (again),
Ben