Linear Regression Seber - Why is the matrix A written in this way?

88 Views Asked by At

Example 1.9 of Seber’s book states the following: enter image description here

The result stating that if $\rho = 0$ it holds that $Q = \sum _i (X_i - \overline X)^2$ has expected value $\sigma ^2 (n-1)$ follows from $\operatorname{tr}(\textbf{A}\Sigma) + \mu ' \textbf A \mu = E[\textbf{X'AX}]$ and the fact that if $\Sigma = \sigma ^2 \textbf{I}_n$ then it follows that $\operatorname{tr}(\textbf{A}\Sigma) = \sigma ^2 \operatorname{tr}(\textbf A)$, with $\mu$ being the expected value vector.

I managed to understand why the matrix multiplication yields the result $\sigma ^2(1-\rho)\textbf A$. What I struggle to understand is why he can/would write write $\textbf A$ in such a way. Where is this matrix $A = [(\delta _{ij} - n^{-1})]$ coming from? It has been a while since I last looked at any linear algebra but when I tried to derive the matrix inducing $Q$ I must have gotten stuck somewhere along the way - or so I think. Perhaps there is something in the notation I am confusing? I don’t know. I would greatly appreciate any help.

2

There are 2 best solutions below

2
On

Write $$Q = \sum_{i}X_{i}^{2} - \frac{1}{n}\left(\sum_{i}X_{i}\right)^2.$$ Let $H$ be the matrix with all entries equal to one. The first term is $X^{T}X$ while the second one is $\frac{1}{n} X^{T}HX$. Thus $$Q = X^{T}X - \frac{1}{n}X^{T}HX = X^{T}\left(I-\frac{1}{n}H\right)X = X^{T}AX.$$

2
On

Define the $n$-vector $\textbf Y=(X_1-\overline X, \ldots, X_n-\overline X)^\top$ and note that $$ \begin{aligned} \textbf Y_1 &= \left(1-\frac 1n\right)X_1-\frac 1n X_2-\ldots-\frac 1n X_n \\ \textbf Y_2 &= \frac 1nX_1-\left(1-\frac 1n\right) X_2-\ldots-\frac 1n X_n \end{aligned} $$ etc, which suggests defining $\textbf A$ as done by Seber, which then yields $\textbf Y = \textbf A \textbf X$.

Next, $Q$ is merely $\|\textbf Y\|_2^2$, and since $\textbf A$ is an orthogonal projection, $$E[Q] = E[\textbf X^\top \textbf A^2 \textbf X] = E[\textbf X^\top \textbf A \textbf X] = E[\operatorname{tr}(\textbf A \textbf X \textbf X^\top)] = \operatorname{tr}(A E[\textbf X \textbf X^\top]) = \operatorname{tr}(\textbf A \Sigma).$$