I am preparing a quantitative finance interview and I am struggling with this exercise:
Consider two data series, X = (x1, x2, . . . , xn) and Y = (y1, y2, . . . , yn), both with mean zero. We use linear regression (ordinary least squares) to regress Y against X (without fitting any intercept), as in Y = aX + $\epsilon$ where $\epsilon$ denotes a series of error terms.
Suppose that ρXY = 0.01. Is the resulting value of a statistically significantly different from 0 at the 95% level if:
i. $n = 10^2$ \ ii. $n = 10^3$ \ iii. $n = 10^4$ \
I already know the relation between a and $\rho$ is given by $$a = \frac{\rho_{XY}}{\sigma_X}$$
But I am struggling with the confidence level part.
Any help would be appreciated. Thank you!
You can use the $F$-test in order to calculate statistically significance, given you hypotheis you have that $$ F_n= \frac{\rho^2}{1-\rho^2}*(n-2) $$
Hence you obtain the following $F_{100} = 0.0098$, $F_{1000} = 0.098$ and $F_{1000}=0.98$. In order to have a significance level at 0.05 you should hav $F>f(1,n-2,1-\alpha)$ where $f(1,n-2,\alpha)$ is the $\alpha$ percentile of a $F$ distribuition with parameters 1 and $n-2$. You can find these values on a table or on some online F-calculator. For the 3 values of $n$ this number is almost 3.5 hence you cannot reject the null hypothesis i.e. you value $\rho =0.01$ is not significative at the given confidence.