Averages: why do we minimize the error function, is it simplicity?

82 Views Asked by At

If we have a sequence of several random numbers one way to predict the next ones minimizing the overall error is the average.

I was explained it to some extent here.

I don't get very well why the function that we minimize is the sum of the squares: $$ \sum_i^n (x_i - \alpha)^2 = f(\alpha) $$

I do understand that using $\sum x_i -\alpha = f(\alpha)$ wouldn't take us far though (apparently).

And is there any visual or geometric way to find the solution instead?

1

There are 1 best solutions below

0
On

I want to add to the comment made above is that the least square approximation doesn't even require the knowledge of distribution of the random variable . So, that makes him special among all the method of estimation . However there are some limitations to it as well because the no of unknowns should be equal to the no of knowns in order to use the least square estimation.