In regression analysis one finds a line that fits best by minimizing the sum of squared errors.
But why squared errors? Why not use the absolute value of the error?
It seems to me that with squared errors the outlyers gain more weight. Why is that justified? And if it is justified to give the outlyers more weight, then why give them exactly this weight? Why not for example take the least sum of exponetial errors?
Edit: I am not so much interested in the fact that it might be easier to calculate. Rather the question is: does squaring the errors result in a better fitting line compared to using the absolute value of the error?
Furthermore I am looking for an answer in layman's terms that can enhance my intuitive understanding.
Many insightful answers here.
I'd like to share something I came across awhile ago that might help you with your edited question:
No, squaring the errors doesn't always result in a better fitting line.
Here's a figure comparing the best fit lines produced by L-1 regression and least squares regression on a dataset with outliers:
Click here for figure
As you've pointed out, outliers adversely affect least squares regression. Here's an instance where least squares regression gives a best fit line that "pans" towards outliers.
Full credit to: matlabdatamining.blogspot.sg/2007/10/l-1-linear-regression.html