I have a dataset with some of the following values:
independent variable n: dependent variable t
196: 8.32E-05
676: 0.000360012
..: ..
2739025: 17.19871902
4422609: 34.82757854
I am trying to match this empirical data to the closest function. For this, I use SPSS -> Analyze -> Regression -> [curve estimation / Nonlinear]. When I analyses the outcome, it appears that the large values are fitted in the same order of magnitude (42 vs 35), whereas the small values are off by multiple orders of magnitude (0.05 vs 0.00008).
I assume this is because of a square sum error measure.
My question: how can I perform an analysis that creates a more balanced fit, preferably using SPSS?
P.S. I apologize if math.stackexchange.com is the wrong forum. If so, please let me know which one to use instead.
In usual least squares analysis you square the error at each point and add the squares to get the total error, which you minimize. If you have values like $35$ which are fitted with $42$ that contributes $(42-35)^2=49$ to the error. If a value like $0.05$ is fitted with $0$ it contributes $0.05^2=0.0025$ to the error, so it doesn't contribute much unitil the fit gets to be something like $7$.
One approach is to take the log of all your data and fit that. You will then be asking the fitter to match each data point within a multiplicative error instead of an additive error. Another approach is to weight some data points more than others in the sum of squares. A third approach is to use a functional form that is guaranteed to go through $(0,0)$