In basic numerical analysis, it is shown that Aitken's method improves on the basic iteration method in speed of convergence in the asymptotic sense (see detail below if desired). Now it seems that this should be meaningless in practice, giving no guarantees of 'faster' for any finite number of iterations. I found that this is not only my feeling, but that this concern is echoed in this Wikipedia article:
Although strictly speaking, a limit does not give information about any finite first part of the sequence, this concept is of practical importance in dealing with a sequence of successive approximations for an iterative method, as then typically fewer iterations are needed to yield a useful approximation if the rate of convergence is higher. This may even make the difference between needing ten or a million iterations insignificant.
My questions then are
Barring empirical evidence, is there ANY formal way of turning the asymptotic result into a result in terms of finite iterations? Even at least probabilistically? Even when conditions are added? This would be an example of what I have in mind: Given a function with condition such and such (smooth, etc.), the convergence is indeed faster in $n$ iterations with probability $p(n)$.
If there are such results, can you point me to some of them?
If there are no such results, should there be no interest in trying to find them? If not why not?
Doesn't this state of affairs 'bother' numerical analysts? If not shouldn't it?
Do people reading this have their own 'intuitions' about when the speed of convergence holds in practice? What are these intuitions? Why not try to formalise them?
Detail
When the series $\{x_n\}$ generated using $x_i = f(x_{i-1})$ converges under the usual conditions, Aitken's method, generating the series $\{x'_n\}$ using $$ x'_n = x_n - \dfrac{(x_{n+1}-x_n)^2}{(x_{n+2}-2x_{n+1}+x_n)} $$ converges faster in the sense that, with $s$ being the solution $f(s)=s$, we have $$ \dfrac{x'_n-s}{x_n-s} \to 0, n \to \infty. $$
No. It is a daily occurrence that a numerical analyst codes up an algorithm which has better theoretical scaling/convergence only to find out that another with 'worse' theoretical bounds performs better for the application at hand. An example is that Gauss-Kronrod quadrature is faster than double-exponential quadrature at double precision, even though double-exponential quadrature has faster theoretical convergence. (However, it turns out that for higher precision, double-exponential wins.)
I don't use my intuition because it's been wrong so often. I use logging and performance benchmarks to watch the convergence rate for the problem at hand. On the other hand, the ideas of backwards stability, condition number, algorithmic scaling and convergence are nonetheless the best theoretical guideposts we have.