In search and optimization, a popular (set of) theorem(s) exist called the "No Free Lunch Theorem(s)". The authors of this theorem describe it as "for any algorithm, any elevated performance over one class of problems is offset by performance over another class" (https://ieeexplore.ieee.org/document/585893). These theorems have important interpretations in many applied fields such as Machine Learning. From a layman's perspective, the implication of the "No Free Lunch Theorem" asserts that there is no "universally best" class of algorithm for all "classes of optimization problems".
I was wondering if a similar type of theorem exists for Polynomial Approximation. For example, we know that there are many methods in mathematics that can be used for function approximation - for instance, Taylor Polynomials, Lagrange Polynomials, Padé Polynomials, etc.
As in the case of Machine Learning algorithms, can any claims similar to those asserted by the "No Free Lunch Theorems" be made regarding different methods of Polynomial Approximations? For example - is it possible that Taylor Polynomials might "better" approximate some functions compared to Padé Polynomials; but by virtue of this fact, there also might exist some other functions that Padé Polynomials approximate better than Taylor Polynomials? Or are have we been able to mathematically determine that some approximation methods are "universally better" than other approximation methods for broad classes of functions?
Thank you!
Note: I suppose that in this context, "better" can be quantified through properties such as "strength of convergence", "size of the radius of convergence" or properties of the "bounded error" (e.g. magnitude) - but I am not sure about this.
There are plenty of functions that cannot be approximated by polynomials at all. The simplest one is probably
$$f(x) = |x|$$
If you actually try to fit a Lagrange polynomial onto $n = 3,5,7,9,11,13$ points, it would end up looking like this:
Polynomials aren't the end-all-be-all of function approximation. We think that they do work quite well on analytic (infinitely differentiable) functions, but "most" functions are actually non-differentiable, like $f(x) = |x|$ is.
Of course, in reality, you would likely not try to fit a polynomial onto $|x|$, but it's reasonable to assume that there are formulas you could derive from the real world that do include absolute values or points where the function isn't differentiable.
A good example for this is superconductivity. The resistance for superconductors actually drops from a measurable, positive value to $0$, as soon as it hits a temperature threshold:
If you didn't know this and tried to fit a polynomial onto some points measured from the red curve, you'd be disappointed, because it wouldn't work.
So all in all, the "No Free Lunch" theorem would be something like:
Meaning that no matter how much you try to approximate such an $f$ with a polynomial $p$, there will always be uncountably infinitely many (indeed, $\beth_1$) points in $\mathbb{R}$, where $p(x)$ is arbitrarily far from $f(x)$.
This is easy to see from the absolute value example. For a function $f$ with the property
$$\lim_{x\to a+} f'(x) \ne \lim_{x\to a-} f'(x)$$
it is "locally an absolute value function" near $a$. (For the actual absolute value function the left side is $1$, while the right side is $-1$.) This means that even in the local area around $a$, it isn't possible to approximate $f$ with any polynomial.