Occasionally, you will literally see people arguing for nonstandard analysis purely to unnest the quantifiers in the definition of a limit. By unnesting, I mean avoiding an exists quantifier that is required to be nested inside the scope of the for all epsilon quantifier, and instead having only foralls that can be made implicit and constraints on the corresponding variables. NSA does this by just constraining $N$ to be infinitely large or $x$ to be infinitely close to $x_0$.
But there's a much simpler way to clean things up so that the quantifiers are no longer nested. Just define functions:
- A sequence $a_n$ in a metric space is convergent with limit L if there is a function $N : \mathbf{R^+} \to \mathbf{N}$ such that $d(a_n,L) < \varepsilon$ for all $\varepsilon > 0$ and $n > N(\varepsilon)$.
- A sequence $a_n$ in a metric space is a Cauchy series if there is a function $N : \mathbf{R^+} \to \mathbf{N}$ such that $d(a_n,a_m) < \varepsilon$ for all $\varepsilon > 0$ and $n,m > N(\varepsilon)$.
- A function $f : X \to Y$ between metric spaces has the limit $L = \lim_{x \to x_0} f(x)$ if there is a function $\delta : \mathbf{R^+} \to \mathbf{R^+}$ so that $d(f(x),L) < \varepsilon$ for any $\varepsilon > 0$ and $d(x,x_0) < \delta(\varepsilon)$.
which lets you write the definitions with only foralls that can be floated out to top level so that you just have one inequality between top level variables implying the other.
Furthermore, actually requiring that this function be defined (occasionally called the modulus of convergence), is both very convenient in epsilon delta proofs, and actually necessary to have for anything related to numerics/computability/constructive analysis or if you generally care about "how many terms do I need" questions.
So this made me curious. Is there any fundamental reason why you would want to write the definition of a limit in terms of nested quantifiers? I'd say tradition, but afaik first order logic is quite a bit newer than the definition of a limit. How did the 19th century analysts state the definition?
Here's a possible "logical" answer that may have to do with it, after thinking about it.
As Mark S said in his comment to the question, a proposition written in a form where the quantifiers are foralls that can be floated to top level is said to be in Skolem normal form.
This is an advantage primarily because the quantifiers can be made implicit and the entire proof ends up only needing propositional logic. This is convenient in $\varepsilon - \delta$ proofs because it allows you to reuse epsilons in many cases.
This is in contrast to many proofs written by analysts which tend to be in continuation passing style. This means that analysts then have to either keep defining new epsilons in the scope of the old ones, or use proofs by contradictions or blanket existence statements to "escape" the continuation.
...and this might be the actual reason why. Analysts write proofs in CPS style because of the definition of limits, which often get escaped by using proofs by contradiction in order to keep steps of proofs short and modular. Which means that they end up with more nonconstructive existence lemmas, which are contagious and lead to more CPS style proofs.
Since most results are then stated in terms of existence without error terms, and those are contagious since in some cases you need choice to escape them, in order to cite existing results you have to keep writing proofs in the same style, and analysts just end up building a stronger tolerance to that specific proof style than mathematicians in other fields and become used to it.