(In what follows, NB that $\mathbb{N}^\times$ is the positive natural numbers, and $\mathbb{N}$ includes 0.)
I am working on Exercise 1 of Amann and Escher Analysis I, and the problem is essentially asking:
Show that if $X$ is a set with a commutative and associative (addition) operation defined thereon, then $\sum_{j=0}^n x_j = \sum_{j=0}^n x_{\sigma(j)}$ where $\sigma$ is any permutation (bijection) of $\{0,1,...,n\}$.
I am really baffled as to how to even start. It's obviously true, but I need to prove it using the recursion theorem of set theory (Amann and Escher's Proposition 5.11). I understand how the standard sum notation ought to interpreted. I've written:
We must first interpret the notation $\sum_{j=0}^n x_{\sigma(j)}$. Given the definition for general associative operations in Example 5.12, one is led to presume that this notation refers to a recursive definition which is induced by the original. The original $\sum_{j=0}^n x_j$ is technically defined by Prop 5.11 in terms of some sequence $x_0,x_1,x_2,...$ in $X$ and some sequence of functions $V_n: X^n \to X, (y_0,...,y_{n-1}) \mapsto y_{n-1} + x_n$ for $n \in \mathbb{N}^\times$. Now take the unique function $f: \mathbb{N} \to X$ (by Prop 5.11) for which $f(0) = x_0$ and for all $n \in \mathbb{N}$, $f(n+1) = V_{n+1}(f(0),...,f(n)) = f(n) + x_{n+1}$. Then define $\sum_{j=0}^n x_j := f(n)$.
I am guessing that I need to interpret the notation $\sum_{j=0}^n x_{\sigma(j)}$ somehow in terms of a recursive definition related to the sequence $V_n$ which defined $\sum_{j=0}^n x_j$ (and then perhaps I need to use induction somehow?), but I can't see how given that $\sigma$ is defined for fixed $n$, among other things. Can anyone provide a hint or even a sketch of a proof, with careful attention paid to what $\sum_{j=0}^n x_{\sigma(j)}$ really means?
For those interested in all the gory details of what I have, I have attached a couple pictures of relevance, but hopefully my question is valid as it stands.

First of all, I don't believe you have sufficient foundation in basic logic to tackle such problems. This is not your fault, naturally, because it is hard to find good teachers. For you to truly grasp how everything in mathematics is built, you would need to learn basic FOL (first-order logic) including an actually usable deductive system such as this one.
As part of that, you need to understand precisely the logical underpinnings of definitions, which are captured by definitorial expansion. An important example is the pair-projection function-symbols
firstandsecondwherefirst(p)andsecond(p)are the first and second items inpfor any ordered pairp. (These can be defined rigorously in this way.) We can then use the notation( S x , T y ↦ E(x,y) )as short for( (S×T) i ↦ E(first(i),second(i)) )and similarly use the notation( S x , T y , U z ↦ E(x,y,z) )as short for( (S×(T×U)) i ↦ E(first(i),first(second(i)),second(second(i))) ).The proposition 5.11 that you cited from your book is actually a very shoddy one from a logical viewpoint. It has the nebulous ellipsis ("..."), which in general is imprecise and results in shoddy conceptual understanding. The real general recursion theorem is:
∀A,S∈set ∀c∈A→S ∀f∈A×(ℕ×S)→S ∃g∈A×ℕ→S ∀x∈A( g(x,0) = c(x) ∧ ∀k∈ℕ ( g(x,k+1) = f(x,(k,g(x,k))) ) ).This allows you to construct the summation fully rigorously like this (skipping minor steps):
I want to emphasize that nowhere in this 100% rigorous construction do we rely on assuming associativity or commutativity of + on ℂ! Only when we want to prove some further properties of Sum would we need those.
~ ~ ~
Now on to your question. As a beginner you should probably not use subscripts if you are dealing with a genuine function. You can see already that the
Sumoperator as defined above requires two inputs, one being a function fromℕ⁺→ℂand the other being a natural number that represents the index to sum to. Don't think of it as $\sum_{i=1}^k f_i$. Think of it as $\sum_{i=1}^k f(i)$, which is what it truly is.Then everything regarding permutations would become crystal-clear too. You want to prove that $\sum_{i=1}^k f(i) = \sum_{i=1}^k f(σ(i))$ for every $f{∈}ℕ⁺{→}ℂ$ and permutation σ on $[1..k]$. Of course, $\sum_{i=1}^k f(σ(i)) = \sum_{i=1}^k (f∘σ)(i)$, and so you ought to know precisely what that means. If conventional notation is confusing, simply use the precise formal notation and it is clear.
In particular, you must supply
Sumwith a function from ℕ⁺→ℂ and the index to sum to. Givenf∈ℕ⁺→ℂ, remember thatf(i)is not a function, butfis, and so isf∘σ. For convenience, let's not ask for the summation operator to be able to handle functions whose domain is not ℕ⁺, otherwise the above definition ofSumis insufficient. (Your cited text makes this mistake as well.) So eliminate this problem, we can define a permutation of $[1..k]$ as a function $f{∈}ℕ⁺{→}ℕ⁺$ such that $Im(f{↾}[1..k]) = [1..k]$.Rigorously:
Now the desired theorem is:
∀f∈ℕ⁺→ℂ ∀k∈ℕ ∀σ∈Perm(k) ( Sum(f,k) = Sum(f∘σ,k) ).How do we prove it? Ignoring all the other technical problems with the cited text, scleaver gives an intuitive method, but there is still a big problem: It is actually not at all easy to prove the theorem for a single arbitrary transposition. I will hence demonstrate an alternative approach:
For any permutation $σ$ on $[1..k]$ define $c(σ)$ to be the number of pairs $⟨i,j⟩$ from $[1..k]^2$ such that $i < j$ and $σ(i) > σ(j)$. It is easy to prove that:
By strong induction along $ℕ$ we can assume that the theorem holds for any permutation $φ$ with $c(φ) < c(σ)$. If $c(σ) = 0$ then we are done. Otherwise by the above let $d∈[1..k{−}1]$ such that $σ(d) > σ(d+1)$. Let $σ'$ be the same as $σ$ except that $σ'(d) = σ(d+1)$ and $σ'(d+1) = σ(d)$. Then we can show that $c(σ') < c(σ)$, and so $\sum_{i=1}^k f(i) = \sum_{i=1}^k (f∘σ')(i)$. We also have:
$\sum_{i=1}^k (f∘σ)(i)$
$ = \Big( \big( \sum_{i=1}^{d-1} (f∘σ)(i) + (f∘σ)(d) \big) + (f∘σ)(d+1) \Big) + \sum_{i=d+2}^{k} (f∘σ)(i)$ [by (✻)]
$ = \Big( \big( \sum_{i=1}^{d-1} (f∘σ')(i) + (f∘σ')(d+1) \big) + (f∘σ')(d) \Big) + \sum_{i=d+2}^{k} (f∘σ')(i)$
$ = \Big( \big( \sum_{i=1}^{d-1} (f∘σ')(i) + (f∘σ')(d) \big) + (f∘σ')(d+1) \Big) + \sum_{i=d+2}^{k} (f∘σ')(i)$
[by associativity and commutativity of + on ℂ]
$ = \sum_{i=1}^k (f∘σ')(i)$ [by (✻)]
$ = \sum_{i=1}^k f(i)$.
(✻) is one key property of summation that must be proven before you can do this kind of manipulation. It relies on associativity, but not commutativity, of + on ℂ. Also note that the "$\sum_{i=d+2}^{k}$" summation notation with a variable lower index can be defined in terms of the more basic summation notation with a fixed lower index of $1$. If you want to make it fully rigorous, you can either define it in terms of
Sum, or in the first place defineSumto take an extra input for the lower index. Both are viable choices.I have chosen not to formalize the entire theorem in full so that you can see enough to convince yourself of how it can be done, and if you are keen enough to try it you can ask me if you get stuck.