Hofer-Zehnder, in section 1.5, proves that every Hamiltonian field on a strictly convex compact regular energy surface carries a periodic orbit. I have understood the proof. What I am wondering about is how he gets the variational principles. The start of the proof finds a function $H$ having the given convex hypersurface as energy surface and satisfying positive homogeneity of degree two, positive definiteness of the Hessian, and $\mathcal{C}^2$ regularity. Then the proof normalizes the period to get period $2\pi$, and then it goes:
For example, the variational principle:
$$\min\int_0^{2\pi}H\big(x(t)\big)\,dt\quad{\rm under}\quad\frac{1}{2}\int_0^{2\pi}\langle Jx,\dot x\rangle=1.$$
[…] has the above equations as Euler equations.
Firstly, what are the Euler equations of a variational principle? And secondly, how did they get this particular principle?
Update
Let me make my question more precise. There are 3-4 points I have trouble with.
- Why do they consider this principle with this strange constraint of that integral being 2? I mean, it isn't even a vector space! It's perhaps not even an affine space, it has no particular properties. Why should this constraint be useful? And why that functional?
- How do I get the Euler equations for this principle, and what are they? If I try to say "directional derivatives 0", I get stuck at once, because how am I supposed to make the variation?
- The book then goes on, and says this functional attains neither the infimum, nor the supremum. How can one prove these claims?
The book then Legendre transforms $H$ into $G$, gives properties of $G$, and introduces the same principle on $G$:
$$\min\int_0^{2\pi}G(\dot z)\,dt\quad{\rm under}\quad\frac{1}{2}\int_0^{2\pi}\langle Jz,\dot z\rangle=1.$$
Again, why should I consider this particular principle, with this particular constraint? I mean, I understood the rest of the proof so I know this is the right principle to consider, but how was it come up with? Is there any reason that would make one think of this principle as a possible correct choice?
Let's not consider a closed path for now but some smooth curve $\gamma:[0,1]\to M$ , and consider the functional $$S[\gamma]:=\int_0^1 \sum_ip_i(\gamma(t)) \mathrm d q_i(\gamma(t))-H(\gamma(t)) \mathrm d t$$ where I'm working in a chart with coordinates $(q_1,\dots,q_n,p_1,\dots,p_n)$. An old problem in classical mechanics (finding trajectories with fixed endpoints) was formulated such that this function is stationary wrt the true physical trajectory. You can read more about how calculus of variations works, so here I will get the first variation of this functional a bit heuristically. To get the variation, we compute $S[\gamma + \epsilon \eta]-S[\gamma]=\epsilon \delta S[\gamma]+O(\epsilon^2)$, where $\eta:[0,1]\to M$ is a variation with $\eta(0)=\eta(1)=0$. In practice, we can get the variation by operating with $\delta$ as if it were an ordinary differential, so for instance $\delta(a b)=a\delta b + b\delta a$, $\delta(f(x,y))=f_x \delta x + f_y \delta y$ etc. The Hamiltonian explicitly depends on $H(p_1(\gamma(t)),q_1(\gamma(t)),\dots,p_n(\gamma(t)),q_n(\gamma(t)))$, so in varying this we get
$$\delta S = \int_0^1 \sum_i\left(\delta p_i \mathrm d q_i + p_i\mathrm d \delta q_i\right)-\sum_i\left(\frac{\partial H}{\partial p_i}\delta p_i + \frac{\partial H}{\partial q_i}\delta q_i\right)\mathrm d t$$ and now we note that $p_i\mathrm d \delta q_i=p_i\delta q_i-\delta q_i\mathrm d p_i$. This means that we may integrate by parts: $$\delta S = \sum_i\int_0^1 \left(\mathrm d q_i -\frac{\partial H}{\partial p_i}\mathrm d t \right) \delta p_i -\left(\mathrm d p_i +\frac{\partial H}{\partial p_i} \mathrm d t \right) \delta q_i + \sum_i p_i\delta q_i|_{0}^1.$$ The last term is usually called a surface term, and it vanishes because the variation vanishes at the endpoints, since the endpoints are fixed.
Requiring $\delta S = 0$ gives $$\frac{\mathrm d q_i}{\mathrm d t}=\frac{\partial H}{\partial p_i},\,\frac{\mathrm d p_i}{\mathrm d t}=-\frac{\partial H}{\partial q_i}$$ along $\gamma$. This is the same as requiring that $\gamma$ be the flow of the Hamiltonian vector field $X_H$ given by $dH=i_{X_H}\omega$ (modulo a minus sign I always forget). Historically, this was the motivation for symplectic geometry, with the symplectic manifold as the cotangent bundle $T^*Q$ of the $n$-dimensional configuration space $Q$, which is nothing more than the manifold where (generalized) coordinates live. Alternatively we may write $$\dot x = J\nabla H$$
These are properly called Hamilton's equations, although they may be viewed as Euler-Lagrange equations, which are given by the same procedure but with a functional $\int L(\dot q,q;t)\mathrm dt$. The book is probably referring to Hamilton's equations as Euler equations, because that's what they are but for this particular functional.
Now, for your particular problem, you have periodic motion. I remember that you asked in this question about why over a closed path we get $$\int_{\gamma} \lambda:=\int_{\gamma} \sum_i p_i\mathrm d q_i = \frac12\int_0^{2\pi}\langle-J\dot x,x\rangle dt$$ The constraint is a normalization of this, basically. And since $\langle J x,Jy\rangle=\langle x,y\rangle$, this really is equivalent to your constraint. To obtain the correct equations of motion (as they are called) from variational calculus, when constraints are present, we use Lagrange multipliers. Consider the problem of making $\int F(\dot x,x,t) \mathrm dt $ stationary, where $F$ is some functional with the constraint $\int M(\dot x,x,t) \mathrm dt= m$.This can mean for instance that the total mass is m, whatever. The correct functional to which we may apply Euler-Lagrange equations is
$$\int\left(F(\dot x,x,t) + \lambda(x)M(\dot x,x,t)\right)\mathrm dt$$
So, by saying minimize $\int H(x(t))\mathrm dt$ along with the constraint, I'm really saying to minimize the so called "action functional", which is the one I started this answer with ($\lambda$ will drop out during calculation, and turn out to be a constant in this case). All of this should answer your questions 1. and 2.
Whether $S[\gamma]$ attains an infimum or supremum isn't important, only whether it's stationary is. I don't know how I would show what this attains a minimum or maximum. That would require the second variation, or something like the non-negativity of the Weierstrass excess (or simply E) function $E(x,y,z,w)=F(x,y,w)-F(x,y,z)-(w-z)F_{y'}(x,y,z)$.