The Setup
Consider a hypergeometric distribution $X$ with parameters $N, n, m,$ i.e. $$\mathbb{P}[X = k] = \frac{{m \choose k} {N-m \choose n-k}}{{N \choose n}},$$ for $k$ running from $0$ to $\min\{n,m\}.$ The mean $\mathbb{E}[X]$ is $\frac{mn}{N}.$ Let $Y$ be a $\mathsf{Bi}(m,n/N)$ distribution, i.e. $$\mathbb{P}[Y = k] = {m \choose k}\left(\frac{n}{N}\right)^k \left(1-\frac{n}{N} \right)^{m-k}.$$ $Y$ also has mean $\frac{mn}{N}.$
(In $X,$ the roles of $m$ and $n$ are interchangeable, so the following should hold for $\mathsf{Bi}(n,m/N)$ also.) In Janson, Łuczak, and Ruciński's Random Graphs (JLR), the proof of Theorem 2.10 (page 29 in my edition) is largely left as an exercise. It depends on the following fact:
The Question
$$(1) \ \ \forall u \in \mathbb{R}, \mathbb{E}[e^{uX}] \leq \mathbb{E}[e^{uY}].$$ That is, the MGF of $X$ is less than or equal to the MGF of $Y.$ Intuitively, this makes sense, since $Y$ seems more likely than $X$ to be far from its mean. How can we prove this?
What I've Tried
JLR references Hoeffding 1963 for this result. I went through the parts of that paper that seemed most relevant to this problem--"Sampling from a Finite Population" and some of "Sums of Dependent Random Variables", but I was unable to find anything that proves $$\forall u \in \mathbb{R}, \mathbb{E}[e^{uX}] \leq \mathbb{E}[e^{uY}].$$ (I skimmed the rest of the paper as well.) One more remark about Hoeffding: when this paper talks about $\mathbb{E}[e^{cZ}]$ it seems to usually require $c>0$, whereas our desired result holds for all $u \in \mathbb{R}.$
I have tried writing out explicit expressions for these MGFs: $$\mathbb{E}[e^{uX}] = \sum_{k} \frac{{m \choose k} {N-m \choose n-k}}{{N \choose n}} e^{uk}$$ and $$\mathbb{E}[e^{uY}] = \sum_{k} {m \choose k} \left(\frac{n}{N} \right)^k \left(1-\frac{n}{N}\right)^{m-k} e^{uk}.$$ These summands share factors of ${m \choose k}$, and $e^{uk} > 0,$ so one might hope to prove $$(2) \ \ \forall k \leq m,n \leq N, \ \ \frac{{N - m \choose n-k}}{{N \choose n}} \leq \left(\frac{n}{N} \right)^k \left(1 - \frac{n}{N} \right)^{m-k}.$$ $(2)$, if true, would prove $(1), \mathbb{E}[e^{uX}] \leq \mathbb{E}[e^{uY}].$ I currently do not believe this inequality is true, based on some numbers I put into Maple (but I may have made an error there). In any case, although this inequality would be sufficient, it is not necessary for proving (1).
I also tried expanding $e^{uX} = 1 + uX + \frac{u^2 X^2}{2} + \cdots,$ and similarly for $e^{uY}.$ For $u \geq 0,$ (probably using Fubini-Tonelli somewhere), it would suffice to prove $$(3) \ \ \forall K \in \mathbb{N}, \ \ \mathbb{E}[X^K] \leq \mathbb{E}[Y^K].$$ I do not know how to do this (although it seems true), and in any case it wouldn't directly prove (1) when $u < 0.$
Let $r^{\underline k}$ denote the falling power $r(r-1)(r-2)\dotsm (r-k+1)$.
The hypergeometric $X$ denotes the number of successes among $m$ draws without replacement from a pool with $N$ options, $n$ of which are successful. So $X^{\underline k}$ denotes the number of $k$-tuples of distinct successes among the $m$ draws. By linearity of expectation, $$ \mathbb E[X^{\underline k}] = m^{\underline k} \cdot \frac{n^{\underline k}}{N^{\underline k}} $$ as there are $m^{\underline k}$ $k$-tuples of distinct draws, and a probability of $\frac{n^{\underline k}}{N^{\underline k}}$ that all $k$ draws in a $k$-tuple are successful.
By similar reasoning, we have $$ \mathbb E[Y^{\underline k}] = m^{\underline k} \frac{n^k}{N^k} $$ since here, the probability that all $k$ draws in a $k$-tuple are successful is just $(\frac nN)^k$.
Comparing these: we have $\frac{n-k}{N-k} \le \frac{n-k+1}{N-k+1} \le \dots \le \frac{n-1}{N-1} \le \frac nN$, so in particular $\frac{n^{\underline k}}{N^{\underline k}} \le \frac{n^k}{N^k}$, and $\mathbb E[X^{\underline k}] \le \mathbb E[Y^{\underline k}]$.
The $k^{\text{th}}$ derivative of $\mathbb E[z^X]$ with respect to $z$ is $\mathbb E[X^{\underline k} z^{X-k}]$, and so in particular the $k^{\text{th}}$ derivative at $1$ is just $\mathbb E[X^{\underline k}]$. Therefore when we compare the Taylor expansion of $\mathbb E[z^X]$ and $\mathbb E[z^Y]$ around $z=1$, the former always has smaller coefficients than the latter.
We conclude that $\mathbb E[z^X] \le \mathbb E[z^Y]$ for $z \ge 1$; in other words, $\mathbb E[e^{uX}] \le \mathbb E[e^{uY}]$ for $u \ge 0$.
For $u\le 0$, let $v=-u$, and note that $m-X$ and $m-Y$ are hypergeometric and binomial, respectively, with the same relationship as $X$ and $Y$. So we have $\mathbb E[e^{v(m-X)}] \le \mathbb E[e^{v(m-Y)}]$; dividing by the constant $e^{vm}$, we get $\mathbb E[e^{-vX}] \le \mathbb E[e^{-vY}]$.