This question about the intuition behind the scaling property of the Fourier transform made me wonder about the corresponding notion for a Fourier series.
The Fourier transform of $f(ax)$ is $\frac{1}{|a|} \mathcal{F(\frac{u}{a})}$. If $a>1$ then the graph of $f(ax)$ is $f$ compressed and so its Fourier transform has frequencies that are higher. However they are scaled down in magnitude.
On the other hand take $f(x)= \cos x$. Its fourier series is trivially itself, with the coefficient of $\cos x$ being $1$. Scaling it to $f(ax) = \cos ax$ where $a>1$ is an integer still has a trivial fourier series, and the coefficient of $\cos ax$ is $1$. This is unlike the Fourier transform where the "coefficients" of each frequency get scaled as the formula shows: $\frac{1}{|a|} \mathcal{F(\frac{u}{a})}$.
Is there a conceptual way to explain this discrepancy?
I think I can see what you're getting at -- you want to view the Fourier series as a Fourier transform that happens to consist of $\delta$ peaks instead of a smoothly varying density, and then you wonder why those $\delta$ peaks don't appear to behave the same under scaling as an ordinary smooth Fourier transform does.
First off, I don't think the default formalism for Fourier transforms actually allows for this viewpoint. It can probably be formalized in some way, though, such as by letting the transform produce (or act on?) a complex measure instead of an ordinary function. Never mind; let's imagine it can work and see where that leads us.
Now, intuitively it is tempting to think of $C\delta(x-x_0)$ as a "density" that is $0$ everywhere except at $x_0$ where it is $C\cdot\infty$. But that doesn't really work -- infinities don't work that way in the first place, and in particular the height of a delta peak doesn't behave quite like an infinitary density. And you've chanced upon one of the differences, namely how they react to scaling.
For an ordinary smooth density function $F$ we can scale it by $a$ by using $G(t)=\frac{1}{a} F(\frac{t}{a})$. The inner factor of $\frac{1}{a}$ does the actual stretching of the horizontal axis, and the outer factor of $\frac{1}{a}$ corrects for the resulting area increase, such that $$\int_{ap}^{aq} G(t)\;dt = \int_{ap}^{aq}\frac1a F(\frac ta)\;dt = \int_p^q F(u)\;du$$ for all intervals $[p,q]$. On the other hand, a delta peak needs no such outer factor, because at the peak itself the $\delta$ function cannot see that its argument has been stretched! Instead we have $$\int_{ap}^{aq} \delta(\frac xa-x_0)\;dx = \int_p^q \delta(x-x_0)\;dx$$ with no outer factor needed to compensate for the horizontal stretching of the peak, because the peak has width 0 and so doesn't stretch at all.
And that is why you're seeing a difference in the behavior of Fourier coefficients versus Fourier transforms.