In his Probability Theory (Dover edition, pages 541 and 542; luckily the pages are available on [Google Books]), Renyi derives Hartley's formula the following way:
Given an alphabet of $N$ elements, we can choose a sequence of $k$ elements. There are $N^k$ such sequences, and if we want to uniquely identify each, we need $n_k$ bits. $n_k$ is characterized by:
$$ 2^{n_k-1} < N^k \leq 2^{n_k} $$
From here we can easily get to the inequality
$$ k \log_2 N \leq n_k < k \log_2 (N+1) $$
Now comes a statement that I don't understand. He writes that “it follows”:
$$ \lim_{k\to\infty} \frac{n_k}{k} = \log_2 N $$
This statement is not obvious to me. Can somebody help me think about why the inequality turns into an equality?
Is this statement in any way connected to what follows immediately after?
Thus for every $\epsilon > 0$ we can find a number $k$ such that if we take the elements of $E$ by ordered groups of $k$, then the identification of one element requires on the average less than $\log_2 N + \epsilon$ binary digits.