In Mackay's Information Theory book (p. 111), he's said that
Let the source alphabet be $\mathcal{A}_X = \{a_1, . . . , a_I\}$, and let the $I$th symbol $a_I$ have the special meaning ‘end of transmission’.
which seems quite arbitrary to me. More importantly, there might involve a risk of corrupting decoding process, since $a_I$ cannot be translated (it's replaced with end of transmission symbol).
To specify my idea, suppose, we have $\mathcal{A}_x=\{a,b,c\}$ and probabilities $p (a) = 0 . 25 , p (b) = 0 . 25 , p (c) = 0 . 5$. If we want to encode a string $abc$, how to decide the end of transmission symbol and its probability? Are we randomly set some probability $p(d)$ as end of transimisson and then rescale $p(a), p(b), p(c)$ according to that probability?
Edited
To be clear, the scope of question is artihmatric coding.
Ideally, you will choose a probability for EOT dependant on how long your message typically is, as though it were part of every message in your corpus. Thus in your probability selecting corpus "abc" is really "abc." And "qrxjundieko" is really "qrxjundieko." And so forth.
Note that my use of a period there implies that there are no other periods in the entire corpus. Really you use a special symbol that's guaranteed not to be there.