Having an old exercise book without a solution to backward-engineer it. I would like the concept to be explained.
Having a race between 17 people, the 1st person has 3/4 posibility of winning and all (16) others have 1/64 posibility each.
If after the race we get a message that the winner was not the 1st, what is the information that message contains, and how much uncertainty is left about the winner of the race?
The event that correspond to having somebody else then the first to win have probability $1/4$, the amount of information is usually measured with the $-\log_2$ function (at least in information theory and for entropy and such information measures), hence the information contained in the message is $-\log_2(1/4) = 2$ bits.
The uncertainty left about the race is the expected amount of information that remains. When we know the first runner isn't the winner then the winner is uniformly distributed over the rest of the $16$ other runners by symmetry hence the entropy is $-16\cdot\frac{1}{16}\log_2\left( \frac{1}{16} \right)=4$ bits.