I am now reading through a book to understand Fano's inequality, but I remember my professor explaining it in a certain way that made it seem so logical.
I will go office hours as soon as possible, but for now can someone please try to explain to me Fano's inequality but not through math just in a "logical" way that makes sense.
Thanks a lot!!
Suppose that we wish to estimate a random variable $X$, by an estimator $\hat{X}$ . Further more assume that $\mathbb{P}(\hat{X} \neq X) = \epsilon$. The question is what can we say about the conditional entropy $H(X\mid\hat{X})$. Intuitively, if $\epsilon$ is very small, then $H(X\mid\hat{X})$ should also be very small. Fano's inequality quantifies this intuition $$ H(X\mid\hat{X})\leq H\left(\epsilon\right)+\epsilon\log\left(|\mathcal{X}\right|-1) $$ where $\mathcal{X}$ is the alphabet of $X$, and $|\mathcal{X}|$ is the cardinality of $\mathcal{X}$.
You can find more information in this book.