Assume we have a function
$$H(x) = -\sum_{i=1}^n x_i \log x_i$$ where $0 < x_i < 1$ $\forall i$ (this is the Shannon entropy if you are familiar with it).
I am reading a paper in which the authors stated the following approximation $$H(x) = -\sum_{i=1}^n x_i \log x_i \approx \sum_{i=1}^n x_i (1-x_i)$$ Can anyone tell me the intuition behind this approximation?
$\log (1-y) = - \sum_{n=1}^\infty \frac{y^n}{n}, |y|<1$.
If $0<x<1$, then $0<1-x<1$, let $y=1-x$,
$\log(1-(1-x))=\log x = -\sum_{n=1}^\infty \frac{(1-x)^n}{n}$.
$-\log x = \sum_{n=1}^\infty \frac{(1-x)^n}{n}$.
Dropping higher order terms: $\log x \approx 1-x$.