Suppose we have a channel which transmits sequences of length n, of 0s and 1s (i.e. A={0,1} to the nth -> B={0,1} to the nth, such that during transmission, it will randomly (with equal probability) choose one digit from the initial sequence, and multiply it by 0.
e.g. we send in 01110, and it happens to choose the third digit, yielding 01010.
How to calculate the capacity of a channel like this? i.e. what is the max over A of (I(A,B))?
I've struggled with this exercise, I see many potential ways of dealing with it (e.g. simply finding the zero point of the derivative of I(A,B), but that would require quite complicated maths, and I am bound to make mistakes).
Perhaps it could be calculated using some theorems, or approximated using some other channels?
I'd be very grateful for any hints.
Remark 1: In my original answer I had misunderstood the channel effect (many thanks to @leonbloy for pointing that out). I have rewritten the answer, hoping I got it right this time :)
Remark 2: As noted in the comments by @leonbloy, the assumption $H(E|X,Y)=0$ I am making below is not correct. I will leave, however, leave this answer, as it may serve as a starting point for someone else to give it a shot. However, the lower bound stated below still holds, as, clearly, $H(E|X,Y)\geq 0$
Let $X \in \{0,1\}^n$ and $Y \in \{0,1\}^n$ denote the input and output of the channel, respectively, and $E\in \{0,1,\ldots,n-1\}$ denote the position of the input signal that the channel replaces with a zero.
I will use the following standard formula for the mutual information between input and output : $$ I(X;Y) = H(Y) - H(Y|X), $$ in bits per $n$ channel uses.
Consider the computation of $H(Y)$ first. Using the following two expressions for the joint entropy $H(Y,E)$, \begin{align} H(Y,E) &= H(E) + H(Y|E),\\ H(Y,E) &= H(Y) + H(E|Y), \end{align} it follows that $$ H(Y) = H(E) + H(Y|E) - H(E|Y). $$
Similarly, we can show that \begin{align} H(Y|X) &= H(E|X)+H(Y|X,E)-H(E|Y,X)\\ &= H(E)+0-0, \end{align} which follows by noting that $E$ is independent of $X$ and that $Y$ and $E$ are completely determined when $X$, $E$ and $Y$, $X$, are known, respectively. Therefore, we have $$ I(X;Y) = H(Y|E) - H(E|Y). $$ Let's compute the two terms of the left hand side.
Regarding $H(Y|E)$, note that when $E$ is known, the corresponding element of $Y$ is also known (equal to zero) and the uncertainty about $Y$ is with respect to its other $n-1$ elements. Assuming that the input symbols are equiprobable, it follows that $H(Y|E) = n-1$.
The tricky part is the computation of $H(E|Y)$. First note that by observing a realization $Y=y$, we know that $E$ must be restricted to one of the positions of $y$ that are zero. Since any of these positions can be the actual one with equal probability, it follows that $$ H(E|y) = \log_2 k_y, $$
for $1\leq k_y\leq n$, where $k_y \triangleq|\{i:y_i=0\}|$ is the number of elements in $y$ that are zero. Note that if $k_y=1$ (one zero in $y$) then we now that the zero element is where the error is and the conditional entropy becomes zero, as expected.
Now,
\begin{align} H(E|Y) &\triangleq \sum_y \mathbb{P}(Y=y) H(E|y)\\ &= \sum_{i=1}^n \mathbb{P}(k_y=i) \log_2 i \\ &= \sum_{i=1}^n \binom{n-1}{i-1} \left(\frac{1}{2} \right)^{i-1} \left(\frac{1}{2} \right)^{n-1-(i-1)}\log_2 i \\ &= \left(\frac{1}{2} \right)^{n-1}\sum_{i=1}^n \binom{n-1}{i-1} \log_2 i \\ \end{align}
where we used the fact that $k_y=i$ can occur if, apart from the error position, there are $i-1$ out of the rest $n-1$ positions of the input that are zero. It appears that the last expression cannot be further simplified. Of course, one could consider the trivial bound \begin{align} H(E|Y) &\leq H(E)\\ &=\log_2 n, \end{align} resulting in $$ I(X;Y) \geq n-1-\log_2 n. $$
Just for fun, I plot below the numerically evaluated normalized mutual information $I(X;Y)/n$ (in bits per channel use). Note that as $n$ increases, it approaches the bound of the ideal channel ($1$ bit per channel use). The bound becomes tight for large $n$.