Suppose that the conditional entropy $H[Y|X]$ between two discrete random variables $x$ and $y$ is zero.
Show that, for all values of $x$ such that $p(x) > 0$, the variable $y$ must be a function of $x$, in other words, for each $x$ there is only one value of $y$ such that $p(y|x)\ne 0$
Therefore, if $H[y|x]=0$, then there exists a function $g$ such that $y=g(x)$.
Would it be sufficient to prove this by looking at Venn Diagram and noticing that if $H[Y|X]=0$, then:
$$H[y] = H[x] - H[x|y]$$
Therefore, $H[y|x]=0$ if and only if the value of $y$ is completely determined by the value of $x$.
I want to say that $H[x|y]$ depends only on $x$ variable and not on variable $y$... but not sure how to prove it.
I guess, the question is: how to make such a bold statement like above. iff $H[y|x]=0$, $y$ is a function of $x$!

Based on the definition of Entropy:
$$H[y|x] = - \sum_{x_i} \sum_{y_j} p(x_i, y_j) \ln p(y_j | x_i)$$
Considering the property of probability, we can obtain that:
$$0 \le p(y_j|x_i) \le 1$$
$$0 \le p(x_i, y_j) \le 1$$
Therefore, we can see that:
$$-p(x_i, x_j) \ln p(y_j | x_i) \ge 0$$
when:
$$0 < p(y_j|x_i) \le 1$$
when:
$$p(y_j|x_i) = 0$$
provided with the fact that:
$$\lim_{p \to 0} p \ln p = 0$$
we can see that:
$$-p(x_i,y_j) \ln p(y_j | x_i) = -p(x_i) p(y_j|x_i) \ln p(y_j|x_i) = 0$$
(here we view p(x) as a constant).
Hence for an arbitrary term in the equation above, we have proven that it cannot be less than 0. In other words, if and only if every term of $H[y_j|x_i]$ equals 0, $H[\mathbf{y}|\mathbf{x}]$ will equal 0.
Therefore, for each possible value of random varuable x, denoted as $x_i$:
$$-\sum_{y_i} p(x_i, y_i) \ln p(y_i|x_i) = 0$$
If there are more than one possible value of random variable "y given $x_i$", $p(y_i|x_i)$, such that:
$$p(y_i|x_i) \ne 0$$
(Because $x_i$, $y_i$ are both "possible", $p(x_i, y_i)$ will also not equal to 0), constrained by:
$$0 \le p(y_i|x_i) \le 1$$
and
$$\sum_j p(y_i|x_i)=1$$
there should be at least two values of y that satisfy:
$$0 < p(y_j|x_i) < 1$$
which ultimately leads to:
$$-\sum_{y_i} p(x_i, y_i) \ln p(y_i|x_i) > 0$$
Therefore, for each possible value of x, there will only be one y such that:
$$p(y|x) \ne 0$$
In other words, y is determined by x.
Note: If y is a function of x, we can obtain the value of y as soon as observing a x. Therefore we will obtain no additional information when observing a $y_j$ given an already observed x.